From 4fc1cf3de1dbb866bfb4e754319eb27084b053d5 Mon Sep 17 00:00:00 2001 From: mrveiss Date: Fri, 3 Apr 2026 23:32:14 +0300 Subject: [PATCH] style(imports): fix isort ordering on 13 Python files (#3408) --- autobot-backend/api/conversation_export.py | 2 +- .../llm_interface_pkg/adapters/__init__.py | 2 +- .../optimization/__init__.py | 2 +- .../optimization/pipeline.py | 4 +- .../llm_interface_pkg/optimization/router.py | 2 +- .../llm_providers/ollama_provider.py | 4 +- .../services/autoresearch/__init__.py | 40 +++++++++---------- .../knowledge_synthesizer_test.py | 2 +- .../autoresearch/meta_eval_harness_test.py | 3 +- .../autoresearch/prompt_optimizer_test.py | 2 +- .../services/autoresearch/scorers_test.py | 2 +- autobot-backend/tests/test_autoresearch_m3.py | 2 +- .../test_cache_response_json_response.py | 2 +- 13 files changed, 34 insertions(+), 35 deletions(-) diff --git a/autobot-backend/api/conversation_export.py b/autobot-backend/api/conversation_export.py index 73a8b5596..53c64e62a 100644 --- a/autobot-backend/api/conversation_export.py +++ b/autobot-backend/api/conversation_export.py @@ -30,8 +30,8 @@ export_conversation_markdown, import_conversation, ) -from utils.chat_utils import get_chat_history_manager, validate_chat_session_id from utils.chat_exceptions import get_exceptions_lazy +from utils.chat_utils import get_chat_history_manager, validate_chat_session_id logger = logging.getLogger(__name__) diff --git a/autobot-backend/llm_interface_pkg/adapters/__init__.py b/autobot-backend/llm_interface_pkg/adapters/__init__.py index c42589ce4..e30e6fb3c 100644 --- a/autobot-backend/llm_interface_pkg/adapters/__init__.py +++ b/autobot-backend/llm_interface_pkg/adapters/__init__.py @@ -10,7 +10,6 @@ from .ai_stack_adapter import AIStackAdapter from .anthropic_adapter import AnthropicAdapter -from .layer_inference_adapter import LayerInferenceAdapter from .base import ( AdapterBase, AdapterConfig, @@ -19,6 +18,7 @@ EnvironmentTestResult, SessionCodec, ) +from .layer_inference_adapter import LayerInferenceAdapter from .ollama_adapter import OllamaAdapter from .openai_adapter import OpenAIAdapter from .process_adapter import ProcessAdapter diff --git a/autobot-backend/llm_interface_pkg/optimization/__init__.py b/autobot-backend/llm_interface_pkg/optimization/__init__.py index ed066d70e..27cc19d3a 100644 --- a/autobot-backend/llm_interface_pkg/optimization/__init__.py +++ b/autobot-backend/llm_interface_pkg/optimization/__init__.py @@ -63,7 +63,6 @@ LayerInferenceEngine, LayerInferenceStats, ) -from .model_inspector import ModelInfo, clear_cache, inspect_model from .meta_eviction import ( EvictionStats, MetaDeviceEvictionManager, @@ -71,6 +70,7 @@ evict_layer_to_meta, get_gpu_memory_allocated, ) +from .model_inspector import ModelInfo, clear_cache, inspect_model from .pipeline import LayerInferencePipeline, PipelineConfig, PreparedPipeline from .profiler import INFERENCE_STAGES, LayeredProfiler from .prompt_compressor import CompressionConfig, CompressionResult, PromptCompressor diff --git a/autobot-backend/llm_interface_pkg/optimization/pipeline.py b/autobot-backend/llm_interface_pkg/optimization/pipeline.py index 11718cc24..d787b54b8 100644 --- a/autobot-backend/llm_interface_pkg/optimization/pipeline.py +++ b/autobot-backend/llm_interface_pkg/optimization/pipeline.py @@ -27,9 +27,9 @@ from .attention_backend import ( AttentionBackendSelector, - ModelConfig as AttentionModelConfig, ) -from .hf_quantizer import HfQuantizerWrapper, QuantizerConfig, QuantizationType +from .attention_backend import ModelConfig as AttentionModelConfig +from .hf_quantizer import HfQuantizerWrapper, QuantizationType, QuantizerConfig from .kv_cache import KVCacheConfig, KVCacheManager, LayerKVCache from .layer_inference import LayerInferenceConfig, LayerInferenceEngine from .meta_eviction import MetaDeviceEvictionManager, clean_memory diff --git a/autobot-backend/llm_interface_pkg/optimization/router.py b/autobot-backend/llm_interface_pkg/optimization/router.py index e0915ef92..8aa96072a 100644 --- a/autobot-backend/llm_interface_pkg/optimization/router.py +++ b/autobot-backend/llm_interface_pkg/optimization/router.py @@ -266,7 +266,7 @@ def get_quantization_kwargs( if not self.config.quantization_enabled: return {} - from .hf_quantizer import HfQuantizerWrapper, QuantizerConfig, QuantizationType + from .hf_quantizer import HfQuantizerWrapper, QuantizationType, QuantizerConfig quant_type_str = self.config.quantization_type type_map = { diff --git a/autobot-backend/llm_providers/ollama_provider.py b/autobot-backend/llm_providers/ollama_provider.py index edca4a3f2..02a919926 100644 --- a/autobot-backend/llm_providers/ollama_provider.py +++ b/autobot-backend/llm_providers/ollama_provider.py @@ -65,9 +65,7 @@ def _ensure_delegate(self): if self._delegate is not None: return self._delegate from llm_interface_pkg.models import LLMSettings - from llm_interface_pkg.providers.ollama import ( - OllamaProvider as _OllamaProvider, - ) + from llm_interface_pkg.providers.ollama import OllamaProvider as _OllamaProvider from llm_interface_pkg.streaming import StreamingManager settings = LLMSettings() diff --git a/autobot-backend/services/autoresearch/__init__.py b/autobot-backend/services/autoresearch/__init__.py index 12fbf9981..25a8a9315 100644 --- a/autobot-backend/services/autoresearch/__init__.py +++ b/autobot-backend/services/autoresearch/__init__.py @@ -11,22 +11,7 @@ Issue #2599: AutoBot-orchestrated loop + web search (M2). """ -from .knowledge_synthesizer import ExperimentInsight, KnowledgeSynthesizer -from .prompt_optimizer import ( - BenchmarkFn, - OptimizationSession, - OptimizationStatus, - PromptOptimizer, - PromptOptTarget, - PromptVariant, -) -from .scorers import ( - HumanReviewScorer, - LLMJudgeScorer, - PromptScorer, - ScorerResult, - ValBpbScorer, -) +from .archive import Archive from .auto_research_agent import ( ApprovalGate, AutoResearchAgent, @@ -37,12 +22,16 @@ SessionStatus, ) from .config import AutoResearchConfig +from .knowledge_synthesizer import ExperimentInsight, KnowledgeSynthesizer +from .meta_agent import MetaAgent, MetaPatch +from .meta_eval_harness import MetaEvalHarness, MetaEvalResult from .models import ( Experiment, ExperimentResult, ExperimentState, ExperimentStats, HyperParams, + VariantArchiveEntry, ) from .osint_engine import ( CorrelatedSignal, @@ -57,13 +46,24 @@ build_default_engine, ) from .parser import ExperimentOutputParser +from .prompt_optimizer import ( + BenchmarkFn, + OptimizationSession, + OptimizationStatus, + PromptOptimizer, + PromptOptTarget, + PromptVariant, +) from .routes import router from .runner import ExperimentRunner +from .scorers import ( + HumanReviewScorer, + LLMJudgeScorer, + PromptScorer, + ScorerResult, + ValBpbScorer, +) from .store import ExperimentStore -from .archive import Archive -from .models import VariantArchiveEntry -from .meta_agent import MetaAgent, MetaPatch -from .meta_eval_harness import MetaEvalHarness, MetaEvalResult __all__ = [ # Models diff --git a/autobot-backend/services/autoresearch/knowledge_synthesizer_test.py b/autobot-backend/services/autoresearch/knowledge_synthesizer_test.py index a6791f951..f4c880cd4 100644 --- a/autobot-backend/services/autoresearch/knowledge_synthesizer_test.py +++ b/autobot-backend/services/autoresearch/knowledge_synthesizer_test.py @@ -6,9 +6,9 @@ from __future__ import annotations import json +from unittest.mock import AsyncMock, MagicMock, patch import pytest -from unittest.mock import AsyncMock, MagicMock, patch from services.autoresearch.knowledge_synthesizer import ( ExperimentInsight, diff --git a/autobot-backend/services/autoresearch/meta_eval_harness_test.py b/autobot-backend/services/autoresearch/meta_eval_harness_test.py index f102d5240..e4e2a3785 100644 --- a/autobot-backend/services/autoresearch/meta_eval_harness_test.py +++ b/autobot-backend/services/autoresearch/meta_eval_harness_test.py @@ -7,7 +7,8 @@ from __future__ import annotations -from unittest.mock import AsyncMock, MagicMock, patch as mock_patch +from unittest.mock import AsyncMock, MagicMock +from unittest.mock import patch as mock_patch import pytest diff --git a/autobot-backend/services/autoresearch/prompt_optimizer_test.py b/autobot-backend/services/autoresearch/prompt_optimizer_test.py index c093501e0..82d6538c9 100644 --- a/autobot-backend/services/autoresearch/prompt_optimizer_test.py +++ b/autobot-backend/services/autoresearch/prompt_optimizer_test.py @@ -6,9 +6,9 @@ from __future__ import annotations import json +from unittest.mock import AsyncMock, MagicMock import pytest -from unittest.mock import AsyncMock, MagicMock from services.autoresearch.archive import Archive from services.autoresearch.config import AutoResearchConfig diff --git a/autobot-backend/services/autoresearch/scorers_test.py b/autobot-backend/services/autoresearch/scorers_test.py index 27fee6157..21dce2f24 100644 --- a/autobot-backend/services/autoresearch/scorers_test.py +++ b/autobot-backend/services/autoresearch/scorers_test.py @@ -6,9 +6,9 @@ from __future__ import annotations import json +from unittest.mock import AsyncMock, MagicMock import pytest -from unittest.mock import AsyncMock, MagicMock from services.autoresearch.models import Experiment, ExperimentResult, ExperimentState from services.autoresearch.scorers import ( diff --git a/autobot-backend/tests/test_autoresearch_m3.py b/autobot-backend/tests/test_autoresearch_m3.py index 664cd96e9..ecf588993 100644 --- a/autobot-backend/tests/test_autoresearch_m3.py +++ b/autobot-backend/tests/test_autoresearch_m3.py @@ -6,9 +6,9 @@ from __future__ import annotations import json +from unittest.mock import AsyncMock, MagicMock import pytest -from unittest.mock import AsyncMock, MagicMock from services.autoresearch.knowledge_synthesizer import KnowledgeSynthesizer from services.autoresearch.models import ( diff --git a/autobot-backend/tests/utils/test_cache_response_json_response.py b/autobot-backend/tests/utils/test_cache_response_json_response.py index d3b658290..945baa818 100644 --- a/autobot-backend/tests/utils/test_cache_response_json_response.py +++ b/autobot-backend/tests/utils/test_cache_response_json_response.py @@ -21,12 +21,12 @@ from utils.advanced_cache_manager import ( _JSON_RESPONSE_ENVELOPE, + SimpleCacheManager, _deserialise_cached_entry, _record_cache_hit, _record_cache_miss, _serialise_response, ) -from utils.advanced_cache_manager import SimpleCacheManager # --------------------------------------------------------------------------- # _serialise_response