Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion autobot-backend/api/conversation_export.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,8 +30,8 @@
export_conversation_markdown,
import_conversation,
)
from utils.chat_utils import get_chat_history_manager, validate_chat_session_id
from utils.chat_exceptions import get_exceptions_lazy
from utils.chat_utils import get_chat_history_manager, validate_chat_session_id

logger = logging.getLogger(__name__)

Expand Down
2 changes: 1 addition & 1 deletion autobot-backend/llm_interface_pkg/adapters/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,6 @@

from .ai_stack_adapter import AIStackAdapter
from .anthropic_adapter import AnthropicAdapter
from .layer_inference_adapter import LayerInferenceAdapter
from .base import (
AdapterBase,
AdapterConfig,
Expand All @@ -19,6 +18,7 @@
EnvironmentTestResult,
SessionCodec,
)
from .layer_inference_adapter import LayerInferenceAdapter
from .ollama_adapter import OllamaAdapter
from .openai_adapter import OpenAIAdapter
from .process_adapter import ProcessAdapter
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -63,14 +63,14 @@
LayerInferenceEngine,
LayerInferenceStats,
)
from .model_inspector import ModelInfo, clear_cache, inspect_model
from .meta_eviction import (
EvictionStats,
MetaDeviceEvictionManager,
clean_memory,
evict_layer_to_meta,
get_gpu_memory_allocated,
)
from .model_inspector import ModelInfo, clear_cache, inspect_model
from .pipeline import LayerInferencePipeline, PipelineConfig, PreparedPipeline
from .profiler import INFERENCE_STAGES, LayeredProfiler
from .prompt_compressor import CompressionConfig, CompressionResult, PromptCompressor
Expand Down
4 changes: 2 additions & 2 deletions autobot-backend/llm_interface_pkg/optimization/pipeline.py
Original file line number Diff line number Diff line change
Expand Up @@ -27,9 +27,9 @@

from .attention_backend import (
AttentionBackendSelector,
ModelConfig as AttentionModelConfig,
)
from .hf_quantizer import HfQuantizerWrapper, QuantizerConfig, QuantizationType
from .attention_backend import ModelConfig as AttentionModelConfig
from .hf_quantizer import HfQuantizerWrapper, QuantizationType, QuantizerConfig
from .kv_cache import KVCacheConfig, KVCacheManager, LayerKVCache
from .layer_inference import LayerInferenceConfig, LayerInferenceEngine
from .meta_eviction import MetaDeviceEvictionManager, clean_memory
Expand Down
2 changes: 1 addition & 1 deletion autobot-backend/llm_interface_pkg/optimization/router.py
Original file line number Diff line number Diff line change
Expand Up @@ -266,7 +266,7 @@ def get_quantization_kwargs(
if not self.config.quantization_enabled:
return {}

from .hf_quantizer import HfQuantizerWrapper, QuantizerConfig, QuantizationType
from .hf_quantizer import HfQuantizerWrapper, QuantizationType, QuantizerConfig

quant_type_str = self.config.quantization_type
type_map = {
Expand Down
4 changes: 1 addition & 3 deletions autobot-backend/llm_providers/ollama_provider.py
Original file line number Diff line number Diff line change
Expand Up @@ -65,9 +65,7 @@ def _ensure_delegate(self):
if self._delegate is not None:
return self._delegate
from llm_interface_pkg.models import LLMSettings
from llm_interface_pkg.providers.ollama import (
OllamaProvider as _OllamaProvider,
)
from llm_interface_pkg.providers.ollama import OllamaProvider as _OllamaProvider
from llm_interface_pkg.streaming import StreamingManager

settings = LLMSettings()
Expand Down
40 changes: 20 additions & 20 deletions autobot-backend/services/autoresearch/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,22 +11,7 @@
Issue #2599: AutoBot-orchestrated loop + web search (M2).
"""

from .knowledge_synthesizer import ExperimentInsight, KnowledgeSynthesizer
from .prompt_optimizer import (
BenchmarkFn,
OptimizationSession,
OptimizationStatus,
PromptOptimizer,
PromptOptTarget,
PromptVariant,
)
from .scorers import (
HumanReviewScorer,
LLMJudgeScorer,
PromptScorer,
ScorerResult,
ValBpbScorer,
)
from .archive import Archive
from .auto_research_agent import (
ApprovalGate,
AutoResearchAgent,
Expand All @@ -37,12 +22,16 @@
SessionStatus,
)
from .config import AutoResearchConfig
from .knowledge_synthesizer import ExperimentInsight, KnowledgeSynthesizer
from .meta_agent import MetaAgent, MetaPatch
from .meta_eval_harness import MetaEvalHarness, MetaEvalResult
from .models import (
Experiment,
ExperimentResult,
ExperimentState,
ExperimentStats,
HyperParams,
VariantArchiveEntry,
)
from .osint_engine import (
CorrelatedSignal,
Expand All @@ -57,13 +46,24 @@
build_default_engine,
)
from .parser import ExperimentOutputParser
from .prompt_optimizer import (
BenchmarkFn,
OptimizationSession,
OptimizationStatus,
PromptOptimizer,
PromptOptTarget,
PromptVariant,
)
from .routes import router
from .runner import ExperimentRunner
from .scorers import (
HumanReviewScorer,
LLMJudgeScorer,
PromptScorer,
ScorerResult,
ValBpbScorer,
)
from .store import ExperimentStore
from .archive import Archive
from .models import VariantArchiveEntry
from .meta_agent import MetaAgent, MetaPatch
from .meta_eval_harness import MetaEvalHarness, MetaEvalResult

__all__ = [
# Models
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -6,9 +6,9 @@
from __future__ import annotations

import json
from unittest.mock import AsyncMock, MagicMock, patch

import pytest
from unittest.mock import AsyncMock, MagicMock, patch

from services.autoresearch.knowledge_synthesizer import (
ExperimentInsight,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,8 @@

from __future__ import annotations

from unittest.mock import AsyncMock, MagicMock, patch as mock_patch
from unittest.mock import AsyncMock, MagicMock
from unittest.mock import patch as mock_patch

import pytest

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -6,9 +6,9 @@
from __future__ import annotations

import json
from unittest.mock import AsyncMock, MagicMock

import pytest
from unittest.mock import AsyncMock, MagicMock

from services.autoresearch.archive import Archive
from services.autoresearch.config import AutoResearchConfig
Expand Down
2 changes: 1 addition & 1 deletion autobot-backend/services/autoresearch/scorers_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,9 +6,9 @@
from __future__ import annotations

import json
from unittest.mock import AsyncMock, MagicMock

import pytest
from unittest.mock import AsyncMock, MagicMock

from services.autoresearch.models import Experiment, ExperimentResult, ExperimentState
from services.autoresearch.scorers import (
Expand Down
2 changes: 1 addition & 1 deletion autobot-backend/tests/test_autoresearch_m3.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,9 +6,9 @@
from __future__ import annotations

import json
from unittest.mock import AsyncMock, MagicMock

import pytest
from unittest.mock import AsyncMock, MagicMock

from services.autoresearch.knowledge_synthesizer import KnowledgeSynthesizer
from services.autoresearch.models import (
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,12 +21,12 @@

from utils.advanced_cache_manager import (
_JSON_RESPONSE_ENVELOPE,
SimpleCacheManager,
_deserialise_cached_entry,
_record_cache_hit,
_record_cache_miss,
_serialise_response,
)
from utils.advanced_cache_manager import SimpleCacheManager

# ---------------------------------------------------------------------------
# _serialise_response
Expand Down
Loading