Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
37 changes: 16 additions & 21 deletions src/memos/mem_scheduler/base_scheduler.py
Original file line number Diff line number Diff line change
Expand Up @@ -846,28 +846,23 @@ def _submit_web_logs(
f"[DIAGNOSTIC] base_scheduler._submit_web_logs called. Message to publish: {message.model_dump_json(indent=2)}"
)

if self.rabbitmq_config is None:
logger.info(
"[DIAGNOSTIC] base_scheduler._submit_web_logs: RabbitMQ config not loaded; skipping publish."
)
return

for message in messages:
message_info = message.debug_info()
logger.info(f"[DIAGNOSTIC] base_scheduler._submit_web_logs: submitted {message_info}")
try:
for message in messages:
# Always call publish; the publisher now caches when offline and flushes after reconnect
logger.info(
f"[DIAGNOSTIC] base_scheduler._submit_web_logs: enqueue publish {message.model_dump_json(indent=2)}"
)
self.rabbitmq_publish_message(message=message.to_dict())
logger.info(
"[DIAGNOSTIC] base_scheduler._submit_web_logs: publish dispatched "
"item_id=%s task_id=%s label=%s",
message.item_id,
message.task_id,
message.label,
)
except Exception as e:
logger.error(f"[DIAGNOSTIC] base_scheduler._submit_web_logs failed: {e}", exc_info=True)

# Always call publish; the publisher now caches when offline and flushes after reconnect
logger.info(
f"[DIAGNOSTIC] base_scheduler._submit_web_logs: enqueue publish {message_info}"
)
self.rabbitmq_publish_message(message=message.to_dict())
logger.info(
"[DIAGNOSTIC] base_scheduler._submit_web_logs: publish dispatched "
"item_id=%s task_id=%s label=%s",
message.item_id,
message.task_id,
message.label,
)
logger.debug(
f"{len(messages)} submitted. {self._web_log_message_queue.qsize()} in queue. additional_log_info: {additional_log_info}"
)
Expand Down