Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
13 changes: 5 additions & 8 deletions src/memos/mem_scheduler/base_scheduler.py
Original file line number Diff line number Diff line change
Expand Up @@ -842,12 +842,7 @@ def _submit_web_logs(
messages = [messages] # transform single message to list

for message in messages:
logger.info(
f"[DIAGNOSTIC] base_scheduler._submit_web_logs called. Message to publish: {message.model_dump_json(indent=2)}"
)

try:
for message in messages:
try:
# Always call publish; the publisher now caches when offline and flushes after reconnect
logger.info(
f"[DIAGNOSTIC] base_scheduler._submit_web_logs: enqueue publish {message.model_dump_json(indent=2)}"
Expand All @@ -860,8 +855,10 @@ def _submit_web_logs(
message.task_id,
message.label,
)
except Exception as e:
logger.error(f"[DIAGNOSTIC] base_scheduler._submit_web_logs failed: {e}", exc_info=True)
except Exception as e:
logger.error(
f"[DIAGNOSTIC] base_scheduler._submit_web_logs failed: {e}", exc_info=True
)

logger.debug(
f"{len(messages)} submitted. {self._web_log_message_queue.qsize()} in queue. additional_log_info: {additional_log_info}"
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,8 @@
from pathlib import Path
from queue import Empty

from pyglet.libs.win32.constants import FALSE

from memos.configs.mem_scheduler import AuthConfig, RabbitMQConfig
from memos.context.context import ContextThread
from memos.dependency import require_python_package
Expand Down Expand Up @@ -325,14 +327,14 @@ def rabbitmq_publish_message(self, message: dict):
f"[DIAGNOSTIC] Publishing {label} message in Cloud Env. "
f"Exchange: {exchange_name}, Routing Key: '{routing_key}'."
)
logger.info(f" - Message Content: {json.dumps(message, indent=2)}")
logger.info(f" - Message Content: {json.dumps(message, indent=2, ensure_ascii=FALSE)}")
elif label == "knowledgeBaseUpdate":
# Original diagnostic logging for knowledgeBaseUpdate if NOT in cloud env
logger.info(
f"[DIAGNOSTIC] Publishing knowledgeBaseUpdate message (Local Env). "
f"Current configured Exchange: {exchange_name}, Routing Key: '{routing_key}'."
)
logger.info(f" - Message Content: {json.dumps(message, indent=2)}")
logger.info(f" - Message Content: {json.dumps(message, indent=2, ensure_ascii=FALSE)}")

with self._rabbitmq_lock:
logger.info(
Expand Down
Loading