feat(memory): add session-based chat history and user metadata retrieval

- Add ChatSessionCache to manage chat history per session
- Add SEARCH_USER_METADATA cypher query for retrieving user entity metadata
- Add "str" mode support to StructResponse for raw text extraction
- Add content_str field to MemorySearchResult for pre-formatted content
- Fix sandbox URL by removing hardcoded port
- Add description field to entity search results
- Remove history from UserInput schema, use session_id instead
This commit is contained in:
Eternity
2026-05-06 16:24:53 +08:00
parent 6b68ee9fc8
commit 3f9740412a
19 changed files with 387 additions and 117 deletions

View File

@@ -158,12 +158,19 @@ class RedisTaskScheduler:
return {"status": status, "task_id": task_id, "result": result_content}
def _cleanup_finished(self):
pending = self.redis.hgetall(PENDING_HASH)
if not pending:
cursor = 0
all_pending = {}
while True:
cursor, batch = self.redis.hscan(PENDING_HASH, cursor=cursor, count=100)
all_pending.update(batch)
if cursor == 0:
break
if not all_pending:
return
now = time.time()
task_ids = list(pending.keys())
task_ids = list(all_pending.keys())
pipe = self.redis.pipeline()
for task_id in task_ids:
@@ -176,7 +183,7 @@ class RedisTaskScheduler:
for task_id, raw_result in zip(task_ids, results):
try:
meta = json.loads(pending[task_id])
meta = json.loads(all_pending[task_id])
lock_key = meta["lock_key"]
dispatched_at = meta.get("dispatched_at", 0)
age = now - dispatched_at
@@ -276,6 +283,22 @@ class RedisTaskScheduler:
return True
return stable_hash(user_id) % self._shard_count == self._shard_index
def _commit_post_dispatch(self, lock_key, task, msg_id, dispatch_lock):
pipe = self.redis.pipeline()
pipe.set(lock_key, task.id, ex=3600)
pipe.hset(PENDING_HASH, task.id, json.dumps({
"lock_key": lock_key,
"dispatched_at": time.time(),
"msg_id": msg_id,
}))
pipe.delete(dispatch_lock)
pipe.set(
f"task_tracker:{msg_id}",
json.dumps({"status": "DISPATCHED", "task_id": task.id}),
ex=86400,
)
pipe.execute()
def _dispatch(self, msg_id, msg_data) -> bool:
user_id = msg_data["user_id"]
task_name = msg_data["task_name"]
@@ -308,27 +331,16 @@ class RedisTaskScheduler:
task_name, user_id, msg_id, e, exc_info=True,
)
return False
for attempt in range(2):
try:
pipe = self.redis.pipeline()
pipe.set(lock_key, task.id, ex=3600)
pipe.hset(PENDING_HASH, task.id, json.dumps({
"lock_key": lock_key,
"dispatched_at": time.time(),
"msg_id": msg_id,
}))
pipe.delete(dispatch_lock)
pipe.set(
f"task_tracker:{msg_id}",
json.dumps({"status": "DISPATCHED", "task_id": task.id}),
ex=86400,
)
pipe.execute()
self._commit_post_dispatch(lock_key, task, msg_id, dispatch_lock)
break
except Exception as e:
logger.error(
"Post-dispatch state update failed for %s: %s",
task.id, e, exc_info=True,
)
time.sleep(0.1)
self.errors += 1
self.dispatched += 1
@@ -367,22 +379,21 @@ class RedisTaskScheduler:
return
for uid, msg in candidates:
queue_key = f"{USER_QUEUE_PREFIX}{uid}"
if self._dispatch(msg["msg_id"], msg):
self.redis.lpop(f"{USER_QUEUE_PREFIX}{uid}")
self.redis.lpop(queue_key)
if self.redis.llen(queue_key) > 0:
self.redis.sadd(READY_SET, uid)
def schedule_loop(self):
self._heartbeat()
self._cleanup_finished()
pipe = self.redis.pipeline()
pipe.smembers(READY_SET)
pipe.delete(READY_SET)
results = pipe.execute()
ready_users = results[0] or set()
ready_users = self.redis.smembers(READY_SET) or set()
my_users = [uid for uid in ready_users if self._is_mine(uid)]
if not my_users:
if my_users:
self.redis.srem(READY_SET, *my_users)
else:
time.sleep(0.5)
return
@@ -445,7 +456,7 @@ class RedisTaskScheduler:
"Scheduler started: instance=%s", self.instance_id,
)
while True:
while self.running:
try:
self.schedule_loop()
@@ -480,9 +491,7 @@ class RedisTaskScheduler:
logger.error("Shutdown cleanup error: %s", e)
scheduler: RedisTaskScheduler | None = None
if scheduler is None:
scheduler = RedisTaskScheduler()
scheduler = RedisTaskScheduler()
if __name__ == "__main__":
import signal

View File

@@ -27,6 +27,7 @@ from app.services import task_service, workspace_service
from app.services.memory_agent_service import MemoryAgentService
from app.services.memory_agent_service import get_end_user_connected_config as get_config
from app.services.model_service import ModelConfigService
from app.utils.tmp_session import ChatSessionCache
load_dotenv()
api_logger = get_api_logger()
@@ -300,60 +301,39 @@ async def read_server(
if knowledge:
user_rag_memory_id = str(knowledge.id)
session_id = user_input.session_id.hex
api_logger.info(
f"Read service: group={user_input.end_user_id}, storage_type={storage_type}, user_rag_memory_id={user_rag_memory_id}, workspace_id={workspace_id}")
f"Read service: group={user_input.end_user_id}, storage_type={storage_type}, user_rag_memory_id={user_rag_memory_id}, workspace_id={workspace_id}, session_id={session_id}")
try:
# result = await memory_agent_service.read_memory(
# user_input.end_user_id,
# user_input.message,
# user_input.history,
# user_input.search_switch,
# config_id,
# db,
# storage_type,
# user_rag_memory_id
# )
# if str(user_input.search_switch) == "2":
# retrieve_info = result['answer']
# history = await SessionService(store).get_history(user_input.end_user_id, user_input.end_user_id,
# user_input.end_user_id)
# query = user_input.message
#
# # 调用 memory_agent_service 的方法生成最终答案
# result['answer'] = await memory_agent_service.generate_summary_from_retrieve(
# end_user_id=user_input.end_user_id,
# retrieve_info=retrieve_info,
# history=history,
# query=query,
# config_id=config_id,
# db=db
# )
# if "信息不足,无法回答" in result['answer']:
# result['answer'] = retrieve_info
memory_config = get_config(user_input.end_user_id, db)
service = MemoryService(
db,
memory_config["memory_config_id"],
end_user_id=user_input.end_user_id
)
session_cache = ChatSessionCache(session_id)
search_result = await service.read(
user_input.message,
SearchStrategy(user_input.search_switch)
SearchStrategy(user_input.search_switch),
history=await session_cache.get_history(),
)
intermediate_outputs = []
sub_queries = set()
for memory in search_result.memories:
sub_queries.add(str(memory.query))
idx = 0
if user_input.search_switch in [SearchStrategy.DEEP, SearchStrategy.NORMAL]:
intermediate_outputs.append({
"type": "problem_split",
"title": "问题拆分",
"data": [
{
"id": f"Q{idx+1}",
"id": f"Q{(idx := idx + 1)}",
"question": question
}
for idx, question in enumerate(sub_queries)
for question in sub_queries
if question
]
})
perceptual_data = [
@@ -375,16 +355,24 @@ async def read_server(
"raw_result": search_result.memories,
"total": len(search_result.memories),
})
result = {
'answer': await memory_agent_service.generate_summary_from_retrieve(
answer = await memory_agent_service.generate_summary_from_retrieve(
end_user_id=user_input.end_user_id,
retrieve_info=search_result.content,
history=[],
query=user_input.message,
config_id=config_id,
db=db
),
"intermediate_outputs": intermediate_outputs
)
await session_cache.append_many(
[
{"role": "user", "content": user_input.message},
{"role": "assistant", "content": answer}
]
)
result = {
'answer': answer,
"intermediate_outputs": intermediate_outputs,
"session_id": session_id,
}
return success(data=result, msg="回复对话消息成功")
@@ -480,9 +468,11 @@ async def read_server_async(
if knowledge: user_rag_memory_id = str(knowledge.id)
api_logger.info(f"Async read: storage_type={storage_type}, user_rag_memory_id={user_rag_memory_id}")
try:
session_id = user_input.session_id.hex
session_cache = ChatSessionCache(session_id)
task = celery_app.send_task(
"app.core.memory.agent.read_message",
args=[user_input.end_user_id, user_input.message, user_input.history, user_input.search_switch,
args=[user_input.end_user_id, user_input.message, await session_cache.get_history(), user_input.search_switch,
config_id, storage_type, user_rag_memory_id]
)
api_logger.info(f"Read task queued: {task.id}")

View File

@@ -43,10 +43,13 @@ class MemoryService:
self,
query: str,
search_switch: SearchStrategy,
history: list | None = None,
limit: int = 10,
) -> MemorySearchResult:
if history is None:
history = []
with get_db_context() as db:
return await ReadPipeLine(self.ctx, db).run(query, search_switch, limit)
return await ReadPipeLine(self.ctx, db).run(query, search_switch, history, limit)
async def forget(self, max_batch: int = 100, min_days: int = 30) -> dict:
raise NotImplementedError

View File

@@ -32,10 +32,12 @@ class Memory(BaseModel):
class MemorySearchResult(BaseModel):
memories: list[Memory]
content_str: str = Field(default="")
@computed_field
@property
def content(self) -> str:
if self.content_str:
return self.content_str
return "\n".join([memory.content for memory in self.memories])
@computed_field

View File

@@ -1,8 +1,9 @@
from app.core.memory.enums import SearchStrategy, StorageType
from app.core.memory.models.service_models import MemorySearchResult
from app.core.memory.pipelines.base_pipeline import ModelClientMixin, DBRequiredPipeline
from app.core.memory.read_services.search_engine.content_search import Neo4jSearchService, RAGSearchService
from app.core.memory.read_services.generate_engine.query_preprocessor import QueryPreprocessor
from app.core.memory.read_services.generate_engine.retrieval_summary import RetrievalSummaryProcessor
from app.core.memory.read_services.search_engine.content_search import Neo4jSearchService, RAGSearchService
class ReadPipeLine(ModelClientMixin, DBRequiredPipeline):
@@ -10,20 +11,30 @@ class ReadPipeLine(ModelClientMixin, DBRequiredPipeline):
self,
query: str,
search_switch: SearchStrategy,
history: list,
limit: int = 10,
includes=None
) -> MemorySearchResult:
memory_l0 = None
if self.ctx.storage_type == StorageType.NEO4J:
memory_l0 = await self._get_search_service(includes).memory_l0()
query = QueryPreprocessor.process(query)
match search_switch:
case SearchStrategy.DEEP:
return await self._deep_read(query, limit, includes)
res = await self._deep_read(query, history, limit, includes)
case SearchStrategy.NORMAL:
return await self._normal_read(query, limit, includes)
res = await self._normal_read(query, history, limit, includes)
case SearchStrategy.QUICK:
return await self._quick_read(query, limit, includes)
res = await self._quick_read(query, limit, includes)
case _:
raise RuntimeError("Unsupported search strategy")
if memory_l0 is not None:
res.content_str = memory_l0.content + '\n' + res.content
res.memories.insert(0, memory_l0)
return res
def _get_search_service(self, includes=None):
if self.ctx.storage_type == StorageType.NEO4J:
return Neo4jSearchService(
@@ -37,10 +48,11 @@ class ReadPipeLine(ModelClientMixin, DBRequiredPipeline):
self.db
)
async def _deep_read(self, query: str, limit: int, includes=None) -> MemorySearchResult:
async def _deep_read(self, query: str, history: list, limit: int, includes=None) -> MemorySearchResult:
search_service = self._get_search_service(includes)
questions = await QueryPreprocessor.split(
query,
history,
self.get_llm_client(self.db, self.ctx.memory_config.llm_model_id)
)
query_results = []
@@ -49,12 +61,18 @@ class ReadPipeLine(ModelClientMixin, DBRequiredPipeline):
query_results.append(search_results)
results = sum(query_results, start=MemorySearchResult(memories=[]))
results.memories.sort(key=lambda x: x.score, reverse=True)
results.content_str = await RetrievalSummaryProcessor.summary(
query,
results.content,
self.get_llm_client(self.db, self.ctx.memory_config.llm_model_id)
)
return results
async def _normal_read(self, query: str, limit: int, includes=None) -> MemorySearchResult:
async def _normal_read(self, query: str, history: list, limit: int, includes=None) -> MemorySearchResult:
search_service = self._get_search_service(includes)
questions = await QueryPreprocessor.split(
query,
history,
self.get_llm_client(self.db, self.ctx.memory_config.llm_model_id)
)
query_results = []
@@ -63,6 +81,11 @@ class ReadPipeLine(ModelClientMixin, DBRequiredPipeline):
query_results.append(search_results)
results = sum(query_results, start=MemorySearchResult(memories=[]))
results.memories.sort(key=lambda x: x.score, reverse=True)
results.content_str = await RetrievalSummaryProcessor.summary(
query,
results.content,
self.get_llm_client(self.db, self.ctx.memory_config.llm_model_id)
)
return results
async def _quick_read(self, query: str, limit: int, includes=None) -> MemorySearchResult:

View File

@@ -0,0 +1,15 @@
You are a Content Condenser for a memory-augmented retrieval system.
Your task is to compress the retrieved content while preserving all information that is highly relevant to the users query.
Guidelines:
Focus only on content related to the query; ignore irrelevant parts.
Remove redundancy, filler, or repeated information only for non-XML content.
Preserve all factual details: names, dates, decisions, code snippets, technical details.
If relevant information is inside XML tags, do not remove, merge, or compress the XML tags or their internal text; keep them fully intact.
Structure multiple relevant points as a compact bullet list or paragraph, depending on density.
If no content is relevant, return exactly: "No relevant information found."
Do not add any knowledge or facts not in the retrieved content.
# [IMPORTANT] OUTPUT ONLY THE CONDENSED CONTENT, DO NOT ATTEMPT TO ANSWER THE QUERY.
# [IMPORTANT] DO NOT REMOVE OR PARAPHRASE HIGHLY RELEVANT INFORMATION.

View File

@@ -21,14 +21,14 @@ class QueryPreprocessor:
return text
@staticmethod
async def split(query: str, llm_client: RedBearLLM):
async def split(query: str, history: list, llm_client: RedBearLLM):
system_prompt = prompt_manager.render(
name="problem_split",
datetime=datetime.now().strftime("%Y-%m-%d"),
)
messages = [
{"role": "system", "content": system_prompt},
{"role": "user", "content": query},
{"role": "user", "content": f"<history>{history}</history><query>{query}</query>"},
]
try:
sub_queries = await llm_client.ainvoke(messages) | StructResponse(mode='json')

View File

@@ -1,11 +1,29 @@
import logging
from app.core.models import RedBearLLM
from app.core.memory.prompt import prompt_manager
from app.core.memory.utils.llm.llm_utils import StructResponse
logger = logging.getLogger(__name__)
class RetrievalSummaryProcessor:
@staticmethod
def summary(content: str, llm_client: RedBearLLM):
return
async def summary(query, content: str, llm_client: RedBearLLM):
system_prompt = prompt_manager.render(
name="retrieval_summary"
)
messages = [
{"role": "system", "content": system_prompt},
{"role": "user", "content": f"<query>{query}</query><content>{content}</content>"},
]
try:
summary = await llm_client.ainvoke(messages) | StructResponse(mode='str')
return summary
except:
logger.error("Failed to generate reply summary, returning original content", exc_info=True)
return content
@staticmethod
def verify(content: str, llm_client: RedBearLLM):
async def verify(query, content: str, llm_client: RedBearLLM):
return

View File

@@ -14,6 +14,8 @@ from app.core.rag.nlp.search import knowledge_retrieval
from app.repositories import knowledge_repository
from app.repositories.neo4j.graph_search import search_graph, search_graph_by_embedding
from app.repositories.neo4j.neo4j_connector import Neo4jConnector
from app.core.memory.read_services.search_engine.result_builder import MetadataBuilder
from app.repositories.neo4j.graph_search import search_user_metadata
logger = logging.getLogger(__name__)
@@ -177,6 +179,22 @@ class Neo4jSearchService:
memories.sort(key=lambda x: x.score, reverse=True)
return MemorySearchResult(memories=memories[:limit])
async def memory_l0(self) -> Memory:
async with Neo4jConnector() as connector:
end_user_id = self.ctx.end_user_id
user_meta = await search_user_metadata(connector, end_user_id)
metadata = MetadataBuilder(user_meta)
memory = Memory(
score=1,
source=Neo4jNodeType.EXTRACTEDENTITY,
query='',
id=end_user_id,
content=metadata.content,
data=metadata.data,
)
return memory
class RAGSearchService:
def __init__(self, ctx: MemoryContext, db: Session):

View File

@@ -42,7 +42,15 @@ class ChunkBuilder(BaseBuilder):
@property
def content(self) -> str:
return self.record.get("content")
parts = ["<chunk>"]
fields = [
("content", self.record.get("content", "")),
]
for tag, value in fields:
if value:
parts.append(f"<{tag}>{value}</{tag}>")
parts.append("</chunk>")
return "".join(parts)
class StatementBuiler(BaseBuilder):
@@ -57,7 +65,15 @@ class StatementBuiler(BaseBuilder):
@property
def content(self) -> str:
return self.record.get("statement")
parts = ["<statement>"]
fields = [
("statement", self.record.get("statement", "")),
]
for tag, value in fields:
if value:
parts.append(f"<{tag}>{value}</{tag}>")
parts.append("</statement>")
return "".join(parts)
class EntityBuilder(BaseBuilder):
@@ -73,10 +89,16 @@ class EntityBuilder(BaseBuilder):
@property
def content(self) -> str:
return (f"<entity>"
f"<name>{self.record.get("name")}<name>"
f"<description>{self.record.get("description")}</description>"
f"</entity>")
parts = ["<entity>"]
fields = [
("name", self.record.get("name", "")),
("description", self.record.get("description", "")),
]
for tag, value in fields:
if value:
parts.append(f"<{tag}>{value}</{tag}>")
parts.append("</entity>")
return "".join(parts)
class SummaryBuilder(BaseBuilder):
@@ -91,7 +113,15 @@ class SummaryBuilder(BaseBuilder):
@property
def content(self) -> str:
return self.record.get("content")
parts = ["<summary>"]
fields = [
("content", self.record.get("content", "")),
]
for tag, value in fields:
if value:
parts.append(f"<{tag}>{value}</{tag}>")
parts.append("</summary>")
return "".join(parts)
class PerceptualBuilder(BaseBuilder):
@@ -114,15 +144,21 @@ class PerceptualBuilder(BaseBuilder):
@property
def content(self) -> str:
return ("<history-file-info>"
f"<file-name>{self.record.get('file_name')}</file-name>"
f"<file-path>{self.record.get('file_path')}</file-path>"
f"<summary>{self.record.get('summary')}</summary>"
f"<topic>{self.record.get('topic')}</topic>"
f"<domain>{self.record.get('domain')}</domain>"
f"<keywords>{self.record.get('keywords')}</keywords>"
f"<file-type>{self.record.get('file_type')}</file-type>"
"</history-file-info>")
parts = ["<history-file-info>"]
fields = [
("file-name", self.record.get("file_name", "")),
("file-path", self.record.get("file_path", "")),
("summary", self.record.get("summary", "")),
("topic", self.record.get("topic", "")),
("domain", self.record.get("domain", "")),
("keywords", self.record.get("keywords", [])),
("file-type", self.record.get("file_type", "")),
]
for tag, value in fields:
if value:
parts.append(f"<{tag}>{value}</{tag}>")
parts.append("</history-file-info>")
return "".join(parts)
class CommunityBuilder(BaseBuilder):
@@ -137,7 +173,54 @@ class CommunityBuilder(BaseBuilder):
@property
def content(self) -> str:
return self.record.get("content")
parts = ["<community>"]
fields = [
("content", self.record.get("content", "")),
]
for tag, value in fields:
if value:
parts.append(f"<{tag}>{value}</{tag}>")
parts.append("</community>")
return "".join(parts)
class MetadataBuilder(BaseBuilder):
@property
def data(self) -> dict:
return {
"id": self.record.get("id", ""),
"aliases_name": self.record.get("aliases", []) or [],
"description": self.record.get("description", ""),
"anchors": self.record.get("anchors", []) or [],
"beliefs_or_stances": self.record.get("beliefs_or_stances", []) or [],
"core_facts": self.record.get("core_facts", []) or [],
"events": self.record.get("events", []) or [],
"goals": self.record.get("goals", []) or [],
"interests": self.record.get("interests", []) or [],
"relations": self.record.get("relations", []) or [],
"traits": self.record.get("traits", []) or [],
}
@property
def content(self) -> str:
parts = ["<user-info>"]
fields = [
("description", self.record.get("description", "")),
("aliases", self.record.get("aliases", [])),
("anchors", self.record.get("anchors", [])),
("beliefs_or_stances", self.record.get("beliefs_or_stances", [])),
("core_facts", self.record.get("core_facts", [])),
("events", self.record.get("events", [])),
("goals", self.record.get("goals", [])),
("interests", self.record.get("interests", [])),
("relations", self.record.get("relations", [])),
("traits", self.record.get("traits", [])),
]
for tag, value in fields:
if value:
parts.append(f"<{tag}>{value}</{tag}>")
parts.append("</user-info>")
return "".join(parts)
def data_builder_factory(node_type, data: dict) -> T:

View File

@@ -17,7 +17,7 @@ async def handle_response(response: type[BaseModel]) -> dict:
class StructResponse:
def __init__(self, mode: Literal["json", "pydantic"], model: Type[BaseModel] = None):
def __init__(self, mode: Literal["json", "pydantic", "str"], model: Type[BaseModel] = None):
self.mode = mode
if mode == "pydantic" and model is None:
raise ValueError("Pydantic model is required")
@@ -31,6 +31,8 @@ class StructResponse:
for block in other.content_blocks:
if block.get("type") == "text":
text += block.get("text", "")
if self.mode == "str":
return text
fixed_json = json_repair.repair_json(text, return_objects=True)
if self.mode == "json":
return fixed_json

View File

@@ -132,7 +132,7 @@ class CodeNode(BaseNode):
async with httpx.AsyncClient(timeout=60) as client:
response = await client.post(
f"{settings.SANDBOX_URL}:8194/v1/sandbox/run",
f"{settings.SANDBOX_URL}/v1/sandbox/run",
headers={
"x-api-key": 'redbear-sandbox'
},

View File

@@ -40,6 +40,7 @@ class MemoryReadNode(BaseNode):
end_user_id=end_user_id,
user_rag_memory_id=state["user_rag_memory_id"],
)
# TODO: Historical Messages -> Used to refer to coreference resolution
search_result = await memory_service.read(
self._render_template(self.typed_config.message, variable_pool),
search_switch=SearchStrategy(self.typed_config.search_switch)

View File

@@ -1296,6 +1296,7 @@ RETURN e.id AS id,
e.name AS name,
e.end_user_id AS end_user_id,
e.entity_type AS entity_type,
e.description AS description,
COALESCE(e.activation_value, e.importance_score, 0.5) AS activation_value,
COALESCE(e.importance_score, 0.5) AS importance_score,
e.last_access_time AS last_access_time,
@@ -1479,6 +1480,21 @@ ORDER BY score DESC
LIMIT $limit
"""
SEARCH_USER_METADATA = """
MATCH (n:ExtractedEntity)
WHERE (n.end_user_id = $end_user_id AND n.entity_type ='用户')
RETURN n.description AS description,
n.aliases AS aliases,
n.anchors AS anchors,
n.beliefs_or_stances AS beliefs_or_stances,
n.core_facts AS core_facts,
n.events AS events,
n.goals AS goals,
n.interests AS interests,
n.relations AS relations,
n.traits AS traits
"""
FULLTEXT_QUERY_CYPHER_MAPPING = {
Neo4jNodeType.STATEMENT: SEARCH_STATEMENTS_BY_KEYWORD,
Neo4jNodeType.EXTRACTEDENTITY: SEARCH_ENTITIES_BY_NAME_OR_ALIAS,

View File

@@ -27,9 +27,9 @@ from app.repositories.neo4j.cypher_queries import (
SEARCH_PERCEPTUAL_BY_USER_ID,
FULLTEXT_QUERY_CYPHER_MAPPING,
USER_ID_QUERY_CYPHER_MAPPING,
NODE_ID_QUERY_CYPHER_MAPPING
NODE_ID_QUERY_CYPHER_MAPPING,
SEARCH_USER_METADATA
)
from app.repositories.neo4j.neo4j_connector import Neo4jConnector
logger = logging.getLogger(__name__)
@@ -513,7 +513,7 @@ async def search_graph_by_embedding(
task_keys = []
for node_type in include:
tasks.append(search_by_embedding(connector, node_type, end_user_id, embedding, limit*2))
tasks.append(search_by_embedding(connector, node_type, end_user_id, embedding, limit * 2))
task_keys.append(node_type.value)
task_results = await asyncio.gather(*tasks, return_exceptions=True)
@@ -557,6 +557,17 @@ async def search_graph_by_embedding(
return results
async def search_user_metadata(
connector: Neo4jConnector,
end_user_id: str
) -> dict:
user_info = await connector.execute_query(
SEARCH_USER_METADATA,
end_user_id=end_user_id
)
return user_info[0] if user_info else {}
async def get_dedup_candidates_for_entities( # 适配新版查询:使用全文索引按名称检索候选实体
connector: Neo4jConnector,
end_user_id: str,

View File

@@ -1,14 +1,15 @@
import uuid
from abc import ABC
from typing import Optional
from pydantic import BaseModel
from pydantic import BaseModel, Field
class UserInput(BaseModel):
message: str
history: list[dict]
search_switch: str
end_user_id: str
session_id: uuid.UUID = Field(default_factory=uuid.uuid4)
config_id: Optional[str] = None

View File

@@ -108,6 +108,7 @@ def create_long_term_memory_tool(
try:
with get_db_context() as db:
memory_service = MemoryService(db, config_id, end_user_id)
# TODO: Historical Messages -> Used to refer to coreference resolution
search_result = asyncio.run(memory_service.read(question, SearchStrategy.QUICK))
# memory_content = asyncio.run(

View File

View File

@@ -0,0 +1,77 @@
import json
import logging
import redis.asyncio as redis
from app.aioRedis import get_redis_connection
logger = logging.getLogger(__name__)
DEFAULT_TTL = 3600
class ChatSessionCache:
"""Cache user-AI conversation history in Redis with TTL-based expiry.
Usage::
cache = ChatSessionCache(session_id="user_123")
await cache.append("user", "Hello")
await cache.append("assistant", "Hi there!")
history = await cache.get_history()
"""
def __init__(self, session_id: str, ttl: int = DEFAULT_TTL):
self.session_id = session_id
self.ttl = ttl
self._key = f"chat:session:{session_id}"
@staticmethod
async def _client() -> redis.StrictRedis:
return await get_redis_connection()
async def append(self, role: str, content: str) -> None:
r = await self._client()
entry = json.dumps({"role": role, "content": content}, ensure_ascii=False)
await r.rpush(self._key, entry)
await r.expire(self._key, self.ttl)
async def append_many(self, messages: list[dict[str, str]]) -> None:
"""Batch append messages. Each dict should have ``role`` and ``content`` keys."""
if not messages:
return
r = await self._client()
entries = [
json.dumps(m, ensure_ascii=False)
for m in messages
if "role" in m and "content" in m
]
if entries:
await r.rpush(self._key, *entries)
await r.expire(self._key, self.ttl)
async def get_history(self) -> list[dict[str, str]]:
r = await self._client()
raw = await r.lrange(self._key, 0, -1)
return [json.loads(item) for item in raw]
async def get_history_text(self, user_label: str = "User", ai_label: str = "Assistant") -> str:
"""Return conversation as a formatted text block."""
history = await self.get_history()
lines = []
for msg in history:
role = msg.get("role", "")
content = msg.get("content", "")
label = user_label if role == "user" else ai_label if role == "assistant" else role
lines.append(f"{label}: {content}")
return "\n".join(lines)
async def reset(self) -> None:
"""Delete the session from Redis."""
r = await self._client()
await r.delete(self._key)
async def touch(self) -> None:
"""Refresh the TTL without modifying data."""
r = await self._client()
await r.expire(self._key, self.ttl)