Skip to content

Commit c2792c6

Browse files
authored
docs: fix search method return value handling in integration and example docs (#3208)
1 parent 2307dc8 commit c2792c6

15 files changed

+25
-25
lines changed

docs/examples/ai_companion.mdx

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -87,8 +87,8 @@ class Companion:
8787

8888
previous_memories = self.memory.search(question, user_id=user_id_to_use)
8989
relevant_memories_text = ""
90-
if previous_memories:
91-
relevant_memories_text = '\n'.join(mem["memory"] for mem in previous_memories)
90+
if previous_memories and previous_memories.get('results'):
91+
relevant_memories_text = '\n'.join(mem["memory"] for mem in previous_memories['results'])
9292

9393
prompt = f"User input: {question}\nPrevious {check_answer} memories: {relevant_memories_text}"
9494

@@ -146,7 +146,7 @@ def print_memories(user_id, label):
146146
memories = ai_companion.get_memories(user_id=user_id)
147147
if memories:
148148
for m in memories:
149-
print(f"- {m['text']}")
149+
print(f"- {m['memory']}")
150150
else:
151151
print("No memories found.")
152152

docs/examples/llamaindex-multiagent-learning-system.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -213,8 +213,8 @@ class MultiAgentLearningSystem:
213213
query="learning machine learning"
214214
)
215215

216-
if memories and len(memories):
217-
history = "\n".join(f"- {m['memory']}" for m in memories)
216+
if memories and memories.get('results'):
217+
history = "\n".join(f"- {m['memory']}" for m in memories['results'])
218218
return history
219219
else:
220220
return "No learning history found yet. Let's start building your profile!"

docs/examples/memory-guided-content-writing.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,7 @@ def apply_writing_style(original_content):
8686
print("No preferences found.")
8787
return None
8888

89-
preferences = "\n".join(r["memory"] for r in results)
89+
preferences = "\n".join(r["memory"] for r in results.get('results', []))
9090

9191
system_prompt = f"""
9292
You are a writing assistant.

docs/examples/openai-inbuilt-tools.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -225,7 +225,7 @@ async function addSampleMemories() {
225225

226226
const getMemoryString = (memories) => {
227227
const MEMORY_STRING_PREFIX = "These are the memories I have stored. Give more weightage to the question by users and try to answer that first. You have to modify your answer based on the memories I have provided. If the memories are irrelevant you can ignore them. Also don't reply to this section of the prompt, or the memories, they are only for your reference. The MEMORIES of the USER are: \n\n";
228-
const memoryString = memories.map((mem) => `${mem.memory}`).join("\n") ?? "";
228+
const memoryString = (memories?.results || memories).map((mem) => `${mem.memory}`).join("\n") ?? "";
229229
return memoryString.length > 0 ? `${MEMORY_STRING_PREFIX}${memoryString}` : "";
230230
};
231231

docs/examples/personal-travel-assistant.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -155,11 +155,11 @@ class PersonalTravelAssistant:
155155

156156
def get_memories(self, user_id):
157157
memories = self.memory.get_all(user_id=user_id)
158-
return [m['memory'] for m in memories['memories']]
158+
return [m['memory'] for m in memories.get('results', [])]
159159

160160
def search_memories(self, query, user_id):
161161
memories = self.memory.search(query, user_id=user_id)
162-
return [m['memory'] for m in memories['memories']]
162+
return [m['memory'] for m in memories.get('results', [])]
163163

164164
# Usage example
165165
user_id = "traveler_123"

docs/integrations/agentops.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -104,7 +104,7 @@ def demonstrate_sync_memory(local_config, sample_messages, sample_preferences, u
104104
results = memory.search(query, user_id=user_id)
105105

106106
if results and "results" in results:
107-
for j, result in enumerate(results):
107+
for j, result in enumerate(results['results']):
108108
print(f"Result {j+1}: {result.get('memory', 'N/A')}")
109109
else:
110110
print("No results found")

docs/integrations/agno.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -128,7 +128,7 @@ def chat_user(
128128
if user_input:
129129
# Search for relevant memories
130130
memories = client.search(user_input, user_id=user_id)
131-
memory_context = "\n".join(f"- {m['memory']}" for m in memories)
131+
memory_context = "\n".join(f"- {m['memory']}" for m in memories.get('results', []))
132132

133133
# Construct the prompt
134134
prompt = f"""

docs/integrations/autogen.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -68,7 +68,7 @@ Create a function to get context-aware responses based on user's question and pr
6868
```python
6969
def get_context_aware_response(question):
7070
relevant_memories = memory_client.search(question, user_id=USER_ID)
71-
context = "\n".join([m["memory"] for m in relevant_memories])
71+
context = "\n".join([m["memory"] for m in relevant_memories.get('results', [])])
7272

7373
prompt = f"""Answer the user question considering the previous interactions:
7474
Previous interactions:
@@ -100,7 +100,7 @@ manager = ConversableAgent(
100100

101101
def escalate_to_manager(question):
102102
relevant_memories = memory_client.search(question, user_id=USER_ID)
103-
context = "\n".join([m["memory"] for m in relevant_memories])
103+
context = "\n".join([m["memory"] for m in relevant_memories.get('results', [])])
104104

105105
prompt = f"""
106106
Context from previous interactions:

docs/integrations/elevenlabs.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -152,7 +152,7 @@ Define the two key memory functions that will be registered as tools:
152152
)
153153

154154
# Extract and join the memory texts
155-
memories = ' '.join([result["memory"] for result in results])
155+
memories = ' '.join([result["memory"] for result in results.get('results', [])])
156156
print("[ Memories ]", memories)
157157

158158
if memories:

docs/integrations/google-ai-adk.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -49,8 +49,8 @@ mem0 = MemoryClient()
4949
def search_memory(query: str, user_id: str) -> dict:
5050
"""Search through past conversations and memories"""
5151
memories = mem0.search(query, user_id=user_id)
52-
if memories:
53-
memory_context = "\n".join([f"- {mem['memory']}" for mem in memories])
52+
if memories.get('results', []):
53+
memory_context = "\n".join([f"- {mem['memory']}" for mem in memories.get('results', [])])
5454
return {"status": "success", "memories": memory_context}
5555
return {"status": "no_memories", "message": "No relevant memories found"}
5656

0 commit comments

Comments
 (0)