Formatting (#2750)
This commit is contained in:
@@ -25,7 +25,8 @@
|
||||
"source": [
|
||||
"# Set up ENV Vars\n",
|
||||
"import os\n",
|
||||
"os.environ['OPENAI_API_KEY'] = \"sk-xxx\"\n"
|
||||
"\n",
|
||||
"os.environ[\"OPENAI_API_KEY\"] = \"sk-xxx\""
|
||||
]
|
||||
},
|
||||
{
|
||||
@@ -133,11 +134,9 @@
|
||||
"assistant_id = os.environ.get(\"ASSISTANT_ID\", None)\n",
|
||||
"\n",
|
||||
"# LLM Configuration\n",
|
||||
"CACHE_SEED = 42 # choose your poison\n",
|
||||
"CACHE_SEED = 42 # choose your poison\n",
|
||||
"llm_config = {\n",
|
||||
" \"config_list\": [\n",
|
||||
" {\"model\": \"gpt-4o\", \"api_key\": os.environ[\"OPENAI_API_KEY\"]}\n",
|
||||
" ],\n",
|
||||
" \"config_list\": [{\"model\": \"gpt-4o\", \"api_key\": os.environ[\"OPENAI_API_KEY\"]}],\n",
|
||||
" \"cache_seed\": CACHE_SEED,\n",
|
||||
" \"timeout\": 120,\n",
|
||||
" \"temperature\": 0.0,\n",
|
||||
@@ -348,7 +347,7 @@
|
||||
"source": [
|
||||
"# Retrieve the memory\n",
|
||||
"relevant_memories = MEM0_MEMORY_CLIENT.search(user_query, user_id=USER_ID, limit=3)\n",
|
||||
"relevant_memories_text = '\\n'.join(mem['memory'] for mem in relevant_memories)\n",
|
||||
"relevant_memories_text = \"\\n\".join(mem[\"memory\"] for mem in relevant_memories)\n",
|
||||
"print(\"Relevant memories:\")\n",
|
||||
"print(relevant_memories_text)\n",
|
||||
"\n",
|
||||
@@ -389,8 +388,8 @@
|
||||
"# - Enables more context-aware and personalized agent responses.\n",
|
||||
"# - Bridges the gap between human input and AI processing in complex workflows.\n",
|
||||
"\n",
|
||||
"class Mem0ProxyCoderAgent(UserProxyAgent):\n",
|
||||
"\n",
|
||||
"class Mem0ProxyCoderAgent(UserProxyAgent):\n",
|
||||
" def __init__(self, *args, **kwargs):\n",
|
||||
" super().__init__(*args, **kwargs)\n",
|
||||
" self.memory = MEM0_MEMORY_CLIENT\n",
|
||||
@@ -399,15 +398,14 @@
|
||||
" def initiate_chat(self, assistant, message):\n",
|
||||
" # Retrieve memory for the agent\n",
|
||||
" agent_memories = self.memory.search(message, agent_id=self.agent_id, limit=3)\n",
|
||||
" agent_memories_txt = '\\n'.join(mem['memory'] for mem in agent_memories)\n",
|
||||
" agent_memories_txt = \"\\n\".join(mem[\"memory\"] for mem in agent_memories)\n",
|
||||
" prompt = f\"{message}\\n Coding Preferences: \\n{str(agent_memories_txt)}\"\n",
|
||||
" response = super().initiate_chat(assistant, message=prompt)\n",
|
||||
" # Add new memory after processing the message\n",
|
||||
" response_dist = response.__dict__ if not isinstance(response, dict) else response\n",
|
||||
" MEMORY_DATA = [{\"role\": \"user\", \"content\": message}, {\"role\": \"assistant\", \"content\": response_dist}]\n",
|
||||
" self.memory.add(MEMORY_DATA, agent_id=self.agent_id)\n",
|
||||
" return response\n",
|
||||
" "
|
||||
" return response"
|
||||
]
|
||||
},
|
||||
{
|
||||
@@ -560,12 +558,12 @@
|
||||
"from cookbooks.helper.mem0_teachability import Mem0Teachability\n",
|
||||
"\n",
|
||||
"teachability = Mem0Teachability(\n",
|
||||
" verbosity=2, # for visibility of what's happening\n",
|
||||
" recall_threshold=0.5,\n",
|
||||
" reset_db=False, # Use True to force-reset the memo DB, and False to use an existing DB.\n",
|
||||
" agent_id=AGENT_ID,\n",
|
||||
" memory_client = MEM0_MEMORY_CLIENT,\n",
|
||||
" )\n",
|
||||
" verbosity=2, # for visibility of what's happening\n",
|
||||
" recall_threshold=0.5,\n",
|
||||
" reset_db=False, # Use True to force-reset the memo DB, and False to use an existing DB.\n",
|
||||
" agent_id=AGENT_ID,\n",
|
||||
" memory_client=MEM0_MEMORY_CLIENT,\n",
|
||||
")\n",
|
||||
"teachability.add_to_agent(user_proxy)"
|
||||
]
|
||||
},
|
||||
|
||||
Reference in New Issue
Block a user