from unittest.mock import Mock, patch import httpx import pytest from mem0.configs.llms.base import BaseLlmConfig from mem0.llms.azure_openai import AzureOpenAILLM MODEL = "gpt-4o" # or your custom deployment name TEMPERATURE = 0.7 MAX_TOKENS = 100 TOP_P = 1.0 @pytest.fixture def mock_openai_client(): with patch("mem0.llms.azure_openai.AzureOpenAI") as mock_openai: mock_client = Mock() mock_openai.return_value = mock_client yield mock_client def test_generate_response_without_tools(mock_openai_client): config = BaseLlmConfig(model=MODEL, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P) llm = AzureOpenAILLM(config) messages = [ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": "Hello, how are you?"}, ] mock_response = Mock() mock_response.choices = [Mock(message=Mock(content="I'm doing well, thank you for asking!"))] mock_openai_client.chat.completions.create.return_value = mock_response response = llm.generate_response(messages) mock_openai_client.chat.completions.create.assert_called_once_with( model=MODEL, messages=messages, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P ) assert response == "I'm doing well, thank you for asking!" def test_generate_response_with_tools(mock_openai_client): config = BaseLlmConfig(model=MODEL, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P) llm = AzureOpenAILLM(config) messages = [ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": "Add a new memory: Today is a sunny day."}, ] tools = [ { "type": "function", "function": { "name": "add_memory", "description": "Add a memory", "parameters": { "type": "object", "properties": {"data": {"type": "string", "description": "Data to add to memory"}}, "required": ["data"], }, }, } ] mock_response = Mock() mock_message = Mock() mock_message.content = "I've added the memory for you." mock_tool_call = Mock() mock_tool_call.function.name = "add_memory" mock_tool_call.function.arguments = '{"data": "Today is a sunny day."}' mock_message.tool_calls = [mock_tool_call] mock_response.choices = [Mock(message=mock_message)] mock_openai_client.chat.completions.create.return_value = mock_response response = llm.generate_response(messages, tools=tools) mock_openai_client.chat.completions.create.assert_called_once_with( model=MODEL, messages=messages, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P, tools=tools, tool_choice="auto", ) assert response["content"] == "I've added the memory for you." assert len(response["tool_calls"]) == 1 assert response["tool_calls"][0]["name"] == "add_memory" assert response["tool_calls"][0]["arguments"] == {"data": "Today is a sunny day."} @pytest.mark.parametrize( "default_headers", [None, {"Firstkey": "FirstVal", "SecondKey": "SecondVal"}], ) def test_generate_with_http_proxies(default_headers): mock_http_client = Mock() mock_http_client_instance = Mock() mock_http_client.return_value = mock_http_client_instance azure_kwargs = {"api_key": "test"} if default_headers: azure_kwargs["default_headers"] = default_headers with ( patch("mem0.llms.azure_openai.AzureOpenAI") as mock_azure_openai, patch("httpx.Client", new=mock_http_client), ): config = BaseLlmConfig( model=MODEL, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P, api_key="test", http_client_proxies="http://testproxy.mem0.net:8000", azure_kwargs=azure_kwargs, ) _ = AzureOpenAILLM(config) mock_azure_openai.assert_called_once_with( api_key="test", http_client=mock_http_client_instance, azure_deployment=None, azure_endpoint=None, api_version=None, default_headers=default_headers, ) mock_http_client.assert_called_once_with(proxies="http://testproxy.mem0.net:8000")