from unittest.mock import Mock, patch import httpx import pytest from mem0.configs.llms.base import BaseLlmConfig from mem0.llms.azure_openai import AzureOpenAILLM MODEL = "gpt-4o" # or your custom deployment name TEMPERATURE = 0.7 MAX_TOKENS = 100 TOP_P = 1.0 @pytest.fixture def mock_openai_client(): with patch('mem0.llms.azure_openai.AzureOpenAI') as mock_openai: mock_client = Mock() mock_openai.return_value = mock_client yield mock_client def test_generate_response_without_tools(mock_openai_client): config = BaseLlmConfig(model=MODEL, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P) llm = AzureOpenAILLM(config) messages = [ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": "Hello, how are you?"} ] mock_response = Mock() mock_response.choices = [Mock(message=Mock(content="I'm doing well, thank you for asking!"))] mock_openai_client.chat.completions.create.return_value = mock_response response = llm.generate_response(messages) mock_openai_client.chat.completions.create.assert_called_once_with( model=MODEL, messages=messages, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P ) assert response == "I'm doing well, thank you for asking!" def test_generate_response_with_tools(mock_openai_client): config = BaseLlmConfig(model=MODEL, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P) llm = AzureOpenAILLM(config) messages = [ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": "Add a new memory: Today is a sunny day."} ] tools = [ { "type": "function", "function": { "name": "add_memory", "description": "Add a memory", "parameters": { "type": "object", "properties": { "data": {"type": "string", "description": "Data to add to memory"} }, "required": ["data"], }, }, } ] mock_response = Mock() mock_message = Mock() mock_message.content = "I've added the memory for you." mock_tool_call = Mock() mock_tool_call.function.name = "add_memory" mock_tool_call.function.arguments = '{"data": "Today is a sunny day."}' mock_message.tool_calls = [mock_tool_call] mock_response.choices = [Mock(message=mock_message)] mock_openai_client.chat.completions.create.return_value = mock_response response = llm.generate_response(messages, tools=tools) mock_openai_client.chat.completions.create.assert_called_once_with( model=MODEL, messages=messages, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P, tools=tools, tool_choice="auto" ) assert response["content"] == "I've added the memory for you." assert len(response["tool_calls"]) == 1 assert response["tool_calls"][0]["name"] == "add_memory" assert response["tool_calls"][0]["arguments"] == {'data': 'Today is a sunny day.'} def test_generate_with_http_proxies(): mock_http_client = Mock(spec=httpx.Client) mock_http_client_instance = Mock(spec=httpx.Client) mock_http_client.return_value = mock_http_client_instance with (patch("mem0.llms.azure_openai.AzureOpenAI") as mock_azure_openai, patch("httpx.Client", new=mock_http_client) as mock_http_client): config = BaseLlmConfig(model=MODEL, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P, api_key="test", http_client_proxies="http://testproxy.mem0.net:8000", azure_kwargs= {"api_key" : "test"}) _ = AzureOpenAILLM(config) mock_azure_openai.assert_called_once_with( api_key="test", http_client=mock_http_client_instance, azure_deployment=None, azure_endpoint=None, api_version=None ) mock_http_client.assert_called_once_with(proxies="http://testproxy.mem0.net:8000")