from unittest.mock import Mock, patch import httpx import pytest from mem0.configs.llms.base import BaseLlmConfig from mem0.llms.azure_openai import AzureOpenAILLM MODEL = "gpt-4o" # or your custom deployment name TEMPERATURE = 0.7 MAX_TOKENS = 100 TOP_P = 1.0 @pytest.fixture def mock_openai_client(): with patch("mem0.llms.azure_openai.AzureOpenAI") as mock_openai: mock_client = Mock() mock_openai.return_value = mock_client yield mock_client def test_generate_response(mock_openai_client): config = BaseLlmConfig( model=MODEL, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P ) llm = AzureOpenAILLM(config) messages = [ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": "Hello, how are you?"}, ] mock_response = Mock() mock_response.choices = [ Mock(message=Mock(content="I'm doing well, thank you for asking!")) ] mock_openai_client.chat.completions.create.return_value = mock_response response = llm.generate_response(messages) mock_openai_client.chat.completions.create.assert_called_once_with( model=MODEL, messages=messages, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P, ) assert response == "I'm doing well, thank you for asking!" @pytest.mark.parametrize( "default_headers", [None, {"Firstkey": "FirstVal", "SecondKey": "SecondVal"}], ) def test_generate_with_http_proxies(default_headers): mock_http_client = Mock() mock_http_client_instance = Mock() mock_http_client.return_value = mock_http_client_instance azure_kwargs = {"api_key": "test"} if default_headers: azure_kwargs["default_headers"] = default_headers with ( patch("mem0.llms.azure_openai.AzureOpenAI") as mock_azure_openai, patch("httpx.Client", new=mock_http_client), ): config = BaseLlmConfig( model=MODEL, temperature=TEMPERATURE, max_tokens=MAX_TOKENS, top_p=TOP_P, api_key="test", http_client_proxies="http://testproxy.mem0.net:8000", azure_kwargs=azure_kwargs, ) _ = AzureOpenAILLM(config) mock_azure_openai.assert_called_once_with( api_key="test", http_client=mock_http_client_instance, azure_deployment=None, azure_endpoint=None, api_version=None, default_headers=default_headers, ) mock_http_client.assert_called_once_with( proxies="http://testproxy.mem0.net:8000" )