Files
t6_mem0/tests/llms/test_ollama.py
2024-08-02 23:45:45 +05:30

82 lines
2.7 KiB
Python

import pytest
from unittest.mock import Mock, patch
from mem0.llms.ollama import OllamaLLM
from mem0.configs.llms.base import BaseLlmConfig
from mem0.llms.utils.tools import ADD_MEMORY_TOOL
@pytest.fixture
def mock_ollama_client():
with patch('mem0.llms.ollama.Client') as mock_ollama:
mock_client = Mock()
mock_client.list.return_value = {"models": [{"name": "llama3.1:70b"}]}
mock_ollama.return_value = mock_client
yield mock_client
@pytest.mark.skip(reason="Mock issue, need to be fixed")
def test_generate_response_without_tools(mock_ollama_client):
config = BaseLlmConfig(model="llama3.1:70b", temperature=0.7, max_tokens=100, top_p=1.0)
llm = OllamaLLM(config)
messages = [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Hello, how are you?"}
]
mock_response = Mock()
mock_response.message = {"content": "I'm doing well, thank you for asking!"}
mock_ollama_client.chat.return_value = mock_response
response = llm.generate_response(messages)
mock_ollama_client.chat.assert_called_once_with(
model="llama3.1:70b",
messages=messages,
options={
"temperature": 0.7,
"num_predict": 100,
"top_p": 1.0
}
)
assert response == "I'm doing well, thank you for asking!"
@pytest.mark.skip(reason="Mock issue, need to be fixed")
def test_generate_response_with_tools(mock_ollama_client):
config = BaseLlmConfig(model="llama3.1:70b", temperature=0.7, max_tokens=100, top_p=1.0)
llm = OllamaLLM(config)
messages = [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Add a new memory: Today is a sunny day."}
]
tools = [ADD_MEMORY_TOOL]
mock_response = Mock()
mock_message = {"content": "I've added the memory for you."}
mock_tool_call = {
"function": {
"name": "add_memory",
"arguments": '{"data": "Today is a sunny day."}'
}
}
mock_message["tool_calls"] = [mock_tool_call]
mock_response.message = mock_message
mock_ollama_client.chat.return_value = mock_response
response = llm.generate_response(messages, tools=tools)
mock_ollama_client.chat.assert_called_once_with(
model="llama3.1:70b",
messages=messages,
options={
"temperature": 0.7,
"num_predict": 100,
"top_p": 1.0
},
tools=tools
)
assert response["content"] == "I've added the memory for you."
assert len(response["tool_calls"]) == 1
assert response["tool_calls"][0]["name"] == "add_memory"
assert response["tool_calls"][0]["arguments"] == {'data': 'Today is a sunny day.'}