69 lines
2.4 KiB
Python
69 lines
2.4 KiB
Python
from unittest.mock import Mock, patch
|
|
|
|
import pytest
|
|
|
|
from mem0.configs.llms.base import BaseLlmConfig
|
|
from mem0.llms.ollama import OllamaLLM
|
|
from mem0.llms.utils.tools import ADD_MEMORY_TOOL
|
|
|
|
|
|
@pytest.fixture
|
|
def mock_ollama_client():
|
|
with patch("mem0.llms.ollama.Client") as mock_ollama:
|
|
mock_client = Mock()
|
|
mock_client.list.return_value = {"models": [{"name": "llama3.1:70b"}]}
|
|
mock_ollama.return_value = mock_client
|
|
yield mock_client
|
|
|
|
|
|
def test_generate_response_without_tools(mock_ollama_client):
|
|
config = BaseLlmConfig(model="llama3.1:70b", temperature=0.7, max_tokens=100, top_p=1.0)
|
|
llm = OllamaLLM(config)
|
|
messages = [
|
|
{"role": "system", "content": "You are a helpful assistant."},
|
|
{"role": "user", "content": "Hello, how are you?"},
|
|
]
|
|
|
|
mock_response = {"message": {"content": "I'm doing well, thank you for asking!"}}
|
|
mock_ollama_client.chat.return_value = mock_response
|
|
|
|
response = llm.generate_response(messages)
|
|
|
|
mock_ollama_client.chat.assert_called_once_with(
|
|
model="llama3.1:70b", messages=messages, options={"temperature": 0.7, "num_predict": 100, "top_p": 1.0}
|
|
)
|
|
assert response == "I'm doing well, thank you for asking!"
|
|
|
|
|
|
def test_generate_response_with_tools(mock_ollama_client):
|
|
config = BaseLlmConfig(model="llama3.1:70b", temperature=0.7, max_tokens=100, top_p=1.0)
|
|
llm = OllamaLLM(config)
|
|
messages = [
|
|
{"role": "system", "content": "You are a helpful assistant."},
|
|
{"role": "user", "content": "Add a new memory: Today is a sunny day."},
|
|
]
|
|
tools = [ADD_MEMORY_TOOL]
|
|
|
|
mock_response = {
|
|
"message": {
|
|
"content": "I've added the memory for you.",
|
|
"tool_calls": [{"function": {"name": "add_memory", "arguments": {"data": "Today is a sunny day."}}}],
|
|
}
|
|
}
|
|
|
|
mock_ollama_client.chat.return_value = mock_response
|
|
|
|
response = llm.generate_response(messages, tools=tools)
|
|
|
|
mock_ollama_client.chat.assert_called_once_with(
|
|
model="llama3.1:70b",
|
|
messages=messages,
|
|
options={"temperature": 0.7, "num_predict": 100, "top_p": 1.0},
|
|
tools=tools,
|
|
)
|
|
|
|
assert response["content"] == "I've added the memory for you."
|
|
assert len(response["tool_calls"]) == 1
|
|
assert response["tool_calls"][0]["name"] == "add_memory"
|
|
assert response["tool_calls"][0]["arguments"] == {"data": "Today is a sunny day."}
|