Add TogetherAI support (#1485)
This commit is contained in:
@@ -3,14 +3,14 @@ from unittest.mock import Mock, patch
|
||||
from mem0.llms.openai import OpenAILLM
|
||||
|
||||
@pytest.fixture
|
||||
def mock_groq_client():
|
||||
with patch('mem0.llms.openai.OpenAI') as mock_groq:
|
||||
def mock_openai_client():
|
||||
with patch('mem0.llms.openai.OpenAI') as mock_openai:
|
||||
mock_client = Mock()
|
||||
mock_groq.return_value = mock_client
|
||||
mock_openai.return_value = mock_client
|
||||
yield mock_client
|
||||
|
||||
|
||||
def test_generate_response_without_tools(mock_groq_client):
|
||||
def test_generate_response_without_tools(mock_openai_client):
|
||||
llm = OpenAILLM()
|
||||
messages = [
|
||||
{"role": "system", "content": "You are a helpful assistant."},
|
||||
@@ -19,18 +19,18 @@ def test_generate_response_without_tools(mock_groq_client):
|
||||
|
||||
mock_response = Mock()
|
||||
mock_response.choices = [Mock(message=Mock(content="I'm doing well, thank you for asking!"))]
|
||||
mock_groq_client.chat.completions.create.return_value = mock_response
|
||||
mock_openai_client.chat.completions.create.return_value = mock_response
|
||||
|
||||
response = llm.generate_response(messages)
|
||||
|
||||
mock_groq_client.chat.completions.create.assert_called_once_with(
|
||||
mock_openai_client.chat.completions.create.assert_called_once_with(
|
||||
model="gpt-4o",
|
||||
messages=messages
|
||||
)
|
||||
assert response.choices[0].message.content == "I'm doing well, thank you for asking!"
|
||||
|
||||
|
||||
def test_generate_response_with_tools(mock_groq_client):
|
||||
def test_generate_response_with_tools(mock_openai_client):
|
||||
llm = OpenAILLM()
|
||||
messages = [
|
||||
{"role": "system", "content": "You are a helpful assistant."},
|
||||
@@ -55,11 +55,11 @@ def test_generate_response_with_tools(mock_groq_client):
|
||||
|
||||
mock_response = Mock()
|
||||
mock_response.choices = [Mock(message=Mock(content="Memory added successfully."))]
|
||||
mock_groq_client.chat.completions.create.return_value = mock_response
|
||||
mock_openai_client.chat.completions.create.return_value = mock_response
|
||||
|
||||
response = llm.generate_response(messages, tools=tools)
|
||||
|
||||
mock_groq_client.chat.completions.create.assert_called_once_with(
|
||||
mock_openai_client.chat.completions.create.assert_called_once_with(
|
||||
model="gpt-4o",
|
||||
messages=messages,
|
||||
tools=tools,
|
||||
|
||||
69
tests/llms/test_together.py
Normal file
69
tests/llms/test_together.py
Normal file
@@ -0,0 +1,69 @@
|
||||
import pytest
|
||||
from unittest.mock import Mock, patch
|
||||
from mem0.llms.together import TogetherLLM
|
||||
|
||||
@pytest.fixture
|
||||
def mock_together_client():
|
||||
with patch('mem0.llms.together.Together') as mock_together:
|
||||
mock_client = Mock()
|
||||
mock_together.return_value = mock_client
|
||||
yield mock_client
|
||||
|
||||
|
||||
def test_generate_response_without_tools(mock_together_client):
|
||||
llm = TogetherLLM()
|
||||
messages = [
|
||||
{"role": "system", "content": "You are a helpful assistant."},
|
||||
{"role": "user", "content": "Hello, how are you?"}
|
||||
]
|
||||
|
||||
mock_response = Mock()
|
||||
mock_response.choices = [Mock(message=Mock(content="I'm doing well, thank you for asking!"))]
|
||||
mock_together_client.chat.completions.create.return_value = mock_response
|
||||
|
||||
response = llm.generate_response(messages)
|
||||
|
||||
mock_together_client.chat.completions.create.assert_called_once_with(
|
||||
model="mistralai/Mixtral-8x7B-Instruct-v0.1",
|
||||
messages=messages
|
||||
)
|
||||
assert response.choices[0].message.content == "I'm doing well, thank you for asking!"
|
||||
|
||||
|
||||
def test_generate_response_with_tools(mock_together_client):
|
||||
llm = TogetherLLM()
|
||||
messages = [
|
||||
{"role": "system", "content": "You are a helpful assistant."},
|
||||
{"role": "user", "content": "Add a new memory: Today is a sunny day."}
|
||||
]
|
||||
tools = [
|
||||
{
|
||||
"type": "function",
|
||||
"function": {
|
||||
"name": "add_memory",
|
||||
"description": "Add a memory",
|
||||
"parameters": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"data": {"type": "string", "description": "Data to add to memory"}
|
||||
},
|
||||
"required": ["data"],
|
||||
},
|
||||
},
|
||||
}
|
||||
]
|
||||
|
||||
mock_response = Mock()
|
||||
mock_response.choices = [Mock(message=Mock(content="Memory added successfully."))]
|
||||
mock_together_client.chat.completions.create.return_value = mock_response
|
||||
|
||||
response = llm.generate_response(messages, tools=tools)
|
||||
|
||||
mock_together_client.chat.completions.create.assert_called_once_with(
|
||||
model="mistralai/Mixtral-8x7B-Instruct-v0.1",
|
||||
messages=messages,
|
||||
tools=tools,
|
||||
tool_choice="auto"
|
||||
)
|
||||
assert response.choices[0].message.content == "Memory added successfully."
|
||||
|
||||
Reference in New Issue
Block a user