Allow ollama llm to take custom callback for handling streaming (#1376)

This commit is contained in:
Aditya Veer Parmar
2024-06-17 21:14:52 +05:30
committed by GitHub
parent c558eae9ce
commit dc0d8e0932
2 changed files with 23 additions and 2 deletions

View File

@@ -2,6 +2,7 @@ import pytest
from embedchain.config import BaseLlmConfig
from embedchain.llm.ollama import OllamaLlm
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
@pytest.fixture
@@ -31,3 +32,20 @@ def test_get_answer_mocked_ollama(ollama_llm_config, mocker):
answer = llm.get_llm_model_answer(prompt)
assert answer == "Mocked answer"
def test_get_llm_model_answer_with_streaming(ollama_llm_config, mocker):
ollama_llm_config.stream = True
ollama_llm_config.callbacks = [StreamingStdOutCallbackHandler()]
mocked_ollama_chat = mocker.patch("embedchain.llm.ollama.OllamaLlm._get_answer", return_value="Test answer")
llm = OllamaLlm(ollama_llm_config)
llm.get_llm_model_answer("Test query")
mocked_ollama_chat.assert_called_once()
call_args = mocked_ollama_chat.call_args
config_arg = call_args[1]["config"]
callbacks = config_arg.callbacks
assert len(callbacks) == 1
assert isinstance(callbacks[0], StreamingStdOutCallbackHandler)