#1155: Add support for OpenAI-compatible endpoint in LLM and Embed (#1197)

This commit is contained in:
Joe
2024-03-04 20:17:20 -06:00
committed by GitHub
parent 6078738d34
commit 11f4ce8fb6
5 changed files with 17 additions and 2 deletions

View File

@@ -39,13 +39,20 @@ class OpenAILlm(BaseLlm):
"model_kwargs": {},
}
api_key = config.api_key or os.environ["OPENAI_API_KEY"]
base_url = config.base_url or os.environ.get("OPENAI_API_BASE", None)
if config.top_p:
kwargs["model_kwargs"]["top_p"] = config.top_p
if config.stream:
callbacks = config.callbacks if config.callbacks else [StreamingStdOutCallbackHandler()]
chat = ChatOpenAI(**kwargs, streaming=config.stream, callbacks=callbacks, api_key=api_key)
chat = ChatOpenAI(
**kwargs,
streaming=config.stream,
callbacks=callbacks,
api_key=api_key,
base_url=base_url,
)
else:
chat = ChatOpenAI(**kwargs, api_key=api_key)
chat = ChatOpenAI(**kwargs, api_key=api_key, base_url=base_url)
if self.tools:
return self._query_function_call(chat, self.tools, messages)