[Fix] Added missing provider for 'vllm' (#1316)
This commit is contained in:
@@ -25,6 +25,7 @@ class LlmFactory:
|
|||||||
"mistralai": "embedchain.llm.mistralai.MistralAILlm",
|
"mistralai": "embedchain.llm.mistralai.MistralAILlm",
|
||||||
"groq": "embedchain.llm.groq.GroqLlm",
|
"groq": "embedchain.llm.groq.GroqLlm",
|
||||||
"nvidia": "embedchain.llm.nvidia.NvidiaLlm",
|
"nvidia": "embedchain.llm.nvidia.NvidiaLlm",
|
||||||
|
"vllm": "embedchain.llm.vllm.VLLM",
|
||||||
}
|
}
|
||||||
provider_to_config_class = {
|
provider_to_config_class = {
|
||||||
"embedchain": "embedchain.config.llm.base.BaseLlmConfig",
|
"embedchain": "embedchain.config.llm.base.BaseLlmConfig",
|
||||||
|
|||||||
Reference in New Issue
Block a user