Add Gemini Model Support to Vercel AI SDK Provider (#3094)

This commit is contained in:
Saket Aryan
2025-07-03 22:24:11 +05:30
committed by GitHub
parent 5b0f1a7cf8
commit 7476c39257
3 changed files with 14 additions and 1 deletions

View File

@@ -225,6 +225,18 @@ const memories = await getMemories(prompt, { user_id: "borat", mem0ApiKey: "m0-x
The `getMemories` function will return an object with two keys: `results` and `relations`, if `enable_graph` is set to `true`. Otherwise, it will return an array of objects.
## Supported LLM Providers
| Provider | Configuration Value |
|----------|-------------------|
| OpenAI | openai |
| Anthropic | anthropic |
| Gemini | gemini |
| Google | google |
| Mistral | mistral |
| Groq | groq |
> **Note**: You can use `google` as provider for Gemini (Google) models. They are same and internally they use `@ai-sdk/google` package.
## Key Features