--- title: Overview --- Mem0 includes built-in support for various popular large language models. Memory can utilize the LLM provided by the user, ensuring efficient use for specific needs. ## Usage To use a llm, you must provide a configuration to customize its usage. If no configuration is supplied, a default configuration will be applied, and `OpenAI` will be used as the llm. For a comprehensive list of available parameters for llm configuration, please refer to [Config](./config). To view all supported llms, visit the [Supported LLMs](./models). ## Structured vs Unstructured Outputs Mem0 supports two types of OpenAI LLM formats, each with its own strengths and use cases: ### Structured Outputs Structured outputs are LLMs that align with OpenAI's structured outputs model: - **Optimized for:** Returning structured responses (e.g., JSON objects) - **Benefits:** Precise, easily parseable data - **Ideal for:** Data extraction, form filling, API responses - **Learn more:** [OpenAI Structured Outputs Guide](https://platform.openai.com/docs/guides/structured-outputs/introduction) ### Unstructured Outputs Unstructured outputs correspond to OpenAI's standard, free-form text model: - **Flexibility:** Returns open-ended, natural language responses - **Customization:** Use the `response_format` parameter to guide output - **Trade-off:** Less efficient than structured outputs for specific data needs - **Best for:** Creative writing, explanations, general conversation Choose the format that best suits your application's requirements for optimal performance and usability.