[chore] Update LLM YAML config docs (#826)
This commit is contained in:
@@ -13,8 +13,8 @@ app:
|
|||||||
|
|
||||||
llm:
|
llm:
|
||||||
provider: openai
|
provider: openai
|
||||||
model: 'gpt-3.5-turbo'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'gpt-3.5-turbo'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 1
|
top_p: 1
|
||||||
|
|||||||
@@ -65,8 +65,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: azure_openai
|
provider: azure_openai
|
||||||
model: gpt-35-turbo
|
|
||||||
config:
|
config:
|
||||||
|
model: gpt-35-turbo
|
||||||
deployment_name: your_llm_deployment_name
|
deployment_name: your_llm_deployment_name
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
@@ -99,8 +99,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: gpt4all
|
provider: gpt4all
|
||||||
model: 'orca-mini-3b.ggmlv3.q4_0.bin'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'orca-mini-3b.ggmlv3.q4_0.bin'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 1
|
top_p: 1
|
||||||
@@ -130,8 +130,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: huggingface
|
provider: huggingface
|
||||||
model: 'google/flan-t5-xxl'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'google/flan-t5-xxl'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 0.5
|
top_p: 0.5
|
||||||
@@ -161,8 +161,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: vertexai
|
provider: vertexai
|
||||||
model: 'chat-bison'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'chat-bison'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
top_p: 0.5
|
top_p: 0.5
|
||||||
|
|
||||||
|
|||||||
@@ -52,8 +52,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: openai
|
provider: openai
|
||||||
model: 'gpt-3.5-turbo'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'gpt-3.5-turbo'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 1
|
top_p: 1
|
||||||
@@ -84,8 +84,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: azure_openai
|
provider: azure_openai
|
||||||
model: gpt-35-turbo
|
|
||||||
config:
|
config:
|
||||||
|
model: gpt-35-turbo
|
||||||
deployment_name: your_llm_deployment_name
|
deployment_name: your_llm_deployment_name
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
@@ -121,8 +121,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: anthropic
|
provider: anthropic
|
||||||
model: 'claude-instant-1'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'claude-instant-1'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 1
|
top_p: 1
|
||||||
@@ -158,8 +158,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: cohere
|
provider: cohere
|
||||||
model: large
|
|
||||||
config:
|
config:
|
||||||
|
model: large
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 1
|
top_p: 1
|
||||||
@@ -189,8 +189,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: gpt4all
|
provider: gpt4all
|
||||||
model: 'orca-mini-3b.ggmlv3.q4_0.bin'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'orca-mini-3b.ggmlv3.q4_0.bin'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 1
|
top_p: 1
|
||||||
@@ -261,8 +261,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: huggingface
|
provider: huggingface
|
||||||
model: 'google/flan-t5-xxl'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'google/flan-t5-xxl'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 0.5
|
top_p: 0.5
|
||||||
@@ -291,8 +291,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: llama2
|
provider: llama2
|
||||||
model: 'a16z-infra/llama13b-v2-chat:df7690f1994d94e96ad9d568eac121aecf50684a0b0963b25a41cc40061269e5'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'a16z-infra/llama13b-v2-chat:df7690f1994d94e96ad9d568eac121aecf50684a0b0963b25a41cc40061269e5'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 0.5
|
top_p: 0.5
|
||||||
@@ -316,8 +316,8 @@ app = App.from_config(yaml_path="config.yaml")
|
|||||||
```yaml config.yaml
|
```yaml config.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: vertexai
|
provider: vertexai
|
||||||
model: 'chat-bison'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'chat-bison'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
top_p: 0.5
|
top_p: 0.5
|
||||||
```
|
```
|
||||||
|
|||||||
@@ -20,8 +20,8 @@ app = App.from_config(yaml_path="gpt4.yaml")
|
|||||||
```yaml gpt4.yaml
|
```yaml gpt4.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: openai
|
provider: openai
|
||||||
model: 'gpt-4'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'gpt-4'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 1
|
top_p: 1
|
||||||
@@ -47,8 +47,8 @@ app = App.from_config(yaml_path="opensource.yaml")
|
|||||||
```yaml opensource.yaml
|
```yaml opensource.yaml
|
||||||
llm:
|
llm:
|
||||||
provider: gpt4all
|
provider: gpt4all
|
||||||
model: 'orca-mini-3b.ggmlv3.q4_0.bin'
|
|
||||||
config:
|
config:
|
||||||
|
model: 'orca-mini-3b.ggmlv3.q4_0.bin'
|
||||||
temperature: 0.5
|
temperature: 0.5
|
||||||
max_tokens: 1000
|
max_tokens: 1000
|
||||||
top_p: 1
|
top_p: 1
|
||||||
|
|||||||
Reference in New Issue
Block a user