add llm examples to configure
This commit is contained in:
		
							parent
							
								
									db3e2c3638
								
							
						
					
					
						commit
						ef1e711986
					
				| 
						 | 
					@ -45,12 +45,22 @@ The default LLM for 01 is GPT-4-Turbo. You can find this in the default profile
 | 
				
			||||||
 | 
					
 | 
				
			||||||
The fast profile uses Llama3-8b served by Groq. You can find this in the fast profile in `software/source/server/profiles/fast.py`.
 | 
					The fast profile uses Llama3-8b served by Groq. You can find this in the fast profile in `software/source/server/profiles/fast.py`.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					```python
 | 
				
			||||||
 | 
					# Set your profile with a hosted LLM
 | 
				
			||||||
 | 
					interpreter.llm.model = "gpt-4o"
 | 
				
			||||||
 | 
					```
 | 
				
			||||||
 | 
					
 | 
				
			||||||
### Local LLMs
 | 
					### Local LLMs
 | 
				
			||||||
 | 
					
 | 
				
			||||||
You can use local models to power 01.
 | 
					You can use local models to power 01.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
Using the local profile launches the Local Explorer where you can select your inference provider and model. The default options include Llamafile, Jan, Ollama, and LM Studio.
 | 
					Using the local profile launches the Local Explorer where you can select your inference provider and model. The default options include Llamafile, Jan, Ollama, and LM Studio.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					```python
 | 
				
			||||||
 | 
					# Set your profile with a local LLM
 | 
				
			||||||
 | 
					interpreter.local_setup()
 | 
				
			||||||
 | 
					```
 | 
				
			||||||
 | 
					
 | 
				
			||||||
### Hosted TTS
 | 
					### Hosted TTS
 | 
				
			||||||
 | 
					
 | 
				
			||||||
01 supports OpenAI and Elevenlabs for hosted TTS
 | 
					01 supports OpenAI and Elevenlabs for hosted TTS
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in New Issue