language model docs
This commit is contained in:
		
							parent
							
								
									ad563e1a5c
								
							
						
					
					
						commit
						d4c4229141
					
				|  | @ -3,36 +3,14 @@ title: "Language Model" | ||||||
| description: "The LLM that powers your 01" | description: "The LLM that powers your 01" | ||||||
| --- | --- | ||||||
| 
 | 
 | ||||||
| ## llamafile |  | ||||||
| 
 |  | ||||||
| llamafile lets you distribute and run LLMs with a single file. Read more about llamafile [here](https://github.com/Mozilla-Ocho/llamafile) |  | ||||||
| 
 |  | ||||||
| ```bash |  | ||||||
| # Set the LLM service to llamafile |  | ||||||
| poetry run 01 --llm-service llamafile |  | ||||||
| ``` |  | ||||||
| 
 |  | ||||||
| ## Llamaedge |  | ||||||
| 
 |  | ||||||
| llamaedge makes it easy for you to run LLM inference apps and create OpenAI-compatible API services for the Llama2 series of LLMs locally. |  | ||||||
| Read more about Llamaedge [here](https://github.com/LlamaEdge/LlamaEdge) |  | ||||||
| 
 |  | ||||||
| ```bash |  | ||||||
| # Set the LLM service to Llamaedge |  | ||||||
| poetry run 01 --llm-service llamaedge |  | ||||||
| ``` |  | ||||||
| 
 |  | ||||||
| ## Hosted Models | ## Hosted Models | ||||||
| 
 | 
 | ||||||
| 01OS leverages liteLLM which supports [many hosted models](https://docs.litellm.ai/docs/providers/). | The default LLM for 01 is GPT-4-Turbo. You can find this in the default profile in `software/source/server/profiles/default.py`. | ||||||
| 
 | 
 | ||||||
| To select your providers | The fast profile uses Llama3-8b served by Groq. You can find this in the fast profile in `software/source/server/profiles/fast.py`. | ||||||
| 
 | 
 | ||||||
| ```bash | ## Local Models | ||||||
| # Set the LLM service |  | ||||||
| poetry run 01 --llm-service openai |  | ||||||
| ``` |  | ||||||
| 
 | 
 | ||||||
| ## Other Models | You can use local models to power 01. | ||||||
| 
 | 
 | ||||||
| More instructions coming soon! | Using the local profile launches the Local Explorer where you can select your inference provider and model. The default options include Llamafile, Jan, Ollama, and LM Studio. | ||||||
|  |  | ||||||
		Loading…
	
		Reference in New Issue
	
	 Mike Bird
						Mike Bird