From 4b97567ed9b80d40b214f12e910a7acc8889d6ce Mon Sep 17 00:00:00 2001 From: Ariya Hidayat Date: Sat, 14 Sep 2024 20:24:52 -0700 Subject: [PATCH] Update README on LLM services configuration --- README.md | 40 +++++++++++++++++++++++----------------- 1 file changed, 23 insertions(+), 17 deletions(-) diff --git a/README.md b/README.md index 74d156a..a6f25ad 100644 --- a/README.md +++ b/README.md @@ -64,71 +64,77 @@ export LLM_API_BASE_URL=http://localhost:3928/v1 ## Using Managed LLM Services -To use [OpenAI GPT model](https://platform.openai.com/docs), configure the environment variable `OPENAI_API_KEY` with your API key: -```bash -export OPENAI_API_KEY="sk-yourownapikey" -``` +Supported LLM services include [Deep Infra](https://deepinfra.com), [Fireworks](https://fireworks.ai), [Groq](https://groq.com), [Hyperbolic](https://www.hyperbolic.xyz), [Lepton](https://lepton.ai), [Novita](https://novita.ai), [Octo](https://octo.ai), [OpenAI](https://platform.openai.com), [OpenRouter](https://openrouter.ai), and [Together](https://www.together.ai). + +For configuration specifics, refer to the relevant section. The examples use Llama-3.1 8B (or GPT-4o Mini for OpenAI), but any LLM with at least 7B parameters should work just as well, such as Mistral 7B, Qwen-2 7B, or Gemma-2 9B. -To utilize other LLM services, populate the relevant environment variables as demonstrated in the following examples: * [Deep Infra](https://deepinfra.com) ```bash export LLM_API_BASE_URL=https://api.deepinfra.com/v1/openai export LLM_API_KEY="yourownapikey" -export LLM_CHAT_MODEL="mistralai/Mistral-7B-Instruct-v0.1" +export LLM_CHAT_MODEL="meta-llama/Meta-Llama-3.1-8B-Instruct" ``` * [Fireworks](https://fireworks.ai/) ```bash export LLM_API_BASE_URL=https://api.fireworks.ai/inference/v1 export LLM_API_KEY="yourownapikey" -export LLM_CHAT_MODEL="accounts/fireworks/models/llama-v3-8b-instruct" +export LLM_CHAT_MODEL="accounts/fireworks/models/llama-v3p1-8b-instruct" ``` * [Groq](https://groq.com/) ```bash export LLM_API_BASE_URL=https://api.groq.com/openai/v1 export LLM_API_KEY="yourownapikey" -export LLM_CHAT_MODEL="gemma-7b-it" +export LLM_CHAT_MODEL="llama-3.1-8b-instant" ``` -* [Lepton](https://lepton.ai) +* [Hyperbolic](https://www.hyperbolic.xyz) ```bash -export LLM_API_BASE_URL=https://mixtral-8x7b.lepton.run/api/v1/ +export LLM_API_BASE_URL=https://api.hyperbolic.xyz/v1 export LLM_API_KEY="yourownapikey" +export LLM_CHAT_MODEL="meta-llama/Meta-Llama-3.1-8B-Instruct" ``` -* [Hyperbolic](https://www.hyperbolic.xyz) +* [Lepton](https://lepton.ai) ```bash -export LLM_API_BASE_URL=https://api.hyperbolic.xyz/v1 +export LLM_API_BASE_URL=https://llama3-1-8b.lepton.run/api/v1 export LLM_API_KEY="yourownapikey" -export LLM_CHAT_MODEL="meta-llama/Meta-Llama-3.1-8B-Instruct" +export LLM_CHAT_MODEL="llama3-1-8b" ``` * [Novita](https://novita.ai) ```bash export LLM_API_BASE_URL=https://api.novita.ai/v3/openai export LLM_API_KEY="yourownapikey" -export LLM_CHAT_MODEL="meta-llama/llama-3-8b-instruct" +export LLM_CHAT_MODEL="meta-llama/llama-3.1-8b-instruct" ``` * [Octo](https://octo.ai) ```bash export LLM_API_BASE_URL=https://text.octoai.run/v1/ export LLM_API_KEY="yourownapikey" -export LLM_CHAT_MODEL="hermes-2-pro-mistral-7b" +export LLM_CHAT_MODEL="meta-llama-3.1-8b-instruct" +``` + +* [OpenAI](https://platform.openai.com) +```bash +export LLM_API_BASE_URL=https://api.openai.com/v1 +export LLM_API_KEY="yourownapikey" +export LLM_CHAT_MODEL="gpt-4o-mini" ``` * [OpenRouter](https://openrouter.ai/) ```bash export LLM_API_BASE_URL=https://openrouter.ai/api/v1 export LLM_API_KEY="yourownapikey" -export LLM_CHAT_MODEL="meta-llama/llama-3-8b-instruct:free" +export LLM_CHAT_MODEL="meta-llama/llama-3.1-8b-instruct" ``` * [Together](https://www.together.ai/) ```bash export LLM_API_BASE_URL=https://api.together.xyz/v1 export LLM_API_KEY="yourownapikey" -export LLM_CHAT_MODEL="meta-llama/Llama-3-8b-chat-hf" +export LLM_CHAT_MODEL="meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo" ```