-
Notifications
You must be signed in to change notification settings - Fork 43
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Local LLMs #2
Comments
This works partially already. If you set your OAI_CONFIG_LIST.json file to contain something like this:
Then at least the autogen_test.py will use the given base_url. This is the default URL for the Kobold server. I don't think the model or api_key matter in this case. You have to start up the Kobold server with the model that will be used. The example_rag.py program, which uses llama_index code, seems to have some hardwired use of the openai.com server. |
The python autogen_modified_group_chat.py and python autogen_standard_group_chat.py gives this error. INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 401 Unauthorized" python example_rag.py gives me INFO:httpx:HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 401 Unauthorized" python example_research.py wants my Github token. I'm not comfortable trusting this code to give it that yet. So this still seems tied into OpenAPI pretty hard. |
I've dived into this a bit and figured out how to place my local LLM stuff into the base code. I will do a pull request soon for it, but I need to fix one thing with misc.py.. So it calls the OpenAIWrapper.. and Open AI from llama_index.llms...
These seem to force the call to
instead of where I put my LLM.. So in a lot of the files (including misc.py), I made this..
and I went through and replaced config_3 and config_4 with config_listLLM. This works fine until the reply back to the user.
So there is some hardcoding of the URL (which makes sense). I'm not sure how to fix it though as it seems something in a package and my skill level isn't good enough to be editing the packages directly. Can there be a flag in the package that if a LLM is defined in .env, then it can use that url instead of the openapi one? |
Thanks @LanceLake and @dlaliberte for looking into this. I'll try to make local LLM compatibility a priority when I get some more time to work on the project. |
|
Allow to specify URL to use to run it with local LLMs like Ooga Booga.
The text was updated successfully, but these errors were encountered: