-
Notifications
You must be signed in to change notification settings - Fork 48
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Separate doc filtration and reasoning LLMs #1
Comments
you can pass in which model you want in theory. And than route this to some reasoning model for |
Feel free to open a PR if I misunderstood you! |
jasonkneen
referenced
this issue
in jasonkneen/reag
Feb 3, 2025
Fixes #1 Separate the document filtration and reasoning LLMs into two distinct models. * **Python Changes:** - Add a new parameter `filtration_model` to the `ReagClient` constructor in `python/src/reag/client.py`. - Update the `query` method to use the `filtration_model` for document filtration and the `model` for generating the final answer. - Update `python/README.md` to reflect the changes in the `ReagClient` class and the new two-step process. * **TypeScript Changes:** - Add a new parameter `filtrationModel` to the `ReagClient` constructor in `typescript/src/client.ts`. - Update the `query` method to use the `filtrationModel` for document filtration and the `model` for generating the final answer. - Update `typescript/README.md` to reflect the changes in the `ReagClient` class and the new two-step process. --- For more details, open the [Copilot Workspace session](https://copilot-workspace.githubnext.com/superagent-ai/reag/issues/1?shareId=XXXX-XXXX-XXXX-XXXX).
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I think it would be pretty powerful to separate this into two separate LLMs. Something like minimax or Gemini 2.0 can do the document pipeline (huge context window and very cheap), then only those documents (or perhaps summarizations of those documents) go to a smart/reasoning LLM for the final answer
The text was updated successfully, but these errors were encountered: