Skip to content

Issues: EleutherAI/lm-evaluation-harness

reproduce llama 3 evals
#2557 opened Dec 10, 2024 by baberabb
Open 6
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

vLLM CUDA OOM for loglikelihood, but not for generate_until asking questions For asking for clarification / support on library usage.
#2698 opened Feb 14, 2025 by lsjlsj5846
Support Arabic Dataset
#2693 opened Feb 13, 2025 by ziadwaelai
Strip the input for the three tasks: FDA, SWDE, and SQuAD_completion. validation For validation of task implementations.
#2690 opened Feb 12, 2025 by Doraemonzzz
Add o3-mini support
#2685 opened Feb 11, 2025 by HelloJocelynLu
add_bos_token causes very unstable results for quantized llama3-70B asking questions For asking for clarification / support on library usage.
#2676 opened Feb 7, 2025 by wenhuach21
Use AWS Bedrock Models
#2669 opened Feb 3, 2025 by nrcoleman
maximum sequence length
#2657 opened Jan 27, 2025 by Raghadalr02
List of num_fewshots
#2656 opened Jan 25, 2025 by AMindToThink
Question about humaneval
#2648 opened Jan 22, 2025 by Shiguang-Guo
add test for main.py
#2639 opened Jan 20, 2025 by baberabb
ProTip! Find all open issues with in progress development work with linked:pr.