-
Notifications
You must be signed in to change notification settings - Fork 2.1k
Issues: EleutherAI/lm-evaluation-harness
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Encountering assert len(indices) == len(inputs) error when using Qwen2vl for MMMU evaluation
#2720
opened Feb 21, 2025 by
Ben81828
How to preprocess a document with the assistance of a tokenizer from a specific Model
#2717
opened Feb 20, 2025 by
p1nksnow
Different models on same tasks gives same results when cache is active
bug
Something isn't working.
#2715
opened Feb 19, 2025 by
salvatore-cipolla
Importing a local module in a task included with include_path
#2713
opened Feb 19, 2025 by
joaormfsilva
[Accuracy gap with official model card due to wrong parsing]
#2707
opened Feb 17, 2025 by
Monstertail
Inconsistent Behavior with max_tokens, Post-Processing, and Cache Options
#2702
opened Feb 15, 2025 by
ntlm1686
vLLM CUDA OOM for For asking for clarification / support on library usage.
loglikelihood
, but not for generate_until
asking questions
#2698
opened Feb 14, 2025 by
lsjlsj5846
Feature request: allow peft revision separate from base model revision
#2696
opened Feb 13, 2025 by
iuliaturc
Strip the input for the three tasks: FDA, SWDE, and SQuAD_completion.
validation
For validation of task implementations.
#2690
opened Feb 12, 2025 by
Doraemonzzz
add_bos_token causes very unstable results for quantized llama3-70B
asking questions
For asking for clarification / support on library usage.
#2676
opened Feb 7, 2025 by
wenhuach21
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.