-
Notifications
You must be signed in to change notification settings - Fork 2.3k
Issues: EleutherAI/lm-evaluation-harness
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
TypeError: top_k_top_p_sampling_from_probs() missing 1 required positional argument: 'top_p'
#2936
opened Apr 28, 2025 by
rangehow
how do evaluate mmlu by local-chat-completions or local-completions?
#2934
opened Apr 26, 2025 by
amdslgl
Some LongBench scores on Llama-3.1-8B-Instruct are far off
#2932
opened Apr 25, 2025 by
cameronshinn
GPQA Preprocessing Function Results in Incorrect Physics Equations
bug
Something isn't working.
validation
For validation of task implementations.
#2907
opened Apr 14, 2025 by
ShayekhBinIslam
Filter not extracting choice selection correctly
validation
For validation of task implementations.
#2905
opened Apr 14, 2025 by
1jamesthompson1
Does lm-eval currently support testing for models in the deekseek r1 category?
#2903
opened Apr 13, 2025 by
Polarisamoon
Improve the behavior of progress bars when using huggingface
#2898
opened Apr 11, 2025 by
Zephyr271828
RuntimeError: 500 Server Error for URL During LM-Eval with gguf Model
#2894
opened Apr 9, 2025 by
amjh83
TypeError
from missing yaml_path
in lm_eval.utils.load_yaml_config
when task uses include
#2884
opened Apr 6, 2025 by
MarieRoald
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.