You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In order to evaluate with CodeMMLU, you can hand me the LLM's response via email (follow format mentioned in HERE).
Since CodeMMLU is intended as a benchmark, not a training dataset, its purpose is to evaluate the performance of pre-trained LLMs, so it might not correctly serve fine-tune purpose.
The CodeMMLU is a great piece of work!
I noticed that the dataset provides task_id, question, and choices columns, but is there an answer column?
How should I handle this dataset if I want to fine-tune an LLM?
The text was updated successfully, but these errors were encountered: