You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We're using LM-Harness to evaluate Qwen/Qwen2-VL-2B-Instruct on task_name: mmmu_val using 4 A100 GPUs w/ 40GB VRAM.
The evaluation starts and progresses OK until the middle of the dataset where it fails with CUDA OOM (out-of-memory). It looks like it may be caused by a single example in the dataset which requires more GPU memory than usual, leading to OOM.
This Qwen2-VL OOM problem was resolved for training, by passing the following params as Qwen2-VL processor_kwargs:
The PR should let you pass min_pixels, max_pixels to the model args, but just so I don't forget, need to think of a better way of passing sub-method kwargs
We're using LM-Harness to evaluate
Qwen/Qwen2-VL-2B-Instruct
ontask_name: mmmu_val
using 4A100
GPUs w/ 40GB VRAM.The evaluation starts and progresses OK until the middle of the dataset where it fails with CUDA OOM (out-of-memory). It looks like it may be caused by a single example in the dataset which requires more GPU memory than usual, leading to OOM.
This Qwen2-VL OOM problem was resolved for training, by passing the following params as Qwen2-VL
processor_kwargs
:Context: https://huggingface.co/Qwen/Qwen2-VL-2B-Instruct#image-resolution-for-performance-boost
For LH-Harness evaluation, I haven't found a way to configure
processor_kwargs
forhf-multimodal
(onlymodel_args
seem configurable):lm-evaluation-harness/lm_eval/models/hf_vlms.py
Line 134 in 5c006ed
Could you please advise if setting
processor_kwargs
is supported? If not, would it be possible to add it forhf-multimodal
?The text was updated successfully, but these errors were encountered: