Loading Error: Unrecognized model type 'qwen2_5_vl'

#10
by kentrolla - opened

I'm getting a ValueError when loading the model:
ValueError: Unrecognized configuration class <class 'transformers.models.qwen2_5_vl.configuration_qwen2_5_vl.Qwen2_5_VLConfig'> for this kind of AutoModel: AutoModelForCausalLM.

I've followed the instructions, installed transformers and accelerate from the provided GitHub repo, and set trust_remote_code=True, but the error persists.

Is there something specific I might've missed or an additional step required?

Thanks!

Qwen org

Hello, Qwen2_5_VL does not belong to AutoModelForCausalLM and cannot be directly loaded via AutoModel. Instead, it should be loaded using the specific class as shown below:

from transformers import Qwen2_5_VLForConditionalGeneration, AutoTokenizer, AutoProcessor
from qwen_vl_utils import process_vision_info

# Default: Load the model on the available device(s)
model = Qwen2_5_VLForConditionalGeneration.from_pretrained(
    "Qwen/Qwen2.5-VL-32B-Instruct", torch_dtype="auto", device_map="auto"
)

# We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
# model = Qwen2_5_VLForConditionalGeneration.from_pretrained(
#     "Qwen/Qwen2.5-VL-32B-Instruct",
#     torch_dtype=torch.bfloat16,
#     attn_implementation="flash_attention_2",
#     device_map="auto",
# )

Thank you!

kentrolla changed discussion status to closed
Your need to confirm your account before you can post a new comment.

Sign up or log in to comment