Is it possible to use qwen_vl_2.5_32b as clip? Will the performance improve.

#14
by codexq - opened

As the text_encoder folder provide qwen_vl_2.5_7b, will qwen_vl_2.5_32b do a better work?

Since the pre-learning was originally done with "qwen_vl_2.5_7b", there is no reason to change it.

did you find an answer? I'm curious as well but it doesn't seem to work when I just swap the 7b clip for 32b

Sign up or log in to comment