You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In my opinion, LLaVA-Video introduces a high-quality dataset, and the model is still based on the LLaVA architecture. Therefore, I think VisionZip can be applied to LLaVA-Video.
I encountered an error, Could u help me?
File "lmms-eval-main/VisionZip/visionzip/llava_arch.py", line 43, in encode_images_visionzip_multi
image_features, keep_idx = self.get_model().get_vision_tower().forward(images)
File "/root/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1688, in getattr
raise AttributeError(f"'{type(self).name}' object has no attribute '{name}'")
AttributeError: 'ClipEncoderLayer' object has no attribute 'metric
No description provided.
The text was updated successfully, but these errors were encountered: