-
Notifications
You must be signed in to change notification settings - Fork 85
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
【bug】can not load cambrian-34b #12
Comments
Hi, could you please provide more information about your case (e.g. the |
transformers in my env is 4.39 , why must transformers==4.37.0 in dependencies ? |
all dependencies is "== " , I wonder if ">" is OK? |
Our training and evaluation are mainly conducted with the specified versions and haven't been extensively tested with higher versions to ensure correctness. But I have tested to run the 34B model with transformers==4.39.0 and it works fine. Could you provide the information about |
A100*8 |
Loading checkpoint shards: 94%|███████████████████████████████████████████████████████████████████████████████████████████████████▍ | 30/32 [00:19<00:01, 1.57it/s] |
when i add some log , is loding "model.mm_projector_aux_0.0.weight" |
Is it because I only kept the second one in mm_vision_tower_aux_list? |
What do you mean by this? You don't need to modify the config if you want to load our trained model. |
because i can only load local path model,Can you list the huggingface download addresses for these four vision models? "mm_vision_tower_aux_list": [ |
For example, CLIP-ViT-SO400M-14-384 seems to have many versions, and I can't search clip-conv-xxL-multi-stage in huggfing face |
|
Hi, how can I set 2 48G gpus? 2024-06-30 15:21:12 PID=57 init.py:49 setup_logging() INFO → 'standard' logger initialized. |
This error seems not related to multiple GPUs. Make sure that all model files are downloaded correctly (e.g. |
@dionren Some of the vision encoders are not from But I have a workaround for your case with 2 48G gpus. This includes the following modifications:
|
I'm gonna try it out. Thanks a ton for your help and the awesome work you've done. It's truly impressive. |
in load_pretrained_model
model = CambrianLlamaForCausalLM.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3531, in from_pretrained
) = cls._load_pretrained_model(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3958, in _load_pretrained_model
new_error_msgs, offload_index, state_dict_index = _load_state_dict_into_meta_model(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 812, in _load_state_dict_into_meta_model
set_module_tensor_to_device(model, param_name, param_device, **set_module_kwargs)
File "/usr/local/lib/python3.10/dist-packages/accelerate/utils/modeling.py", line 348, in set_module_tensor_to_device
raise ValueError(
ValueError: Trying to set a tensor of shape torch.Size([1024, 1152]) in "weight" (which has shape torch.Size([1024, 1024])), this look incorrect.
The text was updated successfully, but these errors were encountered: