You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
E File "/home/nvme-share/home/zhangguangyao/ColossalAI/colossalai/shardformer/modeling/llama.py", line 530, in forward
E query_states, key_states = apply_rotary_pos_emb(query_states, key_states, cos, sin)
E File "/home/nvme-share/home/zhangguangyao/hf_transformers/src/transformers/models/llama/modeling_llama.py", line 206, in apply_rotary_pos_emb
E q_embed = (q * cos) + (rotate_half(q) * sin)
E RuntimeError: The size of tensor a (16) must match the size of tensor b (8) at non-singleton dimension 2
I have found out this failure is introduced after PR #5868 merged. Please take a look.
Environment
No response
The text was updated successfully, but these errors were encountered:
Is there an existing issue for this bug?
🐛 Describe the bug
Main repo test_shard_llama fails for these configs:
The failure message is :
I have found out this failure is introduced after PR #5868 merged. Please take a look.
Environment
No response
The text was updated successfully, but these errors were encountered: