We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4af4c77 commit b077bb5Copy full SHA for b077bb5
src/transformers/models/llama4/configuration_llama4.py
@@ -142,8 +142,6 @@ class Llama4TextConfig(PretrainedConfig):
142
"layers.*.feed_forward.experts.gate_up_proj": "local_packed_rowwise", # row because not linear
143
"layers.*.feed_forward.experts.down_proj": "local_colwise", # col because not linear
144
"layers.*.feed_forward.experts": "local",
145
- "layers.*.feed_forward.shared_expert.gate_proj": "local_colwise",
146
- "layers.*.feed_forward.shared_expert.up_proj": "local_colwise",
147
"layers.*.feed_forward.down_proj": "local_rowwise",
148
"layers.*.feed_forward": "gather",
149
}
0 commit comments