Skip to content

Commit 97cc872

Browse files
[Model] Ignore rotary embed load for Cohere model (#17319)
1 parent 4464109 commit 97cc872

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

vllm/model_executor/models/commandr.py

+4
Original file line numberDiff line numberDiff line change
@@ -418,6 +418,10 @@ def load_weights(self, weights: Iterable[Tuple[str,
418418
loaded_params: Set[str] = set()
419419
for name, loaded_weight in weights:
420420

421+
# Skip loading rotary embeddings since vLLM has its own
422+
if "rotary_emb.inv_freq" in name:
423+
continue
424+
421425
if (self.quant_config is not None and
422426
(scale_name := self.quant_config.get_cache_scale(name))):
423427
# Loading kv cache quantization scales

0 commit comments

Comments
 (0)