Skip to content

Commit

Permalink
[Bugfix] Fix TeleChat2ForCausalLM weights mapper (#11546)
Browse files Browse the repository at this point in the history
Signed-off-by: Jee Jee Li <[email protected]>
  • Loading branch information
jeejeelee authored Dec 27, 2024
1 parent d003f3e commit 2c9b8ea
Showing 1 changed file with 13 additions and 13 deletions.
26 changes: 13 additions & 13 deletions vllm/model_executor/models/telechat2.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,19 +31,6 @@

class TeleChat2Model(LlamaModel):

hf_to_vllm_mapper = WeightsMapper(
orig_to_new_prefix={
"transformer.": "model.",
},
orig_to_new_substr={
".h.": ".layers.",
".self_attention.": ".self_attn.",
".word_embeddings.": ".embed_tokens.",
".dense.": ".o_proj.",
".ln_f.": ".norm.",
},
)

def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""):
# 1. Initialize the LlamaModel with bias
vllm_config.model_config.hf_config.bias = True
Expand Down Expand Up @@ -118,6 +105,19 @@ def load_weights(self, weights: Iterable[Tuple[str,

class TeleChat2ForCausalLM(LlamaForCausalLM):

hf_to_vllm_mapper = WeightsMapper(
orig_to_new_prefix={
"transformer.": "model.",
},
orig_to_new_substr={
".h.": ".layers.",
".self_attention.": ".self_attn.",
".word_embeddings.": ".embed_tokens.",
".dense.": ".o_proj.",
".ln_f.": ".norm.",
},
)

def _init_model(self, vllm_config: VllmConfig, prefix: str = ""):
return TeleChat2Model(vllm_config=vllm_config, prefix=prefix)

Expand Down

0 comments on commit 2c9b8ea

Please sign in to comment.