vllm.model_executor.models.telechat2
TeleChat2ForCausalLM
¶
Bases: LlamaForCausalLM
Source code in vllm/model_executor/models/telechat2.py
hf_to_vllm_mapper
class-attribute
instance-attribute
¶
hf_to_vllm_mapper = WeightsMapper(
orig_to_new_prefix={"transformer.": "model."},
orig_to_new_substr={
".h.": ".layers.",
".self_attention.": ".self_attn.",
".word_embeddings.": ".embed_tokens.",
".dense.": ".o_proj.",
".ln_f.": ".norm.",
},
)
_init_model
¶
_init_model(
vllm_config: VllmConfig,
prefix: str = "",
layer_type: type[Module] = LlamaDecoderLayer,
)
load_weights
¶
Source code in vllm/model_executor/models/telechat2.py
TeleChat2Model
¶
Bases: LlamaModel
Source code in vllm/model_executor/models/telechat2.py
__init__
¶
__init__(*, vllm_config: VllmConfig, prefix: str = '')