[BugFix] Fix vocab size error for ernie model

This commit is contained in:
Jiang-Jia-Jun
2025-07-09 22:13:41 +08:00
parent da20cf681e
commit b1c53fa779

View File

@@ -735,7 +735,8 @@ def initialize_fd_config(config_or_args) -> FDConfig:
# Handle vocabulary size
model_config.ori_vocab_size = model_config_dict.get("vocab_size", -1)
if "Ernie4_5_ForCausalLM" in model_config_dict.get("architectures", []):
archs = model_config_dict.get("architectures", [])
if "Ernie4_5_ForCausalLM" in archs or "Ernie4_5_MoeForCausalLM" in archs:
model_config.ori_vocab_size = getattr(config_or_args, 'ori_vocab_size', model_config.ori_vocab_size)
# Handle DeepseekV3 specific config