[Model] Provide clearer error for missing KV cache quantization scales (#3007)

This commit is contained in:
littledgg
2025-07-24 20:15:00 +08:00
committed by GitHub
parent c40df1802e
commit f37d00e856
3 changed files with 3 additions and 0 deletions

View File

@@ -420,6 +420,7 @@ class DeepseekV3MLAAttention(nn.Layer):
# NOTE(Ryan):Make sure kv_b_proj_bmm loaded before kv_b_proj,
# The same weight key will be poped after kv_b_proj.
self.o_proj.load_state_dict(state_dict)
self.mla_attn.load_state_dict(state_dict)
class DeepSeekV3DecoderLayer(nn.Layer):