|
5 | 5 | #include <map> |
6 | 6 |
|
7 | 7 | static const std::map<llm_arch, const char *> LLM_ARCH_NAMES = { |
| 8 | + { LLM_ARCH_CLIP, "clip" }, // dummy, only used by llama-quantize |
8 | 9 | { LLM_ARCH_LLAMA, "llama" }, |
9 | 10 | { LLM_ARCH_LLAMA4, "llama4" }, |
10 | 11 | { LLM_ARCH_DECI, "deci" }, |
@@ -84,6 +85,7 @@ static const std::map<llm_arch, const char *> LLM_ARCH_NAMES = { |
84 | 85 | { LLM_ARCH_WAVTOKENIZER_DEC, "wavtokenizer-dec" }, |
85 | 86 | { LLM_ARCH_PLM, "plm" }, |
86 | 87 | { LLM_ARCH_BAILINGMOE, "bailingmoe" }, |
| 88 | + { LLM_ARCH_BAILINGMOE2, "bailingmoe2" }, |
87 | 89 | { LLM_ARCH_DOTS1, "dots1" }, |
88 | 90 | { LLM_ARCH_ARCEE, "arcee" }, |
89 | 91 | { LLM_ARCH_ERNIE4_5, "ernie4_5" }, |
@@ -134,6 +136,8 @@ static const std::map<llm_kv, const char *> LLM_KV_NAMES = { |
134 | 136 | { LLM_KV_EXPERT_COUNT, "%s.expert_count" }, |
135 | 137 | { LLM_KV_EXPERT_USED_COUNT, "%s.expert_used_count" }, |
136 | 138 | { LLM_KV_EXPERT_SHARED_COUNT, "%s.expert_shared_count" }, |
| 139 | + { LLM_KV_EXPERT_GROUP_COUNT, "%s.expert_group_count" }, |
| 140 | + { LLM_KV_EXPERT_GROUP_USED_COUNT, "%s.expert_group_used_count" }, |
137 | 141 | { LLM_KV_EXPERT_WEIGHTS_SCALE, "%s.expert_weights_scale" }, |
138 | 142 | { LLM_KV_EXPERT_WEIGHTS_NORM, "%s.expert_weights_norm" }, |
139 | 143 | { LLM_KV_EXPERT_GATING_FUNC, "%s.expert_gating_func" }, |
@@ -275,6 +279,10 @@ static const std::map<llm_kv, const char *> LLM_KV_NAMES = { |
275 | 279 | }; |
276 | 280 |
|
277 | 281 | static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_NAMES = { |
| 282 | + { |
| 283 | + LLM_ARCH_CLIP, |
| 284 | + {}, |
| 285 | + }, |
278 | 286 | { |
279 | 287 | LLM_ARCH_LLAMA, |
280 | 288 | { |
@@ -1941,6 +1949,38 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N |
1941 | 1949 | { LLM_TENSOR_FFN_UP_SHEXP, "blk.%d.ffn_up_shexp" }, |
1942 | 1950 | }, |
1943 | 1951 | }, |
| 1952 | + { |
| 1953 | + LLM_ARCH_BAILINGMOE2, |
| 1954 | + { |
| 1955 | + { LLM_TENSOR_TOKEN_EMBD, "token_embd" }, |
| 1956 | + { LLM_TENSOR_OUTPUT_NORM, "output_norm" }, |
| 1957 | + { LLM_TENSOR_OUTPUT, "output" }, |
| 1958 | + { LLM_TENSOR_ATTN_NORM, "blk.%d.attn_norm" }, |
| 1959 | + { LLM_TENSOR_ATTN_Q_NORM, "blk.%d.attn_q_norm" }, |
| 1960 | + { LLM_TENSOR_ATTN_K_NORM, "blk.%d.attn_k_norm" }, |
| 1961 | + { LLM_TENSOR_ATTN_QKV, "blk.%d.attn_qkv" }, |
| 1962 | + { LLM_TENSOR_ATTN_OUT, "blk.%d.attn_output" }, |
| 1963 | + { LLM_TENSOR_FFN_GATE_INP, "blk.%d.ffn_gate_inp" }, |
| 1964 | + { LLM_TENSOR_FFN_EXP_PROBS_B, "blk.%d.exp_probs_b" }, |
| 1965 | + { LLM_TENSOR_FFN_NORM, "blk.%d.ffn_norm" }, |
| 1966 | + { LLM_TENSOR_FFN_GATE, "blk.%d.ffn_gate" }, |
| 1967 | + { LLM_TENSOR_FFN_DOWN, "blk.%d.ffn_down" }, |
| 1968 | + { LLM_TENSOR_FFN_UP, "blk.%d.ffn_up" }, |
| 1969 | + { LLM_TENSOR_FFN_GATE_EXPS, "blk.%d.ffn_gate_exps" }, |
| 1970 | + { LLM_TENSOR_FFN_DOWN_EXPS, "blk.%d.ffn_down_exps" }, |
| 1971 | + { LLM_TENSOR_FFN_UP_EXPS, "blk.%d.ffn_up_exps" }, |
| 1972 | + { LLM_TENSOR_FFN_GATE_SHEXP, "blk.%d.ffn_gate_shexp" }, |
| 1973 | + { LLM_TENSOR_FFN_DOWN_SHEXP, "blk.%d.ffn_down_shexp" }, |
| 1974 | + { LLM_TENSOR_FFN_UP_SHEXP, "blk.%d.ffn_up_shexp" }, |
| 1975 | + { LLM_TENSOR_NEXTN_EH_PROJ, "blk.%d.nextn.eh_proj" }, |
| 1976 | + { LLM_TENSOR_NEXTN_EMBED_TOKENS, "blk.%d.nextn.embed_tokens" }, |
| 1977 | + { LLM_TENSOR_NEXTN_ENORM, "blk.%d.nextn.enorm" }, |
| 1978 | + { LLM_TENSOR_NEXTN_HNORM, "blk.%d.nextn.hnorm" }, |
| 1979 | + { LLM_TENSOR_NEXTN_SHARED_HEAD_HEAD, "blk.%d.nextn.shared_head_head" }, |
| 1980 | + { LLM_TENSOR_NEXTN_SHARED_HEAD_NORM, "blk.%d.nextn.shared_head_norm" }, |
| 1981 | + { LLM_TENSOR_LAYER_OUT_NORM, "blk.%d.layer_output_norm" }, |
| 1982 | + }, |
| 1983 | + }, |
1944 | 1984 | { |
1945 | 1985 | LLM_ARCH_DOTS1, |
1946 | 1986 | { |
|
0 commit comments