You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
-**[NEW!!!] Local 671B DeepSeek-Coder-V3/R1:** Running its Q4_K_M version using only 14GB VRAM and 382GB DRAM([Tutorial](./doc/en/DeepseekR1_V3_tutorial.md)).
Copy file name to clipboardExpand all lines: csrc/ktransformers_ext/operators/llamafile/moe.h
+3-2Lines changed: 3 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -32,6 +32,7 @@ struct MOEConfig {
32
32
int stride;
33
33
int group_min_len;
34
34
int group_max_len;
35
+
bool use_silu;
35
36
void* gate_proj;
36
37
void* up_proj;
37
38
void* down_proj;
@@ -42,8 +43,8 @@ struct MOEConfig {
42
43
43
44
MOEConfig() {}
44
45
45
-
MOEConfig(int expert_num, int routed_expert_num, int hidden_size, int intermediate_size, int stride, int group_min_len, int group_max_len, void* gate_proj, void* up_proj, void* down_proj, ggml_type gate_type, ggml_type up_type, ggml_type down_type, ggml_type hidden_type)
MOEConfig(int expert_num, int routed_expert_num, int hidden_size, int intermediate_size, int stride, int group_min_len, int group_max_len, bool use_silu, void* gate_proj, void* up_proj, void* down_proj, ggml_type gate_type, ggml_type up_type, ggml_type down_type, ggml_type hidden_type)
0 commit comments