DavidAU commited on
Commit
a038dba
·
verified ·
1 Parent(s): e54ff31

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -3
README.md CHANGED
@@ -36,7 +36,8 @@ language:
36
 
37
  <img src="deadpan-savant.gif" style="float:right; width:300px; height:300px; padding:10px;">
38
 
39
- Savant Commander is a specialized MOE model that allows you to control which expert(s) are assigned to your use case(s) / prompt(s).
 
40
 
41
  The model is composed of 12 DISTILLS (compressed 12x4B MOE) of top closed (GPT5.1, OpenAI 120 GPT Oss, Gemini (3), Claude (2) )
42
  and open source models (Kimi, GLM, Deepseek, Command-A, Jan) all in one.
@@ -81,15 +82,27 @@ Note the name[s] must be in the prompt and/or the system role and can be located
81
 
82
  For best results suggest using the name(s) at the beginning as a "command" / "request" :
83
 
84
-
85
  GLM do ...
86
 
87
  Using Gemini process this prompt:
88
 
89
- However, using the name[s] in the prompt will work in most cases as that is what is being "scanned for" during gating.
 
 
90
 
91
  You MAY want to increase the number of active experts in some cases.
92
 
 
 
 
 
 
 
 
 
 
 
 
93
  <B>IMPORTANT: </B>
94
 
95
  - Min Quant of Q4ks (non imatrix) or IQ3_M (imatrix) ; otherwise it will "snap".
 
36
 
37
  <img src="deadpan-savant.gif" style="float:right; width:300px; height:300px; padding:10px;">
38
 
39
+ Savant Commander is a specialized MOE model that allows you to control which expert(s) are assigned to your use case(s) / prompt(s) ...
40
+ directly (by name(s)), as opposed to having the "choices" made for you.
41
 
42
  The model is composed of 12 DISTILLS (compressed 12x4B MOE) of top closed (GPT5.1, OpenAI 120 GPT Oss, Gemini (3), Claude (2) )
43
  and open source models (Kimi, GLM, Deepseek, Command-A, Jan) all in one.
 
82
 
83
  For best results suggest using the name(s) at the beginning as a "command" / "request" :
84
 
 
85
  GLM do ...
86
 
87
  Using Gemini process this prompt:
88
 
89
+ However, using the name[s] in the prompt will work in most cases as that is what is being "scanned for" during "prompt processing".
90
+
91
+ This model also has NEGATIVE gating to ensure other models not in use are ISOLATED.
92
 
93
  You MAY want to increase the number of active experts in some cases.
94
 
95
+ For trying the model out (example):
96
+
97
+ NAME, Tell me a horror story.
98
+
99
+ Use a different "name" per "new chat" - you will get different thought blocks, output etc etc - in some cases very different
100
+ from each other.
101
+
102
+ SUGGESTED SETTINGS to START:
103
+
104
+ Temp .7, topk 40, top p .95, min p .05, rep pen 1.05,
105
+
106
  <B>IMPORTANT: </B>
107
 
108
  - Min Quant of Q4ks (non imatrix) or IQ3_M (imatrix) ; otherwise it will "snap".