DavidAU commited on
Commit
9e132bc
·
verified ·
1 Parent(s): 208b3d6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -11
README.md CHANGED
@@ -34,7 +34,7 @@ tags:
34
  pipeline_tag: text-generation
35
  ---
36
 
37
- <h2>Experimental Deepseek-R1-Llama3.1 with Brainstorm 40x, 16.5B. (72 layers, 643 tensors) </h2>
38
 
39
  <img src="deepseek.jpg" style="float:right; width:300px; height:300px; padding:10px;">
40
 
@@ -46,6 +46,8 @@ Keep in mind this model is experimental and may require one or more regens to wo
46
 
47
  Brainstorm 40x is by DavidAU, and extends the "decision making" and "creativity" of an LLM/AI.
48
 
 
 
49
  The "thinking/reasoning" tech (for the model at this repo) is from the original Llama 3.1 "Distill" model from Deepseek:
50
 
51
  [ https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B ]
@@ -64,16 +66,16 @@ The Grand Horrors retain all of their "horror/creative power" and are augmented
64
 
65
  <b>CRITICAL SETTINGS:</B>
66
 
67
- 1. Set Temp between 0 and .8, higher than this "think" functions will not activate. The most "stable" temp seems to be .6, with a variance of +-0.05. Lower for more "logic" reasoning, raise it for more "creative" reasoning (max .8 or so). Also set context to at least 4096, to account for "thoughts" generation.
68
- 2. Set "repeat penalty" to 1.02 to 1.08 and "repeat penalty range" to 64-128.
69
- 3. This model requires a Llama 3 Instruct and/or Command-R chat template. (see notes on "System Prompt" / "Role" below)
70
- 4. It may take one or more regens for "thinking" to "activate"... depending on your prompt.
71
- 5. If you enter a prompt without implied "step by step" requirements, "thinking" (one or more) will activate AFTER first generation. You will also get a lot of variations - some will continue the generation, others will talk about how to improve it, and some (ie generation of a scene) will cause the characters to "reason" about this situation. In some cases, the model will ask you to continue generation / thoughts too. In some cases the model's "thoughts" may appear in the generation itself.
72
- 6. State the word size length max IN THE PROMPT for best results, especially for activation of "thinking."
73
- 7. I have found opening a "new chat" per prompt works best with "thinking/reasoning activation", with temp .6
74
- 8. Depending on your AI app, "thoughts" may appear with "<THINK>" and "</THINK>" tags AND/OR the AI will generate "thoughts" directly in the main output or later output(s).
75
- 9. Although quant Q4KM was used for testing/examples, higher quants will provide better generation / more sound "reasoning/thinking".
76
- 10. To repeat: If you exceed temp of .8 or so, "thinking" processes may stop or change form or you will get "normal" model generation.
77
 
78
  ---
79
 
 
34
  pipeline_tag: text-generation
35
  ---
36
 
37
+ <h2>Deepseek-R1-Llama3.1 with Brainstorm 40x, 16.5B. (72 layers, 643 tensors) </h2>
38
 
39
  <img src="deepseek.jpg" style="float:right; width:300px; height:300px; padding:10px;">
40
 
 
46
 
47
  Brainstorm 40x is by DavidAU, and extends the "decision making" and "creativity" of an LLM/AI.
48
 
49
+ Higher temps will result in deeper, richer "thoughts"... and frankly more interesting ones too.
50
+
51
  The "thinking/reasoning" tech (for the model at this repo) is from the original Llama 3.1 "Distill" model from Deepseek:
52
 
53
  [ https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B ]
 
66
 
67
  <b>CRITICAL SETTINGS:</B>
68
 
69
+ 1. Set Temp between 0 and .8, higher than this "think" functions MAY not activate. The most "stable" temp seems to be .6, with a variance of +-0.05. Lower for more "logic" reasoning, raise it for more "creative" reasoning (max .8 or so). Also set context to at least 4096, to account for "thoughts" generation.
70
+ 2. Set "repeat penalty" to 1.02 to 1.08 and "repeat penalty range" to 64-128.
71
+ 3. Temps 1+, 2+ will deepen thoughts, conclusions, and generation thinking.
72
+ 4. This model requires a Llama 3 Instruct and/or Command-R chat template. (see notes on "System Prompt" / "Role" below)
73
+ 5. It may take one or more regens for "thinking" to "activate"... depending on your prompt.
74
+ 6. If you enter a prompt without implied "step by step" requirements, "thinking" (one or more) will activate AFTER first generation. You will also get a lot of variations - some will continue the generation, others will talk about how to improve it, and some (ie generation of a scene) will cause the characters to "reason" about this situation. In some cases, the model will ask you to continue generation / thoughts too. In some cases the model's "thoughts" may appear in the generation itself.
75
+ 7. State the word size length max IN THE PROMPT for best results, especially for activation of "thinking."
76
+ 8. I have found opening a "new chat" per prompt works best with "thinking/reasoning activation", with temp .6
77
+ 9. Depending on your AI app, "thoughts" may appear with "< THINK >" and "</ THINK >" tags AND/OR the AI will generate "thoughts" directly in the main output or later output(s).
78
+ 10. Although quant Q4KM was used for testing/examples, higher quants will provide better generation / more sound "reasoning/thinking".
79
 
80
  ---
81