Triangle104 commited on
Commit
55030c4
·
verified ·
1 Parent(s): a5df84b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +88 -0
README.md CHANGED
@@ -11,6 +11,94 @@ tags:
11
  This model was converted to GGUF format from [`arcee-ai/Arcee-Maestro-7B-Preview`](https://huggingface.co/arcee-ai/Arcee-Maestro-7B-Preview) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
12
  Refer to the [original model card](https://huggingface.co/arcee-ai/Arcee-Maestro-7B-Preview) for more details on the model.
13
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  ## Use with llama.cpp
15
  Install llama.cpp through brew (works on Mac and Linux)
16
 
 
11
  This model was converted to GGUF format from [`arcee-ai/Arcee-Maestro-7B-Preview`](https://huggingface.co/arcee-ai/Arcee-Maestro-7B-Preview) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
12
  Refer to the [original model card](https://huggingface.co/arcee-ai/Arcee-Maestro-7B-Preview) for more details on the model.
13
 
14
+ ---
15
+ Arcee-Maestro-7B-Preview (7B) is Arcee's first reasoning model trained with reinforment learning. It is based on the Qwen2.5-7B DeepSeek-R1 distillation DeepSeek-R1-Distill-Qwen-7B
16
+ with further GPRO training. Though this is just a preview of our
17
+ upcoming work, it already shows promising improvements to mathematical
18
+ and coding abilities across a range of tasks.
19
+
20
+ Intended Use Cases
21
+ -
22
+
23
+ Advanced reasoning
24
+
25
+ Mathematics
26
+
27
+ Coding
28
+
29
+ Training & Fine-Tuning
30
+ -
31
+
32
+
33
+
34
+ Initial Training: Began with DeepSeek-R1-Distill-Qwen-7B
35
+ GRPO:
36
+ Trained on 450,000 verified math problems
37
+ Additional bootstrapped coding examples
38
+
39
+
40
+
41
+
42
+
43
+
44
+
45
+
46
+
47
+ Performance
48
+ -
49
+
50
+
51
+
52
+ Arcee-Maestro-7B-Preview shows strong performance in mathematics as
53
+ well as coding, competing against even O1 preview, a model far
54
+ surprassing its size.
55
+
56
+
57
+
58
+
59
+
60
+
61
+
62
+ Limitations
63
+ -
64
+
65
+
66
+
67
+ Context Length: 128k Tokens (may vary depending on the final tokenizer settings and system resources).
68
+ Knowledge Cut-off: Training data may not reflect the latest events or developments beyond June 2024.
69
+
70
+
71
+
72
+
73
+
74
+
75
+
76
+ Ethical Considerations
77
+ -
78
+
79
+
80
+
81
+ Content Generation Risks: Like any language model,
82
+ Arcee-Maestro-7B-Preview can generate potentially harmful or biased
83
+ content if prompted in certain ways.
84
+
85
+
86
+
87
+
88
+
89
+
90
+
91
+ License
92
+ -
93
+
94
+
95
+
96
+ Arcee-Maestro-7B-Preview (7B) is released under the Apache-2.0 License.
97
+ You are free to use, modify, and distribute this model in both
98
+ commercial and non-commercial applications, subject to the terms and
99
+ conditions of the license.
100
+
101
+ ---
102
  ## Use with llama.cpp
103
  Install llama.cpp through brew (works on Mac and Linux)
104