Pinkstack commited on
Commit
83d59e5
·
verified ·
1 Parent(s): d7c845b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -23,7 +23,7 @@ As per the [Roblox website](https://create.roblox.com/docs/assistant/guide), the
23
 
24
  # print("Stages of pre-training")
25
 
26
- This model was pre-trained in 3 stages.
27
 
28
  - Stage 1: Pre-training on the Pinkstack/roblox-luau-corpus-text & Roblox/luau_corpus on 4096 tokens (the maximum olmo 2 can usually reach)
29
 
@@ -33,6 +33,8 @@ This model was pre-trained in 3 stages.
33
  !stage 3 and onwards were with added layers. the model started with 16 layers, then we merged another 20 to make the model bigger and deeper!
34
  - Stage 3: Training on a mix of Pinkstack/roblox-luau-corpus-text & Roblox/luau_corpus + wikimedia/wikipedia with rope scaling set to 8, aka **32768** tokens of context. We mixed the wikimedia/wikipedia to hopefully improve the general text and knowledge of the model.
35
 
 
 
36
  This repo contains the stage 3 pre-trained/base model.
37
 
38
  unsloth was used for training (https://unsloth.ai/)
 
23
 
24
  # print("Stages of pre-training")
25
 
26
+ This model was continually pre-trained in 3 stages.
27
 
28
  - Stage 1: Pre-training on the Pinkstack/roblox-luau-corpus-text & Roblox/luau_corpus on 4096 tokens (the maximum olmo 2 can usually reach)
29
 
 
33
  !stage 3 and onwards were with added layers. the model started with 16 layers, then we merged another 20 to make the model bigger and deeper!
34
  - Stage 3: Training on a mix of Pinkstack/roblox-luau-corpus-text & Roblox/luau_corpus + wikimedia/wikipedia with rope scaling set to 8, aka **32768** tokens of context. We mixed the wikimedia/wikipedia to hopefully improve the general text and knowledge of the model.
35
 
36
+ In total, the model was continually pre-trained on up to 1.3B tokens.
37
+ # print("Additional information")
38
  This repo contains the stage 3 pre-trained/base model.
39
 
40
  unsloth was used for training (https://unsloth.ai/)