Update model card with revised readme
Browse files
README.md
CHANGED
@@ -1,14 +1,23 @@
|
|
1 |
# QuantaMaths: `sub_d10_l2_h3_t75K_gf_s173289`
|
2 |
|
3 |
-
|
4 |
|
5 |
-
-
|
6 |
-
- **
|
7 |
-
- **
|
8 |
-
- **
|
9 |
-
- **
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
|
11 |
-
|
|
|
|
|
|
|
12 |
|
13 |
**Folder name details**:
|
14 |
- "add", "sub", or "mix": The types of questions the model can predict.
|
@@ -18,18 +27,4 @@ This repository contains a transformer model that can predict addition questions
|
|
18 |
- "t15K" to "t85K", etc.: The number of batches the model was trained on.
|
19 |
- "s372001", etc.: The random seed used in model training.
|
20 |
|
21 |
-
Some folder names also contain:
|
22 |
-
- "ins1": Before training, the model was initialized with a smaller, accurate addition model.
|
23 |
-
- "ins2": Same as ins1, but the inserted attention heads were not allowed to change.
|
24 |
-
- "ins3": Same as ins2, but the inserted MLP layers were also not allowed to change.
|
25 |
|
26 |
-
**Contents**:
|
27 |
-
- `model.pth`: The trained transformer model.
|
28 |
-
- `training_loss.json`: Data gathered during model training (used to plot "loss over training batches").
|
29 |
-
- `behaviors.json`: Facts gathered about the model by direct inspection (attention pattern data, PCA data, digit impact data, etc.).
|
30 |
-
- `features.json`: Facts gathered about hypothesized algorithm features via experimentation, e.g. node P12L0H1 implements the feature A3.ST.
|
31 |
-
|
32 |
-
**Provenance**:
|
33 |
-
- `model.pth` and `training_loss.json` were created by [QuantaMathsTrain.ipynb](https://github.com/PhilipQuirke/quanta_maths/blob/main/notebooks/QuantaMathsTrain.ipynb).
|
34 |
-
- `behaviors.json` and `features.json` were created by [QuantaMathsAnalyse.ipynb](https://github.com/PhilipQuirke/quanta_maths/blob/main/notebooks/QuantaMathsAnalyse.ipynb).
|
35 |
-
- The JSON files are used by [QuantaMathsAlgorithm.ipynb](https://github.com/PhilipQuirke/quanta_maths/blob/main/notebooks/QuantaMathsAlgorithm.ipynb).
|
|
|
1 |
# QuantaMaths: `sub_d10_l2_h3_t75K_gf_s173289`
|
2 |
|
3 |
+
This repository contains a transformer model that can predict subtraction questions.
|
4 |
|
5 |
+
### Model-specific metadata
|
6 |
+
- **Operation type**: subtraction
|
7 |
+
- **Num digits**: 10
|
8 |
+
- **Layers**: 2
|
9 |
+
- **Attention Heads**: 3
|
10 |
+
- **Training steps**: 75,000
|
11 |
+
**Contents**:
|
12 |
+
- `model.pth`: The trained transformer model.
|
13 |
+
- `training_loss.json`: Data gathered during model training (used to plot "loss over training batches").
|
14 |
+
- `behaviors.json`: Facts gathered about the model by direct inspection (attention pattern data, PCA data, digit impact data, etc.).
|
15 |
+
- `features.json`: Facts gathered about hypothesized algorithm features via experimentation, e.g. node P12L0H1 implements the feature A3.ST.
|
16 |
|
17 |
+
**Provenance**:
|
18 |
+
- `model.pth` and `training_loss.json` were created by [QuantaMathsTrain.ipynb](https://github.com/PhilipQuirke/quanta_maths/blob/main/notebooks/QuantaMathsTrain.ipynb).
|
19 |
+
- `behaviors.json` and `features.json` were created by [QuantaMathsAnalyse.ipynb](https://github.com/PhilipQuirke/quanta_maths/blob/main/notebooks/QuantaMathsAnalyse.ipynb).
|
20 |
+
- The JSON files are used by [QuantaMathsAlgorithm.ipynb](https://github.com/PhilipQuirke/quanta_maths/blob/main/notebooks/QuantaMathsAlgorithm.ipynb).
|
21 |
|
22 |
**Folder name details**:
|
23 |
- "add", "sub", or "mix": The types of questions the model can predict.
|
|
|
27 |
- "t15K" to "t85K", etc.: The number of batches the model was trained on.
|
28 |
- "s372001", etc.: The random seed used in model training.
|
29 |
|
|
|
|
|
|
|
|
|
30 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|