Gogryu commited on
Commit
dce860f
Β·
1 Parent(s): f514bed

models update

Browse files
build/asset-manifest.json CHANGED
@@ -1,15 +1,15 @@
1
  {
2
  "files": {
3
  "main.css": "/static/css/main.6f7196f1.css",
4
- "main.js": "/static/js/main.26cb67d8.js",
5
  "static/media/logo-dark-mode.png": "/static/media/logo-dark-mode.0c51ff42111d2779453a.png",
6
  "static/media/logo-light-mode.png": "/static/media/logo-light-mode.6bc3b2d5ed1ec2d7d721.png",
7
  "index.html": "/index.html",
8
  "main.6f7196f1.css.map": "/static/css/main.6f7196f1.css.map",
9
- "main.26cb67d8.js.map": "/static/js/main.26cb67d8.js.map"
10
  },
11
  "entrypoints": [
12
  "static/css/main.6f7196f1.css",
13
- "static/js/main.26cb67d8.js"
14
  ]
15
  }
 
1
  {
2
  "files": {
3
  "main.css": "/static/css/main.6f7196f1.css",
4
+ "main.js": "/static/js/main.c56c182a.js",
5
  "static/media/logo-dark-mode.png": "/static/media/logo-dark-mode.0c51ff42111d2779453a.png",
6
  "static/media/logo-light-mode.png": "/static/media/logo-light-mode.6bc3b2d5ed1ec2d7d721.png",
7
  "index.html": "/index.html",
8
  "main.6f7196f1.css.map": "/static/css/main.6f7196f1.css.map",
9
+ "main.c56c182a.js.map": "/static/js/main.c56c182a.js.map"
10
  },
11
  "entrypoints": [
12
  "static/css/main.6f7196f1.css",
13
+ "static/js/main.c56c182a.js"
14
  ]
15
  }
build/index.html CHANGED
@@ -1 +1 @@
1
- <!doctype html><html lang="en"><head><meta charset="utf-8"/><link rel="icon" href="icon.png"/><meta name="viewport" content="width=device-width,initial-scale=1"/><meta name="theme-color" content="#000000"/><meta name="Titan Takeoff Server" content="Titan Takeoff Server"/><link rel="manifest" crossorigin="use-credentials" href="manifest.json"/><title>Titan Takeoff Server</title><script defer="defer" src="./static/js/main.26cb67d8.js"></script><link href="./static/css/main.6f7196f1.css" rel="stylesheet"></head><body><noscript>You need to enable JavaScript to run this app.</noscript><div id="root"></div></body></html>
 
1
+ <!doctype html><html lang="en"><head><meta charset="utf-8"/><link rel="icon" href="icon.png"/><meta name="viewport" content="width=device-width,initial-scale=1"/><meta name="theme-color" content="#000000"/><meta name="Titan Takeoff Server" content="Titan Takeoff Server"/><link rel="manifest" crossorigin="use-credentials" href="manifest.json"/><title>Titan Takeoff Server</title><script defer="defer" src="./static/js/main.c56c182a.js"></script><link href="./static/css/main.6f7196f1.css" rel="stylesheet"></head><body><noscript>You need to enable JavaScript to run this app.</noscript><div id="root"></div></body></html>
build/static/js/{main.26cb67d8.js β†’ main.c56c182a.js} RENAMED
The diff for this file is too large to render. See raw diff
 
build/static/js/{main.26cb67d8.js.LICENSE.txt β†’ main.c56c182a.js.LICENSE.txt} RENAMED
File without changes
build/static/js/{main.26cb67d8.js.map β†’ main.c56c182a.js.map} RENAMED
The diff for this file is too large to render. See raw diff
 
src/utils/models.ts CHANGED
@@ -7,90 +7,108 @@ type ModelInfo = {
7
  };
8
 
9
  export const MODELS: ModelInfo[] = [
 
10
  {
11
- name: 'LLaMA 2 (7B)',
12
- params: 7,
13
- hidden_size: 4096,
14
- intermediate_size: 11008,
15
- num_hidden_layers: 32,
16
  },
17
  {
18
- name: 'LLaMA 2 (13B)',
19
- params: 13,
20
- hidden_size: 5120,
21
- intermediate_size: 13824,
22
- num_hidden_layers: 40,
23
  },
24
  {
25
- name: 'LLaMA 2 (70B)',
 
 
 
 
 
 
 
 
 
26
  params: 70,
27
  hidden_size: 8192,
28
  intermediate_size: 28672,
29
  num_hidden_layers: 80,
30
  },
31
  {
32
- name: 'Mistral (13B)',
33
- params: 13,
34
  hidden_size: 4096,
35
  intermediate_size: 14336,
36
- num_hidden_layers: 60,
37
  },
 
 
38
  {
39
- name: 'Mistral (13B)',
 
 
 
 
 
 
 
40
  params: 13,
41
  hidden_size: 5120,
42
  intermediate_size: 13824,
43
  num_hidden_layers: 40,
44
  },
45
  {
46
- name: 'Mistral (7B)',
47
  params: 7,
48
  hidden_size: 4096,
49
- intermediate_size: 14336,
50
  num_hidden_layers: 32,
51
  },
 
 
52
  {
53
- name: 'Qwen (7B)',
54
- params: 7,
55
  hidden_size: 4096,
56
- intermediate_size: 22016,
57
- num_hidden_layers: 32,
58
  },
59
  {
60
- name: 'Qwen (1.5 7B)',
61
  params: 7,
62
  hidden_size: 4096,
63
- intermediate_size: 11008,
64
  num_hidden_layers: 32,
65
  },
66
  {
67
- name: 'Qwen (7B)',
68
- params: 7,
69
- hidden_size: 3584,
70
- intermediate_size: 18944,
71
- num_hidden_layers: 28,
72
  },
 
 
73
  {
74
  name: 'Qwen (7B)',
75
  params: 7,
76
  hidden_size: 4096,
77
- intermediate_size: 11008,
78
  num_hidden_layers: 32,
79
  },
80
  {
81
- name: 'Llava (7B)',
82
  params: 7,
83
  hidden_size: 4096,
84
  intermediate_size: 11008,
85
  num_hidden_layers: 32,
86
  },
87
- {
88
- name: 'Llava (1.5 13B)',
89
- params: 13,
90
- hidden_size: 5120,
91
- intermediate_size: 13824,
92
- num_hidden_layers: 40,
93
- },
94
  {
95
  name: 'Llava (1.6 34B)',
96
  params: 34,
@@ -99,19 +117,21 @@ export const MODELS: ModelInfo[] = [
99
  num_hidden_layers: 60,
100
  },
101
  {
102
- name: 'Gemma (2B)',
103
- params: 2,
104
- hidden_size: 2048,
105
- intermediate_size: 16384,
106
- num_hidden_layers: 18,
107
  },
108
  {
109
- name: 'Gemma (2.9B)',
110
- params: 2.9,
111
- hidden_size: 3584,
112
- intermediate_size: 14336,
113
- num_hidden_layers: 42,
114
  },
 
 
115
  {
116
  name: 'Gemma (27B)',
117
  params: 27,
@@ -120,33 +140,28 @@ export const MODELS: ModelInfo[] = [
120
  num_hidden_layers: 46,
121
  },
122
  {
123
- name: 'Mixtral (46B)',
124
- params: 46,
125
- hidden_size: 4096,
126
- intermediate_size: 14336,
127
- num_hidden_layers: 32,
128
  },
129
  {
130
- name: 'LLaMA 3 (8B)',
131
- params: 8,
132
- hidden_size: 4096,
133
- intermediate_size: 14336,
134
- num_hidden_layers: 32,
135
  },
 
 
136
  {
137
- name: 'LLaMA 3 (8B)',
138
- params: 8,
139
  hidden_size: 4096,
140
  intermediate_size: 14336,
141
  num_hidden_layers: 32,
142
  },
143
- {
144
- name: 'LLaMA 3 (70B)',
145
- params: 70,
146
- hidden_size: 8192,
147
- intermediate_size: 28672,
148
- num_hidden_layers: 80,
149
- },
150
  ];
151
 
152
  export default MODELS;
 
7
  };
8
 
9
  export const MODELS: ModelInfo[] = [
10
+ // LLaMA 3.1 Models
11
  {
12
+ name: 'LLaMA 3.1 (405B)',
13
+ params: 405,
14
+ hidden_size: 16384,
15
+ intermediate_size: 53248,
16
+ num_hidden_layers: 126,
17
  },
18
  {
19
+ name: 'LLaMA 3.1 (70B)',
20
+ params: 70,
21
+ hidden_size: 8192,
22
+ intermediate_size: 28672,
23
+ num_hidden_layers: 80,
24
  },
25
  {
26
+ name: 'LLaMA 3.1 (8B)',
27
+ params: 8,
28
+ hidden_size: 4096,
29
+ intermediate_size: 14336,
30
+ num_hidden_layers: 32,
31
+ },
32
+
33
+ // LLaMA 3 Models
34
+ {
35
+ name: 'LLaMA 3 (70B)',
36
  params: 70,
37
  hidden_size: 8192,
38
  intermediate_size: 28672,
39
  num_hidden_layers: 80,
40
  },
41
  {
42
+ name: 'LLaMA 3 (8B)',
43
+ params: 8,
44
  hidden_size: 4096,
45
  intermediate_size: 14336,
46
+ num_hidden_layers: 32,
47
  },
48
+
49
+ // LLaMA 2 Models
50
  {
51
+ name: 'LLaMA 2 (70B)',
52
+ params: 70,
53
+ hidden_size: 8192,
54
+ intermediate_size: 28672,
55
+ num_hidden_layers: 80,
56
+ },
57
+ {
58
+ name: 'LLaMA 2 (13B)',
59
  params: 13,
60
  hidden_size: 5120,
61
  intermediate_size: 13824,
62
  num_hidden_layers: 40,
63
  },
64
  {
65
+ name: 'LLaMA 2 (7B)',
66
  params: 7,
67
  hidden_size: 4096,
68
+ intermediate_size: 11008,
69
  num_hidden_layers: 32,
70
  },
71
+
72
+ // Mistral Models
73
  {
74
+ name: 'Mistral (13B NeuralPivot)',
75
+ params: 13,
76
  hidden_size: 4096,
77
+ intermediate_size: 14336,
78
+ num_hidden_layers: 60,
79
  },
80
  {
81
+ name: 'Mistral (7B)',
82
  params: 7,
83
  hidden_size: 4096,
84
+ intermediate_size: 14336,
85
  num_hidden_layers: 32,
86
  },
87
  {
88
+ name: 'Mistral (13B Amethyst)',
89
+ params: 13,
90
+ hidden_size: 5120,
91
+ intermediate_size: 13824,
92
+ num_hidden_layers: 40,
93
  },
94
+
95
+ // Qwen Models
96
  {
97
  name: 'Qwen (7B)',
98
  params: 7,
99
  hidden_size: 4096,
100
+ intermediate_size: 22016,
101
  num_hidden_layers: 32,
102
  },
103
  {
104
+ name: 'Qwen (1.5 7B)',
105
  params: 7,
106
  hidden_size: 4096,
107
  intermediate_size: 11008,
108
  num_hidden_layers: 32,
109
  },
110
+
111
+ // Llava Models
 
 
 
 
 
112
  {
113
  name: 'Llava (1.6 34B)',
114
  params: 34,
 
117
  num_hidden_layers: 60,
118
  },
119
  {
120
+ name: 'Llava (1.5 13B)',
121
+ params: 13,
122
+ hidden_size: 5120,
123
+ intermediate_size: 13824,
124
+ num_hidden_layers: 40,
125
  },
126
  {
127
+ name: 'Llava (7B)',
128
+ params: 7,
129
+ hidden_size: 4096,
130
+ intermediate_size: 11008,
131
+ num_hidden_layers: 32,
132
  },
133
+
134
+ // Gemma Models
135
  {
136
  name: 'Gemma (27B)',
137
  params: 27,
 
140
  num_hidden_layers: 46,
141
  },
142
  {
143
+ name: 'Gemma (2.9B)',
144
+ params: 2.9,
145
+ hidden_size: 3584,
146
+ intermediate_size: 14336,
147
+ num_hidden_layers: 42,
148
  },
149
  {
150
+ name: 'Gemma (2B)',
151
+ params: 2,
152
+ hidden_size: 2048,
153
+ intermediate_size: 16384,
154
+ num_hidden_layers: 18,
155
  },
156
+
157
+ // Mixtral Models
158
  {
159
+ name: 'Mixtral (46B)',
160
+ params: 46,
161
  hidden_size: 4096,
162
  intermediate_size: 14336,
163
  num_hidden_layers: 32,
164
  },
 
 
 
 
 
 
 
165
  ];
166
 
167
  export default MODELS;