Update model with padded vocab size (1152) for GPU optimization
Browse files- config.json +2 -2
- model.safetensors +2 -2
config.json
CHANGED
@@ -18,6 +18,6 @@
|
|
18 |
"pad_token_id": 1,
|
19 |
"relative_attention_num_buckets": 32,
|
20 |
"torch_dtype": "float32",
|
21 |
-
"transformers_version": "4.
|
22 |
-
"vocab_size":
|
23 |
}
|
|
|
18 |
"pad_token_id": 1,
|
19 |
"relative_attention_num_buckets": 32,
|
20 |
"torch_dtype": "float32",
|
21 |
+
"transformers_version": "4.52.4",
|
22 |
+
"vocab_size": 1152
|
23 |
}
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a1042ac85f65a758a9a0f25c76e3c1f69c7ae8abf8999f183595e9db4cc7ea58
|
3 |
+
size 962704
|