sc2582 commited on
Commit
a8c60c8
·
verified ·
1 Parent(s): 2e69787

upload checkpoint

Browse files
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "_name_or_path": "save/iqlearn/250114_143059/iter1_meta-llama-Llama-3.2-3B-Instruct//checkpoint-500",
3
  "architectures": [
4
- "LlamaForSequenceClassification"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
@@ -14,14 +14,8 @@
14
  "head_dim": 128,
15
  "hidden_act": "silu",
16
  "hidden_size": 3072,
17
- "id2label": {
18
- "0": "LABEL_0"
19
- },
20
  "initializer_range": 0.02,
21
  "intermediate_size": 8192,
22
- "label2id": {
23
- "LABEL_0": 0
24
- },
25
  "max_position_embeddings": 131072,
26
  "mlp_bias": false,
27
  "model_type": "llama",
@@ -41,6 +35,6 @@
41
  "tie_word_embeddings": true,
42
  "torch_dtype": "bfloat16",
43
  "transformers_version": "4.47.1",
44
- "use_cache": true,
45
  "vocab_size": 128256
46
  }
 
1
  {
2
+ "_name_or_path": "rl-llm-agent/Llama-3.2-3B-Instruct-online-dpo-alfworld-iqlearn-iter0",
3
  "architectures": [
4
+ "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
 
14
  "head_dim": 128,
15
  "hidden_act": "silu",
16
  "hidden_size": 3072,
 
 
 
17
  "initializer_range": 0.02,
18
  "intermediate_size": 8192,
 
 
 
19
  "max_position_embeddings": 131072,
20
  "mlp_bias": false,
21
  "model_type": "llama",
 
35
  "tie_word_embeddings": true,
36
  "torch_dtype": "bfloat16",
37
  "transformers_version": "4.47.1",
38
+ "use_cache": false,
39
  "vocab_size": 128256
40
  }
generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 128000,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 128001,
6
+ 128008,
7
+ 128009
8
+ ],
9
+ "temperature": 0.6,
10
+ "top_p": 0.9,
11
+ "transformers_version": "4.47.1"
12
+ }
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e910e5565a2ac0f108bc7bd410d2778983c95a7b68dd8e06f5ed9ba8747def6e
3
  size 4965841415
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d0dd76b00345d23331d8e0f62431fa34ac8d795c36f47ffe8114501f3cf22ae
3
  size 4965841415
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b65246382a3c2003f7b85abbdd6b366a1b9730986fbf405d9933192eaf99d78
3
- size 1459751666
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad778ecd5db56dfdbefeca17f5c7ff9b67c4ed8d1ece2b6330a1f073c9f4a772
3
+ size 2247750386
pytorch_model.bin.index.json CHANGED
@@ -1,8 +1,9 @@
1
  {
2
  "metadata": {
3
- "total_size": 6425505792
4
  },
5
  "weight_map": {
 
6
  "model.embed_tokens.weight": "pytorch_model-00001-of-00002.bin",
7
  "model.layers.0.input_layernorm.weight": "pytorch_model-00001-of-00002.bin",
8
  "model.layers.0.mlp.down_proj.weight": "pytorch_model-00001-of-00002.bin",
@@ -256,7 +257,6 @@
256
  "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00001-of-00002.bin",
257
  "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00001-of-00002.bin",
258
  "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
259
- "model.norm.weight": "pytorch_model-00002-of-00002.bin",
260
- "score.weight": "pytorch_model-00002-of-00002.bin"
261
  }
262
  }
 
1
  {
2
  "metadata": {
3
+ "total_size": 7213504512
4
  },
5
  "weight_map": {
6
+ "lm_head.weight": "pytorch_model-00002-of-00002.bin",
7
  "model.embed_tokens.weight": "pytorch_model-00001-of-00002.bin",
8
  "model.layers.0.input_layernorm.weight": "pytorch_model-00001-of-00002.bin",
9
  "model.layers.0.mlp.down_proj.weight": "pytorch_model-00001-of-00002.bin",
 
257
  "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00001-of-00002.bin",
258
  "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00001-of-00002.bin",
259
  "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
260
+ "model.norm.weight": "pytorch_model-00002-of-00002.bin"
 
261
  }
262
  }
special_tokens_map.json CHANGED
@@ -14,7 +14,7 @@
14
  "single_word": false
15
  },
16
  "pad_token": {
17
- "content": "[PAD]",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
 
14
  "single_word": false
15
  },
16
  "pad_token": {
17
+ "content": "<|eot_id|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8485b8900f2ef3a891091dbe5ef0edc56f6922c0a23c7dfbb4baf64854f7de4
3
- size 17210102
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4e61cf9383a37398dfd3322799286181b91dedc6a4947393255f5efca1ea6da
3
+ size 17210018
tokenizer_config.json CHANGED
@@ -2047,14 +2047,6 @@
2047
  "rstrip": false,
2048
  "single_word": false,
2049
  "special": true
2050
- },
2051
- "128256": {
2052
- "content": "[PAD]",
2053
- "lstrip": false,
2054
- "normalized": false,
2055
- "rstrip": false,
2056
- "single_word": false,
2057
- "special": true
2058
  }
2059
  },
2060
  "bos_token": "<|begin_of_text|>",
@@ -2062,12 +2054,17 @@
2062
  "clean_up_tokenization_spaces": true,
2063
  "eos_token": "<|eot_id|>",
2064
  "extra_special_tokens": {},
 
2065
  "model_input_names": [
2066
  "input_ids",
2067
  "attention_mask"
2068
  ],
2069
  "model_max_length": 131072,
2070
- "pad_token": "[PAD]",
2071
- "padding_side": "right",
2072
- "tokenizer_class": "PreTrainedTokenizerFast"
 
 
 
 
2073
  }
 
2047
  "rstrip": false,
2048
  "single_word": false,
2049
  "special": true
 
 
 
 
 
 
 
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
 
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
2056
  "extra_special_tokens": {},
2057
+ "max_length": 8000,
2058
  "model_input_names": [
2059
  "input_ids",
2060
  "attention_mask"
2061
  ],
2062
  "model_max_length": 131072,
2063
+ "pad_token": "<|eot_id|>",
2064
+ "padding": true,
2065
+ "stride": 0,
2066
+ "tokenizer_class": "PreTrainedTokenizerFast",
2067
+ "truncation": true,
2068
+ "truncation_side": "left",
2069
+ "truncation_strategy": "longest_first"
2070
  }