czczup commited on
Commit
a1968b7
·
verified ·
1 Parent(s): 2d0c52f

Delete vqa-ai2d-test.log

Browse files
Files changed (1) hide show
  1. vqa-ai2d-test.log +0 -183
vqa-ai2d-test.log DELETED
@@ -1,183 +0,0 @@
1
- + CHECKPOINT=work_dirs/InternVL2-2B
2
- + DATASET=vqa-ai2d-test
3
- ++ pwd
4
- + CHECKPOINT=/mnt/petrelfs/wangweiyun/workspace_zyc/VLM-Dev/work_dirs/InternVL2-2B
5
- ++ pwd
6
- + export PYTHONPATH=/mnt/petrelfs/wangweiyun/workspace_zyc/VLM-Dev:/mnt/petrelfs/wangweiyun/workspace_wwy/pkgs/petrel-oss-sdk-2.3.14:/mnt/petrelfs/share_data/wangweiyun/share_pkgs/petrel-oss-sdk-2.3.12:
7
- + PYTHONPATH=/mnt/petrelfs/wangweiyun/workspace_zyc/VLM-Dev:/mnt/petrelfs/wangweiyun/workspace_wwy/pkgs/petrel-oss-sdk-2.3.14:/mnt/petrelfs/share_data/wangweiyun/share_pkgs/petrel-oss-sdk-2.3.12:
8
- + echo 'CHECKPOINT: /mnt/petrelfs/wangweiyun/workspace_zyc/VLM-Dev/work_dirs/InternVL2-2B'
9
- CHECKPOINT: /mnt/petrelfs/wangweiyun/workspace_zyc/VLM-Dev/work_dirs/InternVL2-2B
10
- + MASTER_PORT=63669
11
- + PORT=63665
12
- + GPUS=8
13
- + GPUS_PER_NODE=8
14
- + NODES=1
15
- + export MASTER_PORT=63669
16
- + MASTER_PORT=63669
17
- + export PORT=63665
18
- + PORT=63665
19
- + ARGS=("$@")
20
- + [[ 5 -gt 0 ]]
21
- + case "$1" in
22
- + shift
23
- + [[ 4 -gt 0 ]]
24
- + case "$1" in
25
- + shift
26
- + [[ 3 -gt 0 ]]
27
- + case "$1" in
28
- + shift
29
- + [[ 2 -gt 0 ]]
30
- + case "$1" in
31
- + shift
32
- + [[ 1 -gt 0 ]]
33
- + case "$1" in
34
- + shift
35
- + [[ 0 -gt 0 ]]
36
- + echo 'GPUS: 8'
37
- GPUS: 8
38
- + [[ /mnt/petrelfs/wangweiyun/workspace_zyc/VLM-Dev/work_dirs/InternVL2-2B == */ ]]
39
- + '[' vqa-ai2d-test == mme ']'
40
- + '[' vqa-ai2d-test == caption ']'
41
- + '[' vqa-ai2d-test == caption-coco ']'
42
- + '[' vqa-ai2d-test == caption-flickr30k ']'
43
- + '[' vqa-ai2d-test == caption-nocaps ']'
44
- + '[' vqa-ai2d-test == vqa ']'
45
- + '[' vqa-ai2d-test == vqa-okvqa-val ']'
46
- + '[' vqa-ai2d-test == vqa-textvqa-val ']'
47
- + '[' vqa-ai2d-test == vqa-textvqa-val-ocr ']'
48
- + '[' vqa-ai2d-test == vqa-vizwiz-val ']'
49
- + '[' vqa-ai2d-test == vqa-vizwiz-test ']'
50
- + '[' vqa-ai2d-test == vqa-vqav2-testdev ']'
51
- + '[' vqa-ai2d-test == vqa-ai2d-test ']'
52
- + torchrun --nnodes=1 --node_rank=0 --master_addr=127.0.0.1 --nproc_per_node=8 --master_port=63669 eval/vqa/evaluate_vqa.py --checkpoint /mnt/petrelfs/wangweiyun/workspace_zyc/VLM-Dev/work_dirs/InternVL2-2B --datasets ai2diagram_test --dynamic --max-num 6
53
- [2024-08-07 21:55:03,764] torch.distributed.run: [WARNING]
54
- [2024-08-07 21:55:03,764] torch.distributed.run: [WARNING] *****************************************
55
- [2024-08-07 21:55:03,764] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
56
- [2024-08-07 21:55:03,764] torch.distributed.run: [WARNING] *****************************************
57
- datasets: ['ai2diagram_test']
58
- datasets: ['ai2diagram_test']
59
- datasets: ['ai2diagram_test']
60
- datasets: ['ai2diagram_test']
61
- datasets: ['ai2diagram_test']
62
- datasets: ['ai2diagram_test']
63
- datasets: ['ai2diagram_test']
64
- datasets: ['ai2diagram_test']
65
- Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
66
- Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
67
- Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
68
- Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
69
- Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
70
- Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
71
- Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
72
- Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
73
- [test] total_params: 2.205754368B, use num_beams: 1
74
- [test] image_size: 448
75
- [test] template: internlm2-chat
76
- [test] dynamic_image_size: True
77
- [test] use_thumbnail: True
78
- [test] total_params: 2.205754368B, use num_beams: 1
79
- [test] image_size: 448
80
- [test] template: internlm2-chat
81
- [test] dynamic_image_size: True
82
- [test] use_thumbnail: True
83
- [test] total_params: 2.205754368B, use num_beams: 1
84
- [test] image_size: 448
85
- [test] template: internlm2-chat
86
- [test] dynamic_image_size: True
87
- [test] use_thumbnail: True
88
- [test] total_params: 2.205754368B, use num_beams: 1
89
- [test] image_size: 448
90
- [test] total_params: 2.205754368B, use num_beams: 1[test] template: internlm2-chat
91
-
92
- [test] image_size: 448[test] total_params: 2.205754368B, use num_beams: 1
93
- [test] dynamic_image_size: True
94
- [test] use_thumbnail: True
95
-
96
- [test] image_size: 448
97
- [test] template: internlm2-chat
98
- [test] dynamic_image_size: True[test] template: internlm2-chat
99
-
100
- [test] use_thumbnail: True
101
- [test] dynamic_image_size: True
102
- [test] use_thumbnail: True
103
- [test] total_params: 2.205754368B, use num_beams: 1
104
- [test] image_size: 448
105
- [test] template: internlm2-chat
106
- [test] dynamic_image_size: True
107
- [test] use_thumbnail: True
108
- [test] total_params: 2.205754368B, use num_beams: 1
109
- [test] image_size: 448
110
- [test] template: internlm2-chat
111
- [test] dynamic_image_size: True
112
- [test] use_thumbnail: True
113
-
114
- warnings.warn(
115
- /mnt/petrelfs/wangweiyun/miniconda3/envs/internvl/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:392: UserWarning: `do_sample` is set to `False`. However, `temperature` is set to `0.0` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `temperature`.
116
- warnings.warn(
117
- /mnt/petrelfs/wangweiyun/miniconda3/envs/internvl/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:392: UserWarning: `do_sample` is set to `False`. However, `temperature` is set to `0.0` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `temperature`.
118
- warnings.warn(
119
- /mnt/petrelfs/wangweiyun/miniconda3/envs/internvl/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:392: UserWarning: `do_sample` is set to `False`. However, `temperature` is set to `0.0` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `temperature`.
120
- warnings.warn(
121
- /mnt/petrelfs/wangweiyun/miniconda3/envs/internvl/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:392: UserWarning: `do_sample` is set to `False`. However, `temperature` is set to `0.0` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `temperature`.
122
- warnings.warn(
123
- /mnt/petrelfs/wangweiyun/miniconda3/envs/internvl/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:392: UserWarning: `do_sample` is set to `False`. However, `temperature` is set to `0.0` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `temperature`.
124
- warnings.warn(
125
- /mnt/petrelfs/wangweiyun/miniconda3/envs/internvl/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:392: UserWarning: `do_sample` is set to `False`. However, `temperature` is set to `0.0` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `temperature`.
126
- warnings.warn(
127
- /mnt/petrelfs/wangweiyun/miniconda3/envs/internvl/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:392: UserWarning: `do_sample` is set to `False`. However, `temperature` is set to `0.0` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `temperature`.
128
- warnings.warn(
129
-
130
-
131
-
132
-
133
-
134
-
135
-
136
-
137
- Evaluating ai2diagram_test ...
138
- Results saved to results/ai2diagram_test_240807215603.json
139
- ai2diagram_test {'accuracy': 0.7412564766839378}
140
- write results to file results/work_dirs_InternVL2-2B.txt
141
- ['/mnt/petrelfs/wangweiyun/workspace_zyc/VLM-Dev/work_dirs/InternVL2-2B', 'ai2diagram_test', {'accuracy': 0.7412564766839378}]
142
- write results to file results/work_dirs_InternVL2-2B.txt
143
- write results to file results/work_dirs_InternVL2-2B.txt
144
- write results to file results/work_dirs_InternVL2-2B.txtwrite results to file results/work_dirs_InternVL2-2B.txt
145
-
146
- write results to file results/work_dirs_InternVL2-2B.txt
147
- write results to file results/work_dirs_InternVL2-2B.txt
148
- write results to file results/work_dirs_InternVL2-2B.txt
149
- + '[' vqa-ai2d-test == vqa-vqav2-val ']'
150
- + '[' vqa-ai2d-test == vqa-gqa-testdev ']'
151
- + '[' vqa-ai2d-test == vqa-docvqa-val ']'
152
- + '[' vqa-ai2d-test == vqa-docvqa-test ']'
153
- + '[' vqa-ai2d-test == vqa-chartqa-test ']'
154
- + '[' vqa-ai2d-test == vqa-infovqa-val ']'
155
- + '[' vqa-ai2d-test == vqa-infovqa-test ']'
156
- + '[' vqa-ai2d-test == vqa-chartqa-test-human ']'
157
- + '[' vqa-ai2d-test == vqa-chartqa-test-augmented ']'
158
- + '[' vqa-ai2d-test == vqa-ocrvqa-val ']'
159
- + '[' vqa-ai2d-test == vqa-ocrvqa-test ']'
160
- + '[' vqa-ai2d-test == refcoco ']'
161
- + '[' vqa-ai2d-test == refcoco-val ']'
162
- + '[' vqa-ai2d-test == llava-bench ']'
163
- + '[' vqa-ai2d-test == pope ']'
164
- + '[' vqa-ai2d-test == tiny_lvlm ']'
165
- + '[' vqa-ai2d-test == mmvet ']'
166
- + '[' vqa-ai2d-test == cmmmu ']'
167
- + '[' vqa-ai2d-test == mmbench-dev-en ']'
168
- + '[' vqa-ai2d-test == mmbench-dev-cn ']'
169
- + '[' vqa-ai2d-test == mmbench-test-en ']'
170
- + '[' vqa-ai2d-test == mmbench-test-cn ']'
171
- + '[' vqa-ai2d-test == ccbench-dev ']'
172
- + '[' vqa-ai2d-test == scienceqa ']'
173
- + '[' vqa-ai2d-test == mmmu-dev ']'
174
- + '[' vqa-ai2d-test == mmmu-val ']'
175
- + '[' vqa-ai2d-test == mmmu-test ']'
176
- + '[' vqa-ai2d-test == mmmu-dev-cot ']'
177
- + '[' vqa-ai2d-test == mmmu-val-cot ']'
178
- + '[' vqa-ai2d-test == mmmu-test-cot ']'
179
- + '[' vqa-ai2d-test == mmvp ']'
180
- + '[' vqa-ai2d-test == mathvista-testmini ']'
181
- + '[' vqa-ai2d-test == mathvista-test ']'
182
- + '[' vqa-ai2d-test == seed ']'
183
- + '[' vqa-ai2d-test == mvbench ']'