#!/bin/bash python -m llava.eval.model_vqa_loader \ --model-path /mnt/petrelfs/zhuchenglin/LLaVA/checkpoints/llava-v1.5-13b-pretrain \ --model-base lmsys/vicuna-13b-v1.5 \ --question-file ./playground/data/eval/vizwiz/llava_test.jsonl \ --image-folder ./playground/data/eval/vizwiz/test \ --answers-file ./playground/data/eval/vizwiz/answers/llava-v1.5-13b.jsonl \ --temperature 0 \ --conv-mode vicuna_v1 python scripts/convert_vizwiz_for_submission.py \ --annotation-file ./playground/data/eval/vizwiz/llava_test.jsonl \ --result-file ./playground/data/eval/vizwiz/answers/llava-v1.5-13b.jsonl \ --result-upload-file ./playground/data/eval/vizwiz/answers_upload/llava-v1.5-13b.json