File size: 443 Bytes
5fa1a76 |
1 2 3 4 5 6 7 8 9 10 11 12 13 |
With the [Trainer] class, you can enable JIT mode for CPU inference by setting the --jit_mode_eval flag: python run_qa.py \ --model_name_or_path csarron/bert-base-uncased-squad-v1 \ --dataset_name squad \ --do_eval \ --max_seq_length 384 \ --doc_stride 128 \ --output_dir /tmp/ \ --no_cuda \ --jit_mode_eval For PyTorch >= 1.14.0, JIT-mode could benefit any model for prediction and evaluation since the dict input is supported in jit.trace. |