Llama Batch Inference of Llama-2-70B-Chat-GPTQ

#50
by Ivy111 - opened

I want to do llama batch inference so that it can run in multiple parallel environments. Can I implement it using Llama-2-70B-Chat-GPTQ?

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment