Llama Batch Inference of Llama-2-70B-Chat-GPTQ

#50
by Ivy111 - opened

I want to do llama batch inference so that it can run in multiple parallel environments. Can I implement it using Llama-2-70B-Chat-GPTQ?

Sign up or log in to comment