original model [weblab-10b-instruction-sft](https://huggingface.co/matsuo-lab/weblab-10b-instruction-sft) This is 4bit GPTQ Version. The size is smaller and the execution speed is faster, but the inference performance may be a little worse. Benchmark results are in progress. I will upload it at a later date.