OpenThaiGPT 1.0.0 7B beta GPTQ 4 bit
OpenThaiGPT 1.0.0 7B beta GPTQ 4 bit
ทดสอบ การใช้งาน Model แบบ GPTQ แบบ 4 bit
GPTQ-for-LLaMA
4 bits quantization of LLaMA using GPTQ
GPTQ is SOTA one-shot weight quantization method
This code is based on GPTQ
There is a pytorch branch that allows you to use groupsize
and act-order
together.
Original Model ( FP16 )
https://huggingface.co/Adun/openthaigpt-1.0.0-beta-7b-ckpt-hf
GPTQ 4 bit Model
https://huggingface.co/Adun/openthaigpt-1.0.0-7b-chat-beta-gptq-4bit
Demo Code
Text generation web UI
A Gradio web UI for Large Language Models.
Setting Parameters
Test on
Hardware
- CPU : Intel i9-7900 3.3GHz
- GPU : NVIDIA RTX-2060 6GB
Software OS : Ubuntu 20.04.6 LTS
Youtube Demo
ChatBot
Q&A
Run Demo code
Reference
GPTQ-for-LLaMa https://github.com/amphancm/GPTQ-for-LLaMa
OpenThaiGPT https://openthaigpt.aieat.or.th/
ความคิดเห็น
แสดงความคิดเห็น