Comments (4)
Hi, thanks for reaching out. At the first glance, the model just use the LlamaForCausalLM
class, so the current codebase should be able to support it. Have you tried applying it to the model and see if it works? Thanks!
from llm-awq.
Actually, people tried applying the codebase to another salesforce model here #32 , which turns out to work.
from llm-awq.
XGen is a LLaMa model and is already supported. Posting an easy-to-use script here for you to get started. Note that this assumes you have:
- built AWQ
pip install tiktoken
- the following script in the llm-awq directory
On a 3090, I am getting 76 tokens/s with this model.
hfuser="Salesforce"
model_name="xgen-7b-8k-inst"
group_size=128
repo_path="$hfuser/$model_name"
model_path="/workspace/llm-awq/$model_name"
search_result_path="/workspace/llm-awq/$model_name/$model_name-awq-search.pt"
quantized_model_path="/workspace/llm-awq/$model_name/$model_name-w4-g$group_size.pt"
git clone https://huggingface.co/$repo_path
python3 -m awq.entry --model_path $model_path \
--w_bit 4 --q_group_size $group_size \
--run_awq --dump_awq $search_result_path
python3 -m awq.entry --model_path $model_path \
--w_bit 4 --q_group_size $group_size \
--load_awq $search_result_path \
--q_backend real \
--dump_quant $quantized_model_path
cd tinychat
python3 demo.py --model_type llama \
--model_path $model_path \
--q_group_size $group_size \
--load_quant $quantized_model_path \
--precision W4A16
from llm-awq.
Great, thanks so much.
from llm-awq.
Related Issues (20)
- No module named 'awq_inference_engine' HOT 2
- No such file or directory: "VILA1.5-13b-AWQ/llm/model-00001-of-00006.safetensors" HOT 8
- tinychat.serve.model_worker_new.py AWQ model in training mode
- how to support to custom module like mla in deep-seek-v2
- openAI-compatible tinychat API?
- AWQ kernel Issue
- Can you provide examples code to run inference on video QA? HOT 2
- AWQ and VILA dependency compatible issue HOT 2
- google.protobuf.message.DecodeError: Error parsing message HOT 1
- Is this a bug for the quantization phase? HOT 1
- Rocm support request
- Invalid Characters
- Memory increases significantly during inference
- Invalid Compute Capability when building Docker pytorch:23.12 HOT 1
- Request for Semi-Structured Sparse Matrix Support in AWQ Kernel
- Illegal memory access for LLama-3-70B
- 显卡要求
- How to load and infer the VILA-1.5-40B-AWQ model on multiple GPUs? I currently have 4 A30✖️24GB GPUs and a cuda out of memory error occurs.
- Add support for GPUs with compute capability lower than 8.0 for awq/kernels installation
- Plans for running model on other devices?
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from llm-awq.