Comments (4)
I get this, weight is fake int4, in calculation, actually is int16
from llm-awq.
![屏幕截图 2024-03-19 112200](https://private-user-images.githubusercontent.com/32669962/313904056-694c1d56-9d45-47e1-8f71-51ecdc544047.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTM0Njg2MDgsIm5iZiI6MTcxMzQ2ODMwOCwicGF0aCI6Ii8zMjY2OTk2Mi8zMTM5MDQwNTYtNjk0YzFkNTYtOWQ0NS00N2UxLThmNzEtNTFlY2RjNTQ0MDQ3LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDA0MTglMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwNDE4VDE5MjUwOFomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTU2ZDIyMjY1NTk0Mzc3Y2Q4NWNiM2Q5ODlkMzNiNGE5YTk1M2I2OTYzMWFhMmRlNDNhMzcwNjY0ZDRhZmQ3NTgmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.jOn-_nDHquiLTqx1HovUvfB61AeUcBmzDMFyvMe_lLQ)
from llm-awq.
below is my script to do quant
python -m awq.entry --model_path $MODEL
--w_bit 4 --q_group_size 128
--run_awq --dump_awq awq/llava_w4/llava-v1.6-vicuna-7b-w4-g128.pt
python -m awq.entry --model_path $MODEL
--w_bit 4 --q_group_size 128
--load_awq awq/llava_w4/llava-v1.6-vicuna-7b-w4-g128.pt
--q_backend real --dump_quant awq/llava_w4/llava-v1.6-vicuna-7b-w4-g128-awq.pt
from llm-awq.
I get this, weight is fake int4, in calculation, actually is int16
If it's convenient for you, could you explain it?
from llm-awq.
Related Issues (20)
- `RuntimeError: probability tensor contains either `inf`, `nan` or element < 0` when running LLaVA demo HOT 2
- Llava weight
- awq_inference_engine has no attribute 'gemm_forward_cuda_new' HOT 4
- reproduce Llama2 7b failure : RuntimeError: The expanded size of the tensor (4608) must match the existing size (4096) at non-singleton dimension 3. Target sizes: [65, 32, 512, 4608]. Tensor sizes: [65, 1, 512, 4096] HOT 3
- RuntimeError: Unknown Layout in CUDA Kernel Execution
- Use awq to quantize Deepseek-coder-33B-instruct model
- run_awq.<locals>.Catcher.forward() error
- KeyError: 'llava_llama' HOT 1
- Error while generating real quantized weights for VILA
- Possible Bug in "_search_module_scale" Function
- AWQ for non-transformer layers?
- Out of memory in Jetson Orin NX 8GB
- Inquiry about Minimum GPU Requirements HOT 1
- when q-group-size = -1,the code will not run
- Weight Packing Format
- illegal memory access when input tokens < 8
- Grok-1 AWQ
- can awq support 3-bit,2-bit, 8-bit quantization? HOT 1
- awq_inference_engine is missing from source, so quantizing custom models fails HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from llm-awq.