Comments (3)
Since AQLM employs vector quantization, each codes parameter encodes 8 fp16
weights of the uncompressed model for the 1x16 setup. On top of that, codebooks and scales also contribute a nontrivial number of parameters.
In the end. There really are ~6.5b parameters in ISTA-DASLab/Mixtral-8x7B-Instruct-v0_1-AQLM-2Bit-1x16-hf
: some of them are int16
codes encoding larger vectors, some of them serve more complicated roles.
To learn more about the representation, please refer to either the original AQLM paper or this nice blogpost explanation.
from aqlm.
Mixtral-8x7B-Instruct-v0_1 has approximately 46 billion parameters, obviously the number of parameters being output here is incorrect
from aqlm.
Since AQLM employs vector quantization, each codes parameter encodes 8
fp16
weights of the uncompressed model for the 1x16 setup. On top of that, codebooks and scales also contribute a nontrivial number of parameters. In the end. There really are ~6.5b parameters inISTA-DASLab/Mixtral-8x7B-Instruct-v0_1-AQLM-2Bit-1x16-hf
: some of them areint16
codes encoding larger vectors, some of them serve more complicated roles. To learn more about the representation, please refer to either the original AQLM paper or this nice blogpost explanation.
from aqlm.
Related Issues (20)
- Issues while attempting LLaMA-3 Quantization HOT 22
- Request: Phi-3-mini-128k-instruct Support HOT 2
- Query on Evaluation Support for C4 Validation HOT 5
- KV Cache Quantization HOT 4
- Minor race condition in CPU 2x8 inference code HOT 3
- Finetuning ISTA-DASLab/Mistral-7B-Instruct-v0.2-AQLM-2Bit-2x8: RuntimeError: CUDA error: invalid argument HOT 2
- Actual bitrate of models on github? HOT 5
- Request for the Llama-2-13B with AQLM (2x8 scheme) HOT 3
- How to run perplexity eval on HF hub models? HOT 3
- when load Llama, AutoConfig will occur error. HOT 2
- Request for Nvidia's RAG Implementation of Llama-3-70B "ChatQA 1.5" HOT 8
- Can you please share the *end-to-end* quantization script+config (including data used) for each model you've already quantized? (particularly llama-3 and miqu - i.e. 70B models) HOT 5
- FV tuning based on GPTQ HOT 6
- aqlm/inference_kernels/cuda_kernel.py HOT 2
- NaNs in sequence classifier output HOT 1
- Using pv-tuning on other quantization methods HOT 1
- How to import and use it in my existent code that loads LLMs? HOT 1
- [Feature Request] Gemma2 support & models HOT 1
- Quantization on multi-node GPUs HOT 2
- Performance issues with ~2bit quantization HOT 6
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from aqlm.