-
Notifications
You must be signed in to change notification settings - Fork 571
Open
Description
I recently came across your CCQ quantization method as described in https://arxiv.org/pdf/2507.07145, and I'm have some questions about the coding setup. My understanding is that the method does the following
- First encode vectors of dimension 3 with a (L=2, N=3, S=2) and vectors of dimension 4 with a (3, 4, 2) setup to get a compressed bitstream and set of scales
- Further compress the bitstreams by doing vector quantization to a "meta codebook"
- Compress the scales
My questions are:
- The baseline encoding scheme appears to just be QTIP where L = L, N = the vector dimension, V = 1, and S = K = the bitrate. The codebook is a direct lookup (bit to integer) codebook with a scale. The bitshift decoding section in your paper corresponds to the bitshift trellis in QTIP. Can you clarify how the base encoding scheme is fundamentally different from QTIP and not just a variant of QTIP?
- If my understanding is correct, does step 2 cluster the 3D vectors together and the 4D vectors together, or the bitstream together as a 7D vector?
- The end result of this seems to be a 4D + 3D vector quantizer or 7D VQ. If its 4D and 3D (two separate vector quantizers), why not just use unstructured 3D and 4D VQ? The codebook should be small enough to fit in cache in both cases.
- Why not use entropy coding since in step 2 instead of higher order vector quantization?
- Is the rate distortion tradeoff of doing step 2 better than just doing TCQ at the target bitrate?
- What is the "base" rounding algorithm? The paper states that within each vector in step 1, you find the minimize MSE representation, which corresponds to rounding to the nearest. However, RTN is known to do very poorly regardless of how good your quantizer is. Do you use GPTQ, AWQ, or something else?
I would have sent an email to the authors except the paper doesn't list the author emails.
Thanks
Metadata
Metadata
Assignees
Labels
No labels