Results 311 to 320 of about 1,866,673 (365)
Some of the next articles are maybe not open access.
[Proceedings] 1992 IEEE International Symposium on Circuits and Systems, 2003
A technique of vector quantization based on neural networks called neural vector quantization is investigated. A neural vector quantizer has been developed for data compression and is capable of faster parallel quantization than conventional vector quantizers. The architecture, dynamics, and training strategies are presented.
J. Hanson, Zhongde Wang
openaire +2 more sources
A technique of vector quantization based on neural networks called neural vector quantization is investigated. A neural vector quantizer has been developed for data compression and is capable of faster parallel quantization than conventional vector quantizers. The architecture, dynamics, and training strategies are presented.
J. Hanson, Zhongde Wang
openaire +2 more sources
Extension of two-stage vector quantization-lattice vector quantization
IEEE Transactions on Communications, 1997This paper is the extension of two-stage vector quantization-(spherical) lattice vector quantization (VQ-(S)LVQ) recently introduced by Pan and Fischer (see IEEE Trans. Inform. Theory, vol.41, p.155, 1995). First, according to high resolution quantization theory, generalized vector quantization-lattice vector quantization (G-VQ-LVQ) is formulated in ...
openaire +2 more sources
QuIP#: Even Better LLM Quantization with Hadamard Incoherence and Lattice Codebooks
International Conference on Machine LearningPost-training quantization (PTQ) reduces the memory footprint of LLMs by quantizing their weights to low-precision. In this work, we introduce QuIP#, a weight-only PTQ method that achieves state-of-the-art results in extreme compression regimes ($\le$ 4 ...
Albert Tseng+4 more
semanticscholar +1 more source
2006
Publisher Summary A quantization strategy that works with sequences or blocks of output provides some improvement in performance over scalar quantization. This chapter provides an overview of vector quantization. The vector of source outputs forms the input to the vector quantizer. At both the encoder and decoder of the vector quantizer, one has a set
openaire +3 more sources
Publisher Summary A quantization strategy that works with sequences or blocks of output provides some improvement in performance over scalar quantization. This chapter provides an overview of vector quantization. The vector of source outputs forms the input to the vector quantizer. At both the encoder and decoder of the vector quantizer, one has a set
openaire +3 more sources
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
Neural Information Processing SystemsLLMs are seeing growing use for applications which require large context windows, and with these large context windows KV cache activations surface as the dominant contributor to memory consumption during inference.
Coleman Hooper+6 more
semanticscholar +1 more source
Vector quantization of neural networks
IEEE Transactions on Neural Networks, 1998The problem of vector quantizing the parameters of a neural network is addressed, followed by a discussion of different algorithms applicable for quantizer design. Optimal, as well as several suboptimal quantization schemes are described. Simulations involving nonlinear prediction of speech signals are presented to compare the performance of different ...
W.C. Chu, N.K. Bose
openaire +3 more sources
Unrestricted multistage vector quantizers
IEEE Transactions on Information Theory, 1992A low-complexity, high-performance vector quantization technique for medium-to-high-bit-rate encoding based on multistage (residual) VQ is introduced. The design, a variable-rate version of multistage VQ, is the vector extension of the piecewise uniform scalar quantizers of P.F. Swaszek and J.B.
openaire +4 more sources
Trellis coded vector quantization
IEEE Transactions on Communications, 1992A vector generalization of trellis coded quantization (TCQ), called trellis coded vector quantization (TCVQ), and experimental results showing its performance for an i.i.d. Gaussian source are presented. For a given rate, TCVQ yields a lower distortion that TCQ at the cost of an increase in implementation complexity. In addition, TCVQ allows fractional
H.S. Wang, Nader Moayeri
openaire +2 more sources
2008 7th World Congress on Intelligent Control and Automation, 2008
This paper proposes a new spectrum vector quantization algorithm (SVQ). SVQ conducts vector quantization in spectrum space. It is characterized by some novels. The first is the informed initialization of prototypes, which is achieved by a modified support vector clustering procedure. The second is the SVD-based spectrum analysis. This technique employs
Ping Ling+4 more
openaire +2 more sources
This paper proposes a new spectrum vector quantization algorithm (SVQ). SVQ conducts vector quantization in spectrum space. It is characterized by some novels. The first is the informed initialization of prototypes, which is achieved by a modified support vector clustering procedure. The second is the SVD-based spectrum analysis. This technique employs
Ping Ling+4 more
openaire +2 more sources
Constrained Vector Quantization
1992There is no better way to quantize a single vector than to use VQ with a codebook that is optimal for the probability distribution describing the random vector. However, direct use of VQ suffers from a serious complexity barrier that greatly limits its practical use as a complete and self-contained coding ...
Robert M. Gray, Allen Gersho
openaire +2 more sources