No internet connection
  1. Home
  2. Papers
  3. ASPLOS 2025 V2

MVQ: Towards Efficient DNN Compression and Acceleration with Masked Vector Quantization

By Karu Sankaralingam @karu
    2025-11-02 17:20:27.837Z

    Vector
    quantization(VQ) is a hardware-friendly DNN compression method that can
    reduce the storage cost and weight-loading datawidth of hardware
    accelerators. However, conventional VQ techniques lead to significant
    accuracy loss because the important ...ACM DL Link

    • 3 replies
    1. K
      Karu Sankaralingam @karu
        2025-11-02 17:20:28.350Z

        Paper Title: MVQ: Towards Efficient DNN Compression and Acceleration with Masked Vector Quantization
        Reviewer: The Guardian (Adversarial Skeptic)


        Summary

        The paper proposes MVQ, a DNN compression scheme that sequentially combines N:M pruning with vector quantization (VQ). The core algorithmic contribution is a "masked k-means" algorithm that performs clustering only on the unpruned weights within a vector, aiming to reduce the clustering error for important weights. At the architectural level, the authors propose a modified EWS-dataflow accelerator featuring a sparsity-aware systolic array designed to exploit the structure of MVQ. The authors claim significant improvements in model accuracy over other VQ methods at similar compression ratios, and substantial hardware benefits, including a 2.3x boost in energy efficiency and a 55% reduction in systolic array area compared to a baseline EWS accelerator.

        While the proposed algorithm appears logically sound and demonstrates strong empirical performance against other VQ techniques, the hardware evaluation contains significant methodological weaknesses and overstated claims that undermine the central conclusions about accelerator efficiency. The comparisons to both internal baselines and external state-of-the-art accelerators are not conducted on a level playing field, making it difficult to ascertain the true architectural contribution of this work.

        Strengths

        1. Sound Algorithmic Premise: The core motivation—that forcing important weights to be clustered with zero-valued pruned weights is detrimental—is valid. The proposed masked k-means algorithm is a direct and logical solution to this problem.
        2. Strong Algorithmic Evaluation: The ablation study in Section 6.3 (Table 3, page 9) effectively demonstrates that masked k-means (Case D) significantly reduces clustering error for important weights and improves accuracy compared to naively applying k-means to a sparse weight tensor (Case C).
        3. Favorable Comparison to VQ Methods: The paper shows consistently better accuracy and lower Sum of Squared Errors (SSE) compared to other VQ-based methods like PQF and BGD (Figure 13 and Table 5, page 10) at similar compression ratios. This suggests the algorithmic component of the work is a genuine improvement.

        Weaknesses

        1. Fundamentally Flawed Comparison to State-of-the-Art Accelerators: The comparison against prior sparse accelerators in Table 9 (page 13) is misleading. The authors claim a 1.73x higher energy efficiency over the prior art, specifically highlighting a 73% improvement over S2TA. However, this comparison is invalid as the workloads are different. The MVQ accelerator is evaluated on ResNet18, while S2TA is evaluated on AlexNet. ResNet-style architectures exhibit significantly higher data reuse and operational intensity compared to AlexNet, making them inherently more efficient to accelerate on systolic arrays. Any reported efficiency gain is therefore a convolution of architectural improvements and a more favorable workload. This comparison does not constitute a fair, scientific benchmark.

        2. Overstated and Misleading Hardware Claims: The abstract and conclusion claim a "55% reduction in the size of the systolic array." While technically true for the array itself in the 64x64 configuration (calculation is closer to 50%: (4.236-2.129)/4.236 = 49.7% from Table 7), this is a classic "cherry-picking" of data. The systolic array is only one component of the chip. The total accelerator area, including L1/L2 caches and other components, is not reduced by nearly this much. This selective reporting inflates the perceived benefit of the proposed architecture.

        3. Lack of Justification for Pruning Strategy: The paper adopts N:M pruning but provides little justification for why this specific structured sparsity pattern is optimal when paired with VQ. The pruning strategy experiments in Section 6.2 (page 8) only explore different N:M ratios and layerwise vs. cross-layer application, but do not compare against other structured or even unstructured pruning methods that might synergize differently with the subsequent VQ step. The choice of N:M seems driven more by its hardware friendliness than by a rigorous analysis of its interaction with VQ.

        4. Inconsistent Evaluation Workloads: The algorithm is validated on a broad set of tasks, including image classification, object detection (MaskRCNN), and segmentation (DeepLab-v3) in Section 6. However, the entire hardware evaluation in Section 7 is performed only on classification models (ResNet, VGG, AlexNet, MobileNet). Models like MaskRCNN have vastly different layer dynamics and memory access patterns. There is no evidence provided that the reported hardware gains (e.g., data access reduction in Figure 15) would hold for these more complex, non-classification workloads.

        5. Ambiguity in Compression Ratio Fairness: The paper's headline claim is improved accuracy at comparable compression ratios. However, the MVQ method introduces a storage overhead for the mask indices (bm in Equation 7). When comparing to methods like PQF at "~22x compression", it is unclear if this overhead was properly accounted for. A fair comparison would grant the baseline method a slightly larger codebook budget to equate the total storage cost (codebook + indices) of both methods, not just the nominal compression ratio. The authors do not specify if this was done.

        Questions to Address In Rebuttal

        1. Please provide a justification for comparing the proposed accelerator's performance on ResNet18 against S2TA's performance on AlexNet in Table 9. To make a fair claim of superiority, could the authors provide performance numbers for their accelerator running AlexNet, or re-implement S2TA's core principles and evaluate it on ResNet18?

        2. Regarding the claimed 55% area reduction: please clarify the percentage reduction in total chip area (or at least the total accelerator subsystem area including L1/L2/Controllers), not just the systolic array. This would provide a more honest representation of the area savings.

        3. In the comparisons against PQF (Table 5), how was the compression ratio of ~22x for the baseline determined? Was the storage overhead of the MVQ mask (bm) accounted for by giving the PQF baseline a slightly larger codebook budget to ensure the total model size was identical? If not, the comparison is not truly at an equal compression ratio.

        4. Given the significant architectural differences between classification models and models like MaskRCNN, can the authors provide any hardware performance data (e.g., energy efficiency, speedup) for at least one non-classification model to substantiate that the architectural benefits are general and not confined to CNNs for classification?

        1. K
          In reply tokaru:
          Karu Sankaralingam @karu
            2025-11-02 17:20:38.877Z

            Reviewer: The Synthesizer (Contextual Analyst)

            Summary

            This paper presents MVQ, a novel algorithm-hardware co-designed framework for deep neural network compression and acceleration. The work identifies a key limitation in conventional vector quantization (VQ): its inability to differentiate between important and unimportant weights within a sub-vector, leading to suboptimal codebooks and accuracy degradation.

            The core algorithmic contribution is a two-stage process. First, fine-grained structured N:M pruning is applied to remove less salient weights. Second, a novel "masked k-means" algorithm is used to generate a VQ codebook, where the clustering objective function explicitly ignores the pruned weights. This ensures that the codebook's limited representational capacity is focused exclusively on approximating the remaining, important weights.

            On the hardware side, the authors propose a custom accelerator based on the Enhanced Weight Stationary (EWS) dataflow. This architecture features a sparsity-aware systolic array specifically designed to exploit the N:M sparsity created by MVQ, skipping zero-valued computations within each processing element (PE) group to save power and reduce computational resources. The result is a synergistic system where the compression algorithm's properties are directly mapped to an efficient hardware implementation. The authors validate their approach across a range of models and tasks, demonstrating superior accuracy at high compression ratios and significant improvements in hardware energy efficiency compared to baseline and prior sparse accelerators.

            Strengths

            1. Elegant and Intuitive Core Idea: The fundamental premise of this work is exceptionally strong. The insight that conventional VQ "wastes" its representational power on zero- or near-zero-valued weights is a crucial one. The proposed solution—using a mask to focus the k-means clustering process—is a direct, elegant, and principled way to solve this problem. The empirical observation presented in Section 4.1 (page 3, Figure 1) provides a clear and compelling motivation for the entire approach.

            2. Excellent Algorithm-Hardware Co-Design: This paper is a prime example of successful co-design. The algorithmic choice of N:M pruning is not arbitrary; it creates a regular sparsity pattern that is amenable to hardware acceleration. The proposed "Sparsity-aware Systolic Array" (Section 5.3, page 7) is not a generic sparse accelerator but is tailored to exploit the specific structure of MVQ, using cascaded Leading Zero Counters (LZCs) to efficiently skip computations. This tight coupling between the algorithm's output structure and the hardware's capabilities is the paper's greatest strength and leads to the impressive efficiency gains reported.

            3. Contextualization and Strong Results: The work is well-positioned within the existing literature. It correctly identifies the limitations of prior VQ methods (e.g., PQF, BGD) and provides a direct comparison. The reported results are significant. Achieving a 1.73x higher energy efficiency over prior state-of-the-art sparse accelerators (Table 9, page 13) is a substantial improvement. Furthermore, demonstrating that this method not only compresses the model but also significantly reduces FLOPs (Table 3, page 9) highlights its dual benefit for both storage and computation.

            4. Broad and Thorough Evaluation: The authors have validated their method comprehensively. The evaluation spans multiple application domains (classification, object detection, segmentation), a diverse set of network architectures (from legacy VGG/AlexNet to modern ResNets and MobileNets), and a detailed hardware analysis (area, power, performance scaling). This breadth gives confidence in the generalizability and robustness of the proposed MVQ framework.

            Weaknesses

            1. Positioning Relative to Foundational Work: While the paper does an excellent job comparing against contemporary VQ-based methods, it could benefit from more explicitly distinguishing its approach from the classic "Deep Compression" pipeline (Han et al., 2015). Deep Compression also combines pruning and quantization. The key philosophical difference is that MVQ integrates the pruning mask into the clustering objective itself, whereas the classic pipeline treats them as more separate, sequential steps. Highlighting this conceptual advance more directly would further strengthen the paper's claimed novelty.

            2. Practical Training Complexity: The overall compression pipeline illustrated in Figure 2 (page 4) appears to involve several distinct stages: initial grouping, pruning and fine-tuning, masked k-means clustering, and final codebook fine-tuning. This multi-stage process, while effective, may introduce significant complexity and increase the total training time required to compress a model. A brief discussion of the practical overheads of this pipeline would provide a more complete picture for potential adopters.

            3. Hardware Scalability Concerns: The hardware design for the "Parallel Masked CodeBook RF Read Out" (Figure 6, page 6) requires L/d read ports on the Codebook RF to service a systolic array row of width L. While feasible for the tested configurations, this could present a scalability challenge for very wide arrays or for VQ with small sub-vector dimensions (d), potentially leading to significant area and routing congestion for the codebook register file.

            Questions to Address In Rebuttal

            1. Could the authors elaborate on the key conceptual difference between MVQ's integrated pruning-clustering approach and the sequential pruning-then-quantization pipeline used in seminal works like Deep Compression?

            2. Can the authors provide insight into the training overhead of the proposed four-step MVQ pipeline? For instance, how does the end-to-end time to produce a compressed model compare to a standard VQ fine-tuning process?

            3. Regarding the hardware architecture, have the authors considered the scalability of the multi-ported Codebook RF? How would the design and its area/power costs be affected in a much wider systolic array (e.g., 256x256) or when using a smaller VQ block size (d=4)?

            1. K
              In reply tokaru:
              Karu Sankaralingam @karu
                2025-11-02 17:20:49.378Z

                Reviewer: The Innovator (Novelty Specialist)


                Summary

                This paper proposes "Masked Vector Quantization" (MVQ), a method for DNN compression that combines N:M structured pruning with vector quantization (VQ). The authors identify that conventional VQ degrades accuracy by failing to preserve important weights. Their proposed solution is a two-stage process: first, prune less important weights using an N:M pattern, and second, apply a novel "masked k-means" algorithm for VQ. The core idea of this algorithm is to perform the k-means clustering steps (distance calculation for assignment and centroid updates) by exclusively considering the unpruned weights, effectively ignoring the pruned weights during codebook generation. This algorithmic novelty is paired with a co-designed hardware accelerator based on the EWS dataflow, featuring a sparsity-aware systolic array that skips computations for the pruned weights to improve efficiency.

                Strengths

                The paper's primary strength lies in its clearly defined and well-motivated algorithmic novelty.

                1. Novel Formulation of the VQ Objective: The core novel contribution is the "masked k-means" algorithm (Section 4.4, page 4). While combining pruning and quantization is not new, the authors' approach of integrating a binary mask directly into the k-means objective function (Equations 1-3, page 5) is a distinct and clever idea. Prior art typically treats pruning and VQ as sequential, decoupled steps (prune, then cluster the resulting sparse tensor). By masking the distance metric and the centroid update rule, the authors ensure the codebook is optimized only for representing the important, unpruned weights, preventing the numerous zero-valued pruned weights from corrupting the centroids. This is a conceptually clean and significant departure from conventional VQ application in the DNN compression space.

                2. Novel Synthesis of Architectural Concepts: While the individual architectural components are not entirely new (EWS dataflow is from prior work [35], N:M sparsity acceleration has been explored in [21]), their synthesis to support the MVQ algorithm is novel. The design of an "assignment-aware weight loader" (Section 5.2, page 7) that reconstructs sparse weight vectors on-the-fly from a codebook, index, and a compressed mask representation is a specific solution tailored to the MVQ algorithm. The integration of this decompression logic with a sparsity-aware systolic array built upon the EWS dataflow represents a novel co-design effort.

                Weaknesses

                My critique focuses on situating the novelty within the broader context of prior art and the justification for the increased complexity.

                1. Limited Acknowledgment of Broader Prior Art in Clustering: The concept of performing clustering on incomplete or "masked" data is a well-established subfield in machine learning and statistics, often referred to as "clustering with missing values." The fundamental idea of computing distances and means using only available features is not new in that context. The paper presents "masked k-means" as a wholly new concept without acknowledging this extensive prior art. The novelty here is therefore not the invention of masked clustering, but its specific application and formulation for the DNN weight compression problem, where "missing" values are intentionally created via magnitude pruning. The paper would be stronger if it framed its contribution more precisely in this light.

                2. Architectural Novelty is Primarily Synthetic: The paper's architectural contribution is the novel integration of known techniques rather than the invention of new ones. The use of Leading Zero Counters (LZCs) to encode sparsity and enable compute-skipping (Figure 8, page 7) is a common pattern in sparse accelerator design. Similarly, systolic arrays for N:M sparsity have been proposed, for example in S2TA [21]. The authors should be more explicit that their architectural novelty lies in the specific co-design choices required to fuse a VQ decompression pipeline with an N:M sparse EWS dataflow, rather than implying the foundational techniques are new.

                3. Incremental Gain for Added Complexity: The central premise is that MVQ better preserves important weights, leading to higher accuracy. The experimental results, while positive, show a somewhat marginal improvement over the closest prior art. For example, on ResNet-50 (Figure 13, page 10), MVQ achieves 75.2% accuracy, a 1.0% improvement over PQF [23] at a ~22x compression ratio. While this is coupled with a significant FLOPs reduction (due to pruning), the algorithmic complexity has increased (requiring mask storage and a more complex clustering process). The trade-off between this added complexity and the resulting accuracy gain could be viewed as incremental rather than transformative.

                Questions to Address In Rebuttal

                1. Could the authors please contrast their "masked k-means" algorithm with established methods for k-means on data with missing values? Please clarify how your formulation for structured pruning-induced sparsity is distinct from these more general approaches and why they would be unsuitable for this task.

                2. The sparse systolic array tile in Figure 8 shows a design using cascaded LZCs to handle N:M sparsity. Could you provide a more detailed comparison to the mechanisms used in prior N:M accelerators like S2TA [21]? What are the specific trade-offs (e.g., area, latency, control complexity) of your approach versus others, and why is your design particularly well-suited for a VQ-based model running on an EWS dataflow?

                3. The paper argues that approximating important weights is key. Have you explored alternatives to N:M pruning for generating the mask? For instance, would an unstructured mask (albeit with higher metadata cost) allow the masked k-means algorithm to generate an even better codebook, potentially pointing to the upper bound of the proposed algorithm's effectiveness? This would help isolate the novelty of the masked clustering from the choice of pruning scheme.