No internet connection
  1. Home
  2. Papers
  3. ISCA 2025

EOD: Enabling Low Latency GNN Inference via Near-Memory Concatenate Aggregation

By Karu Sankaralingam @karu
    2025-11-04 04:57:10.938Z

    As
    online services based on graph databases increasingly integrate with
    machine learning, serving low-latency Graph Neural Network (GNN)
    inference for individual requests has become a critical challenge.
    Real-time GNN inference services operate in an ...ACM DL Link

    • 3 replies
    1. K
      Karu Sankaralingam @karu
        2025-11-04 04:57:11.444Z

        Review Form

        Reviewer: The Guardian (Adversarial Skeptic)


        Summary

        The paper presents EOD, a co-designed hardware/software system for low-latency Graph Neural Network (GNN) inference in an inductive setting. The core idea is to mitigate the latency caused by neighborhood explosion and data preparation by precomputing hidden features for all training nodes. To manage the resulting memory overhead, the authors propose a "multi-layer concatenate" compression scheme (ZVC). These algorithmic changes are supported by a custom DIMM-based Near-Memory Processing (NMP) architecture designed for efficient aggregation of the precomputed, compressed features.

        While the paper identifies a valid and important problem, the proposed solution rests on a foundation of precomputation that introduces critical methodological flaws. The evaluation framework compares fundamentally different amounts of online work, leading to inflated performance claims, and fails to account for the substantial hidden costs required to maintain the system's accuracy, thereby undermining its claimed real-world viability.

        Strengths

        1. Clear Problem Identification: The paper correctly identifies that the Preparation and Memcpy steps, rather than pure computation, are the primary bottlenecks for real-time GNN inference services. The latency breakdown in Figure 3(a) effectively motivates the need to address these stages.

        2. Novel Compression Heuristic: The observation of an inverse relationship in sparsity between consecutive GNN layers (Figure 5) and the proposed multi-layer concatenation method to exploit this for ZVC compression (Section 4.2) is a clever algorithmic insight. This technique effectively increases the minimum number of zero-values per node, enhancing the efficacy of the chosen compression scheme.

        Weaknesses

        1. Fundamental Evaluation Flaw: Unaccounted Maintenance Cost: The entire premise of precomputation introduces data staleness. As new nodes and edges are added during online service, the precomputed embeddings for training nodes become outdated. The authors acknowledge this and show the resulting accuracy degradation in Figure 15. Their proposed solution is a "periodic re-precomputation (refresh)." This refresh operation is computationally equivalent to performing GNN inference on the entire training graph, a massively expensive and high-latency process. This cost is completely omitted from the performance evaluation. A system that requires periodic high-latency downtime or background computation to remain accurate cannot be fairly evaluated on its low-latency inference capabilities alone. This omission is a critical flaw that invalidates the claims of providing a practical low-latency solution.

        2. Misleading Baseline Comparison: The headline speedup claims (e.g., 17.9x geometric mean end-to-end) are derived from comparing the proposed EOD system against a standard GPU baseline. This is an apples-to-oranges comparison. The GPU baseline performs the full L-hop neighborhood traversal and feature gathering online (Preparation step), while EOD offloads this expensive traversal to an offline precomputation step. EOD's online workload is fundamentally smaller and simpler. A fair comparison would be against a GPU baseline that also leverages precomputed data (the "GPUpre" case in Figure 14). As the paper itself states on page 12, the speedup of EOD over GPUpre is a much more modest 1.14-1.35×. The massive reported speedups are an artifact of an inequitable experimental setup, not a revolutionary performance gain in like-for-like inference.

        3. Overstated Aggregation Performance: The paper prominently features aggregation speedups of over 900x (Abstract, Figure 13). While technically representing the performance on one sub-task, this is misleading. For a system paper focused on end-to-end latency, this cherry-picked metric dramatically inflates the perceived contribution. The end-to-end latency, which is the only metric that matters to the end-user, shows far more modest gains.

        4. Unquantified Cumulative Accuracy Loss: The proposed system introduces at least three distinct approximations:
          a. Pruning of target-to-train edges (Section 4.1).
          b. An "adjusted ReLU threshold" to increase sparsity (Section 4.2).
          c. The inherent staleness of precomputed embeddings between refreshes (Section 6.4).
          The paper analyzes these in isolation (or not at all, in the case of pruning's specific % drop) but never presents a clear analysis of their cumulative impact. The final accuracy of the EOD system at the moment just before a refresh cycle, compared to a gold-standard model with no approximations, is never reported. This leaves the true accuracy cost of the system entirely ambiguous.

        Questions to Address In Rebuttal

        1. What is the wall-clock time required for the "periodic re-precomputation" on the datasets evaluated (e.g., Products, Reddit)? How does this "maintenance latency" compare to the aggregated inference latency served during one refresh period? Please justify how a system requiring such a costly refresh operation can be considered a "low-latency" solution in a continuously operating online service.

        2. Please justify the use of the standard GPU as the primary baseline for end-to-end speedup claims, given that it performs a fundamentally larger online workload (L-hop traversal) than EOD. Why should the results not be primarily framed in comparison to the "GPUpre" baseline, which represents a more direct, apples-to-apples comparison of inference hardware for a precomputation-based algorithm?

        3. Provide a table detailing the final, end-to-end inference accuracy of the EOD system under the combined effects of all three approximations (pruning, ReLU thresholding, and data staleness at the end of a refresh cycle) compared to a non-approximated baseline model.

        4. The multi-layer concatenation compression scheme appears to depend on the properties of the ReLU activation function. How does this technique perform with other common GNN models that use different activation functions (e.g., LeakyReLU, GeLU) or architectures that do not rely on simple activation sparsity (e.g., GAT)? Please comment on the generality of this core contribution.

        1. K
          In reply tokaru:
          Karu Sankaralingam @karu
            2025-11-04 04:57:21.957Z

            Of course. Here is a peer review of the paper from the perspective of "The Synthesizer."


            Review Form:

            Reviewer: The Synthesizer (Contextual Analyst)

            Summary

            This paper presents EOD, a hardware-software co-designed system aimed at tackling the critical challenge of low-latency inference for Graph Neural Networks (GNNs) in an inductive, real-time setting. The authors correctly identify that the primary bottlenecks are not just the GNN computation itself, but the extensive data preparation and host-to-device data transfer, exacerbated by the "neighborhood explosion" problem.

            The core contribution is an elegant decoupling of the inference process. The authors propose to precompute the computationally heavy and data-intensive propagation among the existing "train" nodes offline. This transforms the online inference problem from an L-hop graph traversal to L separate 1-hop aggregations from the precomputed train node features to the new "target" nodes. To manage the storage overhead of these precomputed features, they introduce a novel "concatenated ZVC" compression scheme that exploits sparsity patterns across different GNN layers. This algorithmic approach is supported by a DIMM-based Near-Memory Processing (NMP) architecture designed specifically to perform the parallel 1-hop aggregations on the compressed data, thereby minimizing data movement and leveraging memory-side parallelism. The result is a system that dramatically reduces end-to-end latency for small-batch GNN serving.

            Strengths

            The true strength of this paper lies in its insightful problem formulation and the holistic, co-designed solution it proposes.

            1. Addressing the Right Problem: The GNN acceleration literature is crowded with work on full-batch training or transductive inference. This paper wisely targets the inductive, mini-batch inference scenario, which is far more representative of real-world, latency-sensitive applications like fraud detection and real-time recommendation. The analysis in Section 3 (page 4), which clearly shows that data preparation and memory copy dominate latency, is a crucial observation that correctly frames the entire problem and motivates the need for a system-level solution beyond a simple compute accelerator.

            2. Elegant Algorithmic/Hardware Co-Design: This is not a paper that simply throws hardware at a problem. The algorithmic innovation—precomputing train-to-train propagation—is the key enabler. This single decision fundamentally changes the nature of the online workload, making it vastly more regular and amenable to a specialized NMP architecture. The subsequent compression scheme (Section 4.2, page 5) is a clever and necessary component to make the precomputation practical from a memory capacity perspective. The hardware is therefore not a generic NMP solution, but one tailored to execute the specific "concatenate aggregation" task enabled by the algorithm. This synergy is the paper's most compelling feature.

            3. Connecting Disparate Research Threads: This work serves as an excellent synthesis of several important trends in computer science. It sits at the intersection of:

              • GNNs: Taking models like GraphSAGE out of the lab and into production environments.
              • Near-Memory Processing: Applying the principles pioneered in domains like recommendation systems (e.g., RecNMP) to the unique data access patterns of GNN inference.
              • Systems for ML: Recognizing that deploying ML is a full-stack problem, where data movement, I/O, and preprocessing are often more critical than the matrix multiplications themselves.

            By building these bridges, the paper provides a valuable blueprint for future research in practical ML systems.

            Weaknesses

            The weaknesses of the work are not in its core idea, which is sound, but in the assumptions that bound its current applicability. As a synthesizer, I see these less as flaws and more as the most fertile grounds for future work.

            1. The Static "Training Graph" Assumption: The precomputation strategy is highly effective but hinges on the set of training nodes and their features being largely static. In many real-world systems (e.g., social networks or e-commerce platforms), the graph is constantly evolving with new users, products, and interactions that become part of the "known" graph. The paper's proposed solution of "periodic re-precomputation" (Section 6.4, page 12) is a practical but reactive fix. This approach avoids the crucial question of how to incrementally and efficiently update the precomputed embeddings as the base graph changes, which is a significant challenge for real-world deployment.

            2. Single-Node System Abstraction: The evaluation is performed in the context of a single server with an in-memory graph database. This is a reasonable starting point, but the largest and most valuable graphs are almost always distributed across a cluster of machines. The paper acknowledges this limitation in the discussion (Section 7, page 13). In a distributed setting, the data-fetching and aggregation step would involve network latency, which could easily become the new dominant bottleneck, potentially negating some of the benefits of NMP. The paper would be strengthened by a more thorough discussion of how the EOD paradigm might extend to a distributed memory or storage environment.

            3. Limited Generality of Observations: The "concatenated ZVC" compression method is motivated by the empirical observation that sparsity in hidden features can be inversely correlated across layers (Figure 5, page 6). This is an interesting finding, but its generality is not established. It may be a specific artifact of the GraphSAGE model with ReLU activations on the tested datasets. It is unclear if this property holds for other popular architectures (e.g., GATs) or different activation functions, which could limit the effectiveness of the proposed compression scheme.

            Questions to Address In Rebuttal

            1. Could the authors elaborate on the cost and complexity of the "refresh" or re-precomputation step? In a production environment with a constant stream of updates, how would one determine the optimal refresh frequency to balance the trade-off between inference accuracy (which decays as the graph becomes stale) and the computational cost of the refresh?

            2. While a full distributed implementation is beyond scope, could the authors speculate on the architectural changes EOD would require to function in a sharded graph environment? For instance, would NMP modules need to communicate with each other (e.g., via a technology like DIMM-Link, mentioned in their citations), and how would the precomputed features be managed and accessed across the cluster?

            3. Regarding the concatenated ZVC compression: Have you investigated if the observed inverse sparsity correlation holds for GNN models other than GraphSAGE, or for different non-linearities besides ReLU? How critical is a high compression ratio to the overall performance of EOD, and how gracefully does the system's performance degrade if the compression is less effective?

            1. K
              In reply tokaru:
              Karu Sankaralingam @karu
                2025-11-04 04:57:32.609Z

                Review Form

                Reviewer: The Innovator (Novelty Specialist)

                Summary

                This paper, "EOD: Enabling Low Latency GNN Inference via Near-Memory Concatenate Aggregation," proposes a co-designed hardware/software system to accelerate real-time, inductive Graph Neural Network (GNN) inference. The authors identify two primary bottlenecks: the data preparation/transfer overhead and the "neighborhood explosion" problem.

                The core of their proposed solution consists of three main components:

                1. An algorithmic optimization based on precomputing the hidden features for all training nodes (tr-to-tr propagation), thereby reducing the online inference workload to L separate 1-hop aggregations instead of a single L-hop aggregation.
                2. A compression scheme that concatenates the hidden features of a node across all GNN layers before applying Zero Value Compression (ZVC), which aims to improve the compression ratio by averaging out sparsity variations between layers.
                3. A DIMM-based Near-Memory Processing (NMP) architecture specifically designed to accelerate the tr-to-tar aggregation step on data stored in this concatenated and compressed format.

                While the paper demonstrates substantial performance improvements, the novelty of the core constituent ideas is limited. The primary contribution lies in the specific synthesis and co-optimization of these components into a functional and high-performance system for a very specific workload.

                Strengths

                1. Novel System-Level Integration: The primary strength of this work is the coherent and tightly-coupled integration of pre-existing concepts (precomputation, ZVC, NMP) into a specialized system. While the individual parts are not new, their combination to solve the low-latency inductive GNN inference problem is a novel system-level engineering contribution.

                2. Minor Algorithmic Novelty in Compression: The idea of concatenating hidden features across multiple layers before applying ZVC (Section 4.2, Pages 5-6) is a clever, albeit incremental, optimization. By combining feature vectors that may have anti-correlated sparsity patterns (as shown in Figure 5), the authors create a more favorable data distribution for compression. I have not seen this specific technique applied in prior GNN acceleration work.

                3. Specialized Hardware Co-design: The NMP architecture is novel in its specialization. It is not a generic NMP unit; it is purpose-built to handle the proposed concatenated ZVC data format, including a custom instruction (Agg-instruction, Figure 10, Page 8) and an integrated ZVC decompressor (Figure 9d, Page 8). This demonstrates a deep co-design.

                Weaknesses

                The central weakness of this paper, from a novelty perspective, is that its foundational algorithmic and architectural pillars are built upon well-established prior art.

                1. Core Algorithmic Premise is Not Novel: The main algorithmic trick—pruning tar-to-tr edges and precomputing tr-to-tr propagation to simplify online inference (Section 4.1, Page 4-5)—is not a new idea. This exact concept of decoupling training and test node computations to accelerate inductive inference has been explored before. Specifically, the work by Si et al., "Serving graph compression for graph neural networks" (ICLR 2023) [51], which the authors cite, proposes this very same decoupling strategy as its central contribution. The current paper re-implements this known technique as the basis for its hardware acceleration, but the fundamental insight is not original.

                2. Architectural Paradigm is Not Novel: The use of DIMM-based NMP for accelerating aggregation-heavy workloads is a known pattern. Prior works such as RecNMP [24] and Tensordimm [28] established the viability of placing logic on a DIMM's buffer chip to process embedding table lookups for recommendation models. More directly, GNNear [64] and GraNDe [61] have already proposed using DIMM-based NMP to accelerate GNNs. EOD follows this established architectural template. Its novelty is confined to the specific logic implemented within the NMP unit, which is tailored to its specific data format, rather than a fundamentally new NMP architecture.

                3. Compression Primitive is Standard: The use of Zero Value Compression (ZVC) to exploit activation sparsity is a standard technique in ML accelerators, as seen in works like Rhu et al., "Compressing DMA engine" (HPCA 2018) [47]. The novelty in EOD lies only in the pre-processing step (concatenation), not in the compression mechanism itself.

                In essence, the paper takes a known algorithm from [51], implements it on a known architectural template from [24, 61, 64], and uses a standard compression primitive [47] with a minor pre-processing twist. The impressive speedup numbers are a result of aggressively applying this known precomputation algorithm, which fundamentally changes the problem from an L-hop graph traversal to L independent table lookups—a task for which NMP is known to be highly effective.

                Questions to Address In Rebuttal

                1. The core precomputation strategy described in Section 4.1 appears functionally identical to the decoupling method proposed in your cited work [51]. Please clarify, with technical precision, what the novel algorithmic difference is between your method and the one in [51]. If there is no significant difference, please justify why building hardware for a known algorithm constitutes a sufficient contribution for this venue.

                2. The multi-layer concatenation technique for ZVC is presented as a key enabler. Is this technique generally applicable to other GNN acceleration scenarios, or is its utility strictly limited to a context where all layers' hidden features are precomputed and stored? Its novelty is proportional to its generality.

                3. Given that the primary performance gain comes from transforming the problem into one that is embarrassingly parallel (L independent 1-hop aggregations), how much of the benefit is from the NMP architecture versus the algorithm itself? Figure 14a shows that even a "GPUpre" baseline (precomputation on GPU) achieves a significant speedup (12.0-16.4x). This suggests the algorithm is the dominant factor. Please argue for the significance of the NMP architecture's novel aspects beyond simply being a good fit for a known memory-bound problem.