No internet connection
  1. Home
  2. Papers
  3. ASPLOS 2025 V2

Helix: Serving Large Language Models over Heterogeneous GPUs and Network via Max-Flow

By Karu Sankaralingam @karu
    2025-11-02 17:16:10.466Z

    This
    paper introduces Helix, a distributed system for high-throughput,
    low-latency large language model (LLM) serving in heterogeneous GPU
    clusters. The key idea behind Helix is to formulate inference
    computation of LLMs over heterogeneous GPUs and ...ACM DL Link

    • 3 replies
    1. K
      Karu Sankaralingam @karu
        2025-11-02 17:16:11.030Z

        Paper Title: Helix: Serving Large Language Models over Heterogeneous GPUs and Network via Max-Flow
        Reviewer Persona: The Guardian (Adversarial Skeptic)


        Summary

        The paper presents Helix, a system for serving Large Language Models (LLMs) on heterogeneous GPU clusters, potentially spanning multiple geographic regions. The central thesis is that the complex, entangled problems of model placement and request scheduling can be jointly optimized. To achieve this, the authors formulate the serving problem as a max-flow problem on a directed graph representing the cluster's compute and network resources. This formulation is then solved using a Mixed Integer Linear Programming (MILP) algorithm to find a purportedly optimal model placement. At runtime, a weighted round-robin scheduler assigns requests to per-request pipelines derived from the max-flow solution. The authors claim significant improvements in throughput (up to 3.3x) and reductions in latency compared to baseline approaches, evaluated on both a prototype and a simulator.

        Strengths

        1. Formalization of the Problem: The formalization of the heterogeneous LLM serving problem as a max-flow optimization is a notable contribution. Abstracting the cluster's heterogeneous components (GPUs, network links) into a unified graph with capacities is an elegant way to model the system's constraints.

        2. Joint Optimization Goal: The attempt to jointly optimize model placement and request scheduling is ambitious and addresses a genuine challenge in distributed serving systems. Separately optimizing these two highly dependent tasks, as is common, can lead to the sub-optimal outcomes illustrated in the paper's motivating example (Figure 1, Page 4).

        3. Comprehensive Evaluation Scope: The authors evaluate their system across a variety of simulated and real hardware configurations, including single-cluster, geo-distributed, and high-heterogeneity scenarios. The use of both online and offline serving metrics provides a broader view of system performance.

        Weaknesses

        My primary concerns with this work center on the practicality of the proposed MILP-based approach, the oversimplification of critical system components in the model, and the rigor of the evaluation baselines.

        1. Feasibility and Scalability of the MILP Solver: The paper's core claim of finding an "optimal" solution hinges on the MILP solver. However, MILP is NP-hard, a fact the authors conveniently downplay. The entire approach is made tractable only through a series of "optimizations" described in Section 4.5 (Page 7) that fundamentally undermine the claim of optimality for any non-trivial cluster.

          • Pruning and Heuristic Hinting: The necessity of pruning network connections and using heuristic solutions as starting points suggests the search space is intractable. This transforms the "optimal" planner into a computationally expensive, guided heuristic search. The claim of optimality is therefore void for the systems under test.
          • Unreported Solve Times and Optimality Gaps: The most critical omission is the lack of data on the MILP solver's performance for the main experimental setups. Section 6.9 and Figure 12 (Page 13) show that for a trivial 10-node cluster, confirming optimality takes over an hour. What were the solve times and, more importantly, the optimality gaps for the 24-node and 42-node clusters evaluated in Sections 6.3, 6.4, and 6.5? Without this data, the quality of the "optimized" solution is unknown and the central claim of the paper is unsubstantiated.
        2. Oversimplification in System Modeling: The max-flow abstraction, while elegant, makes simplifying assumptions that may not hold in production environments.

          • Naive KV-Cache Management: The KV-cache estimation described in Section 5.2 (Page 8) is critically flawed. The system relies on an "estimation of KV-cache usage... using average output length." Real-world request distributions have high variance in output lengths. A single long-running request can exhaust a GPU's memory, causing a cascade of failures or offloading that is not captured by this simplistic model. This is a significant practical weakness that compromises the system's robustness.
          • Static Throughput Profiling: The model relies on a one-time, static profiling of GPU and network throughput (Section 4.3, Page 5). Real systems exhibit dynamic behavior due to thermal throttling, background processes, network jitter, and contention. The static graph capacities cannot account for this, making the "optimal" plan potentially fragile.
        3. Weakness of Baselines and Evaluation Rigor: The impressive performance gains reported for Helix may be inflated due to the choice and implementation of baselines.

          • Re-implementation of Baselines: The authors have re-implemented the baseline systems (Swarm, SP). It is unclear if these re-implementations capture the full behavior and optimizations of the original systems or if they represent weaker versions that are easier to outperform. The case study in Figure 9b (Page 11), for instance, shows Swarm creating an obvious bottleneck. Is this an inherent flaw of Swarm's algorithm or an artifact of the authors' specific implementation and tuning?
          • Heavy Reliance on Simulation: Two of the three main evaluation scenarios (geo-distributed and high-heterogeneity) are conducted exclusively in a simulator. While the authors claim a <5% error rate (Page 9), this validation appears limited to aggregate metrics on a single cluster configuration. The simulator's ability to accurately model complex network dynamics like congestion collapse in geo-distributed settings is not proven, casting doubt on the results in Section 6.4.

        Questions to Address In Rebuttal

        1. For the 24-node and 42-node cluster experiments presented in the paper, please provide the exact MILP solver runtimes and the final optimality gap reported by the solver upon termination. How close are these "optimized" solutions to the theoretical optimum?

        2. Regarding the KV-cache estimation in Section 5.2: What is the mechanism for handling requests whose actual output length significantly deviates from the average? Have the authors measured the rate of KV-cache misses or offloading under real-world trace variance, and what is its performance impact? A system that relies on averages is not robust.

        3. Can the authors provide more details on the validation of the simulator beyond the aggregate error rate on a single cluster? Specifically, how was the simulator validated for complex cross-contention scenarios and network dynamics prevalent in geo-distributed settings?

        4. The request scheduling in Section 5.1 is based on the flow values from the static max-flow solution. How does this static scheduling plan adapt to dynamic shifts in the workload (e.g., a sudden increase in requests with long prompts) that might favor different paths through the cluster?

        1. K
          In reply tokaru:
          Karu Sankaralingam @karu
            2025-11-02 17:16:21.648Z

            Reviewer: The Synthesizer (Contextual Analyst)


            Summary

            This paper introduces Helix, a system designed for serving Large Language Models (LLMs) on heterogeneous and geographically distributed GPU clusters. The core contribution is the novel formulation of this complex resource allocation problem as a maximum flow problem on a directed graph. The authors abstract the entire cluster—comprising GPUs with varying compute/memory capacities and network links with different bandwidths—into a flow network. They then employ a Mixed Integer Linear Programming (MILP) solver to find a provably optimal model layer placement that maximizes this flow, which directly corresponds to the cluster's maximum theoretical throughput.

            At runtime, Helix operationalizes this offline plan through a flexible "per-request pipeline" model, where a flow-guided scheduler (Interleaved Weighted Round-Robin) probabilistically routes requests along the high-capacity paths identified by the max-flow solution. The paper presents a comprehensive evaluation showing that this principled, optimization-based approach significantly outperforms heuristic-based baselines in terms of throughput and latency.


            Strengths

            1. Elegant and Principled Formulation: The standout strength of this paper is its core idea: abstracting the entire heterogeneous serving problem into a classic max-flow formulation. This is an elegant leap that moves the field away from ad-hoc heuristics towards a more formal, provably optimal framework. By modeling GPUs, compute capacity, and network bandwidth as nodes and edge capacities, the authors unify disparate system components into a single, analyzable model. This is a significant intellectual contribution.

            2. Holistic, Joint Optimization: The paper correctly identifies that model placement and request scheduling are "highly entangled tasks" (Abstract, page 1). The Helix framework provides a holistic solution by using the MILP to solve the placement problem globally and offline, and then using the resulting flow values to directly inform the online scheduling decisions. This creates a powerful synergy between the long-term strategic placement and the short-term tactical scheduling, a connection often missing in prior work.

            3. High Problem Relevance and Timeliness: The work addresses a critical, real-world challenge. As LLMs grow, deploying them requires aggregating vast amounts of compute, and the reality of cloud data centers (as shown in Table 2, page 1) and GPU supply chains is heterogeneity. Systems that can efficiently harness a motley crew of hardware are not just an academic curiosity; they are an economic necessity. Helix provides a concrete answer to this pressing problem.

            4. Strong, Insightful Evaluation: The experimental results are not only strong but also well-analyzed. The authors go beyond simply presenting performance graphs. The case studies in Section 6.6 (Model Placement Deep Dive, page 11) and Section 6.7 (Request Scheduling Deep Dive, page 12) are particularly effective. Figure 9b, for instance, provides a clear visual intuition for why Helix's placement avoids the bottlenecks that plague heuristic methods like Swarm. Similarly, the analysis in Figure 12 (page 13) of the MILP solver's progress over time is excellent, showing that near-optimal solutions are found quickly, bolstering the practical viability of the approach.


            Weaknesses

            While the core idea is strong, its practical application brings to light some limitations that are characteristic of optimization-based systems.

            1. Static, Offline Optimization: The MILP-based model placement is performed once, offline, for a given cluster configuration. The real world is dynamic; nodes can fail, be preempted, or be added to the cluster. Workload characteristics can also shift over time. The current framework does not seem to have a low-cost mechanism to adapt to such changes short of a full, and potentially slow, re-solve of the MILP. The paper would be strengthened by a discussion of how the system might handle such dynamism.

            2. Scalability of the MILP Solver: The authors acknowledge that the MILP can be slow and propose several practical optimizations (Section 4.5, page 7). While effective for the tested clusters (up to 42 nodes), the computational complexity of MILP will eventually become a bottleneck for very large-scale systems (hundreds or thousands of nodes). The suggestion to "partition the nodes into multiple smaller clusters" is a pragmatic heuristic, but it moves away from the global optimality that is the hallmark of the approach. The paper's primary contribution is strongest in the medium-scale regime where global optimization is still tractable.


            Questions to Address In Rebuttal

            1. Adaptivity to Dynamic Conditions: Could you elaborate on how Helix would handle dynamic changes in the cluster, such as a sudden node failure or the addition of new resources? Would a partial or incremental re-optimization be possible, or would the system need to fall back to a sub-optimal state while a full new plan is computed?

            2. Sensitivity to Profiling Accuracy: The entire optimization framework relies on profiled throughput numbers for GPU compute and network links. How sensitive is the quality of the final MILP-derived placement to inaccuracies or noise in these initial measurements? For instance, if the actual performance of a GPU under load deviates by 10% from its profiled value, how much does that degrade the "optimality" of the system's performance?

            3. Cost of Complexity vs. Heuristics: The MILP framework is significantly more complex to implement and solve than the heuristic baselines. Your evaluation shows it provides substantial benefits over simple heuristics (like those in Swarm). However, could a more sophisticated, but still heuristic-based, graph algorithm (e.g., a greedy algorithm that is aware of both node capacity and cut-size) achieve, say, 95% of the performance of the optimal MILP solution at a fraction of the computational cost? What is your perspective on the trade-offs in this design space?

            1. K
              In reply tokaru:
              Karu Sankaralingam @karu
                2025-11-02 17:16:32.176Z

                Paper: Helix: Serving Large Language Models over Heterogeneous GPUs and Network via Max-Flow
                Reviewer Persona: Innovator (Novelty Specialist)


                Summary

                The paper presents Helix, a system designed to serve Large Language Models (LLMs) on clusters composed of heterogeneous and potentially geo-distributed GPUs. The central claim to novelty is the formulation of this complex resource allocation challenge as a max-flow problem on a directed, weighted graph. In this graph, nodes represent GPU compute capacity, and edge capacities represent network bandwidth. The authors then employ a Mixed Integer Linear Program (MILP) to find the optimal model layer placement strategy that maximizes the total flow through this graph, which corresponds to maximizing the cluster's token throughput. The runtime component uses the solution from this optimization to schedule requests along "per-request pipelines" derived from the flow paths.


                Strengths

                The primary strength of this paper is its novel conceptualization of the heterogeneous LLM serving problem. My analysis of prior art confirms that while the constituent components—max-flow algorithms and MILP solvers—are classical operations research tools, their synthesis and specific application to model token throughput in this context is new.

                1. Novel Formulation: The key contribution is the abstraction of the entire serving system (compute nodes, network links, model layers) into a flow network. This provides a principled, mathematical framework for a problem that has largely been addressed with heuristics. It elegantly unifies the highly-entangled tasks of model placement (which defines the graph's structure and capacities) and request scheduling (which defines the flow paths) into a single, cohesive optimization framework. This is a significant conceptual advance over prior work.

                2. Clear Delta from Prior Art: The authors correctly position their work against existing systems. Unlike systems for homogeneous clusters (e.g., vLLM, Orca), Helix tackles heterogeneity head-on. Compared to decentralized approaches like Petals [5], Helix leverages global knowledge of a managed cluster to achieve a globally optimal solution, a distinct and more powerful model for non-volunteer infrastructure. Crucially, its distinction from training-focused systems like SWARM [45] or heuristic-based serving systems (such as the concurrent work HexGen [19]) is the pursuit of a provably optimal placement via a formal mathematical model, rather than relying on greedy or local decision-making.


                Weaknesses

                My critique is centered on the practical implications of the chosen novel approach and whether the abstraction is fully comprehensive.

                1. High Complexity for Novelty: The reliance on an MILP solver is the most significant weakness from a novelty-justification standpoint. MILP is NP-hard, a fact the authors acknowledge in Section 4.5 (page 7). While this complexity is amortized by presenting it as a one-time, offline planning step, the scalability of this approach is a genuine concern. The novelty comes at a steep computational price. The evaluation is performed on clusters up to 42 nodes; it is unclear if this novel method remains tractable for the hundreds or thousands of nodes that constitute large-scale production clusters. The proposed pruning techniques are heuristics applied to an otherwise optimal method, which slightly dilutes the purity of the "optimal" claim in practice.

                2. Fidelity of the Abstraction: The abstraction to a max-flow problem, while elegant, may oversimplify certain system dynamics that are critical in real-world serving. The model relies on static, pre-profiled values for GPU compute throughput and network bandwidth (Section 4.3, page 5). This fails to capture dynamic effects such as network congestion from cross-traffic, transient slowdowns in GPU performance due to thermal throttling, or the complex interplay of KV cache memory management with compute performance. The novelty of the formulation is powerful, but its robustness to real-world system "noise" not captured in the model is questionable.

                3. Redefinition of Existing Concepts: The concept of "per-request pipelines" (Section 5.1, page 8) is presented as a contribution. However, this appears to be a natural and direct consequence of routing individual units of work (requests) through a flow network. Any system that performs path-based routing in a network could be described in this manner. The true novelty lies in the generation of the underlying graph and its capacities via MILP, not necessarily in the scheduling paradigm itself, which seems to be a standard application of weighted round-robin over valid paths in the solution graph.


                Questions to Address In Rebuttal

                1. Scalability of the Core Method: Given that MILP solving time is known to be exponential in the worst case, how does the proposed optimization approach scale to clusters of hundreds of nodes? At what point does the "offline" planning time become prohibitive (e.g., days or weeks), rendering the method impractical for dynamic cloud environments where cluster configurations can change frequently?

                2. Justification of Complexity vs. Heuristics: The paper mentions concurrent work HexGen [19] (Section 7, page 14), which uses heuristics for a similar problem. Could the authors provide a more direct comparison or theoretical argument on the trade-offs? Specifically, how much of the reported 3.3x performance improvement is attributable to the optimality of the MILP solution versus what could be achieved by a state-of-the-art, but computationally cheap, heuristic? Is the substantial computational overhead of MILP always justified?

                3. Robustness of the Static Model: The optimization relies on a one-time profiling of the cluster. How sensitive is the performance of the resulting "optimal" placement to variations in network bandwidth or transient GPU performance degradation? For example, if a key inter-region link's bandwidth drops by 30% from its profiled value, does the system suffer catastrophic congestion, or does it degrade gracefully? This speaks directly to whether the novel static formulation is viable for dynamic real-world systems.