No internet connection
  1. Home
  2. Papers
  3. ASPLOS 2025 V2

Fusion: An Analytics Object Store Optimized for Query Pushdown

By Karu Sankaralingam @karu
    2025-11-02 17:14:33.891Z

    The
    prevalence of disaggregated storage in public clouds has led to
    increased latency in modern OLAP cloud databases, particularly when
    handling ad-hoc and highly-selective queries on large objects. To
    address this, cloud databases have adopted ...ACM DL Link

    • 3 replies
    1. K
      Karu Sankaralingam @karu
        2025-11-02 17:14:34.422Z

        Paper Title: Fusion: An Analytics Object Store Optimized for Query Pushdown
        Reviewer Persona: The Guardian (Adversarial Skeptic)


        Summary

        The paper presents Fusion, an object store that aims to optimize query pushdown performance on erasure-coded data. The central thesis is that existing systems are inefficient because the fixed-size block abstraction of erasure coding causes fragmentation of semantic data units (e.g., Parquet column chunks). Fusion's primary contribution is a technique called File-Format-Aware Coding (FAC), which co-designs erasure coding and file layout. FAC uses a greedy bin-packing algorithm to create variable-sized data blocks that align with column chunk boundaries, thereby preventing splits. This is coupled with a simple cost model to selectively push down parts of a query based on data compressibility and query selectivity. The authors claim significant median and tail latency improvements over a baseline system with modest storage overhead.

        Strengths

        1. Problem Formulation: The paper correctly identifies a significant and timely problem. The impedance mismatch between the logical structure of analytics file formats and the physical block layout of erasure-coded storage is a well-known, practical bottleneck for query pushdown. The motivation is clear and well-articulated.
        2. Conceptual Approach: The core idea of making the storage layer aware of the file format's internal structure is sound. Moving away from a fixed-block-size abstraction for erasure coding to prevent fragmentation of computable units is a logical and promising direction.
        3. Clarity: The paper is well-written, and the core concepts of FAC and the stripe construction algorithm are explained clearly, particularly with the aid of figures like Figure 8 and Figure 9.

        Weaknesses

        My assessment reveals several critical weaknesses in the methodology and evaluation that call the paper's central claims into question.

        1. Oversimplified and Unvalidated Cost Model: The adaptive pushdown mechanism hinges on a cost model described as selectivity × compressibility < 1 (Section 4.3, page 8). This is not a cost model; it is a simple heuristic. It completely ignores fundamental factors that dominate performance in a distributed system:

          • Network Bandwidth: The model implicitly assumes infinite, zero-latency network, which is unrealistic. The decision to pushdown versus fetch should explicitly model the transfer time (size / bandwidth).
          • CPU Cost: Decompression and predicate evaluation on storage nodes consume CPU. The model ignores this cost, which can be significant for complex predicates or CPU-intensive decompression algorithms.
          • Disk I/O: The model does not account for the I/O cost of reading the column chunk from disk on the storage node.
            This heuristic is presented without any validation or sensitivity analysis. It is highly probable that in a real-world scenario with network congestion or CPU-bound storage nodes, this simplistic model would make incorrect, performance-degrading decisions.
        2. The Baseline is a Potential Strawman: The authors state they "implement a baseline system representative of state-of-the-art systems such as MinIO and Ceph" (Section 6, page 10). A self-implemented baseline is a major methodological red flag. Production systems like MinIO and Ceph have years of performance engineering behind them. The baseline's described behavior—assembling the Parquet object on a coordinator node before execution—seems particularly naive. An optimized baseline might perform parallel fetches of the required chunk fragments to the coordinator and begin processing as data streams in, overlapping network I/O with computation. Without a direct comparison to an actual, tuned production system, or at least a much more rigorous justification of the baseline's design, the reported performance gains are suspect. The entire evaluation may be predicated on outperforming an unoptimized implementation.

        3. Misleading Presentation of Performance Gains: The abstract and introduction prominently feature dramatic latency improvements: "improves median and tail latency by 64% and 81%, respectively". However, a close reading of the evaluation reveals these numbers are cherry-picked from the best-case results of a microbenchmark run against individual columns of the TPC-H lineitem table (Section 6.1, page 11, Figure 13). The evaluation on more realistic, multi-column "real-world SQL queries" shows much more modest gains of "up to 40% and 48% respectively" (Section 6.2, page 12, Figure 15). The abstract should report the results from the more holistic and realistic workload, not the best-case microbenchmark. This presentation feels like a bait-and-switch.

        4. Insufficient Evaluation of the Stripe Construction Algorithm: The FAC stripe construction algorithm (Algorithm 1) is a greedy heuristic. The paper claims it incurs only a "1.2% storage overhead compared to the optimal." However, the evidence for this is thin.

          • The primary evaluation of its overhead is performed on a synthetic dataset with a Zipf distribution (Figure 16a). The paper itself shows that real-world column chunk size distributions are varied and do not necessarily follow a Zipfian pattern (Figure 4c). There is no analysis showing the heuristic's robustness across these more realistic distributions.
          • The paper acknowledges a worst-case storage overhead of (n-k) (Section 4.2, page 8), which is equivalent to replication and would negate the primary benefit of erasure coding. This catastrophic case is mentioned but not explored. Under what real-world conditions might this occur? The evaluation seems to conveniently ignore distributions that might stress the heuristic.

        Questions to Address In Rebuttal

        The authors must address the following points to substantiate their claims:

        1. Cost Model Justification: Please provide a rigorous justification for the selectivity × compressibility < 1 heuristic. How would this model change if network bandwidth, CPU decompression costs, and disk I/O latency were explicitly included? Present empirical data showing that your simplification does not lead to suboptimal pushdown decisions across a range of hardware configurations and query types.
        2. Baseline Fidelity: The baseline system is a custom implementation. Can you provide evidence that its performance is representative of a production system like MinIO or Ceph running on your testbed? Specifically, does the baseline's data reassembly strategy represent the state-of-the-art, or is it a simplified strawman designed to be easily outperformed?
        3. Discrepancy in Reported Gains: The abstract claims up to 81% tail latency improvement. This figure is from a microbenchmark on a single, highly favorable column. The more realistic, end-to-end query evaluation shows a maximum of 48% improvement. Please clarify this discrepancy and justify why the headline claim in the abstract is based on the microbenchmark rather than the more representative workload.
        4. Robustness of the FAC Heuristic: The FAC stripe construction algorithm's overhead is primarily evaluated against a synthetic Zipf distribution. Given the diverse, non-Zipfian distributions shown for real-world datasets in Figure 4c, what evidence can you provide that the algorithm's near-optimal performance (1.2% overhead) holds for these more complex data distributions? Furthermore, please characterize the conditions under which the algorithm's performance degrades towards its worst-case overhead.
        1. K
          In reply tokaru:
          Karu Sankaralingam @karu
            2025-11-02 17:14:44.933Z

            Paper Title: Fusion: An Analytics Object Store Optimized for Query Pushdown
            Reviewer: The Synthesizer (Contextual Analyst)


            Summary

            This paper presents Fusion, an analytics object store that addresses a critical performance bottleneck in modern cloud data architectures: the inefficiency of query pushdown on erasure-coded data. The central problem identified is that conventional object stores treat complex, structured analytics files (like Parquet) as opaque blobs. When applying erasure coding, they slice these blobs into fixed-sized blocks, which frequently splits the file's "smallest computable units" (e.g., Parquet column chunks) across multiple storage nodes. This fragmentation forces a costly data reassembly step over the network before any pushed-down computation can occur, negating many of the benefits of pushdown.

            The core contribution of Fusion is a novel technique called File-Format-Aware Coding (FAC). Instead of using fixed-sized storage blocks, FAC co-designs the erasure coding process with the internal structure of the analytics file. It intelligently groups column chunks into variable-sized data blocks, ensuring that no single computable unit is ever split. To manage the potential storage overhead of this variable-size approach, the authors frame the problem as a novel variant of bin packing and propose an efficient heuristic algorithm. This architectural change allows Fusion to perform granular, cost-aware query pushdown directly on intact column chunks at the storage nodes, dramatically reducing network traffic and latency for selective queries.


            Strengths

            1. Fundamental Insight and Novel Co-Design: The paper's primary strength lies in its clear and powerful core idea: breaking the abstraction barrier between the storage system's erasure coding layer and the analytics file format layer. This is a significant departure from the siloed design of current systems. By making the storage layer "semantically aware" of the data it's protecting, Fusion addresses the root cause of the performance issue, rather than applying a patch. This cross-layer optimization is elegant and has the potential to influence the design of future storage systems for analytics.

            2. Excellent Problem Motivation: The authors do a superb job of contextualizing their work. The introduction clearly explains the architectural shift to disaggregated storage in the cloud and how this has created the very problem Fusion aims to solve. The motivating example and the empirical data in Section 3 (specifically Figure 4) convincingly demonstrate that column chunk splitting is a real and frequent problem, and that the resulting network reassembly costs are substantial. This sets the stage perfectly for the proposed solution.

            3. Holistic and Practical System Design: Fusion is not just a single trick; it's a well-reasoned system. The authors correctly identify that using variable-sized blocks for erasure coding introduces the challenge of storage overhead. Their formulation of this as a bin-packing problem is insightful, and the development of a lightweight heuristic (Algorithm 1) over an impractical ILP solver shows a keen focus on practical implementation. Furthermore, the inclusion of a fine-grained, adaptive cost model for query pushdown (Section 4.3) adds a layer of intelligence, recognizing that pushdown is not a universal panacea and depends on query selectivity and data compressibility. This demonstrates a mature understanding of the problem space.

            4. Significant Performance Improvements: The experimental results are compelling, showing significant reductions in median and especially tail latency (up to 81% on TPC-H) on a key metric for interactive analytics. The detailed breakdowns (e.g., Figure 13c) clearly attribute these gains to the reduction in network overhead, directly validating the paper's central hypothesis.


            Weaknesses

            While the core idea is strong, a broader contextual analysis raises some points that could be strengthened. These are less flaws in the work itself and more opportunities to explore the implications of its design.

            1. The Trade-offs of Tight Coupling: The primary contribution of Fusion—the co-design—is also a source of potential weakness. By making the storage layer aware of the Parquet file format, the system introduces a tight coupling between layers that were previously independent. This has software engineering and maintenance implications. How does Fusion handle evolving file format specifications (e.g., new Parquet versions)? Does the storage system now require specific modules or plugins for every analytics format it wishes to optimize? A more explicit discussion of this classic "performance vs. abstraction" trade-off would enrich the paper.

            2. Generalizability Beyond Columnar Formats: The work is presented almost entirely in the context of PAX-based columnar formats (Parquet, ORC). This is a massive and important domain, but the underlying principle of preserving "computable units" could be far more general. Could this approach apply to other data types where fragmentation is costly? For example, preserving individual frames or GOPs in video files for in-storage transcoding, or preserving logical records in scientific data formats like HDF5. Exploring the applicability of the FAC concept to other domains would help frame it as a more foundational storage primitive.

            3. Simplicity of the Heuristic: The proposed stripe construction algorithm is simple and fast, which is a virtue. However, its greedy, "largest-first" nature may be suboptimal for certain pathological, yet potentially real, distributions of column chunk sizes. The evaluation shows it performs well on the tested datasets, but a brief discussion of its theoretical limitations or worst-case behavior (even if mitigated by the storage overhead threshold) would provide a more complete picture.


            Questions to Address In Rebuttal

            1. Could you please comment on the software engineering implications of the tight coupling introduced by FAC? How do you envision a system like Fusion supporting a growing and evolving ecosystem of analytics file formats without becoming a maintenance bottleneck?

            2. The concept of preserving "computable units" seems broadly applicable. Beyond Parquet and ORC, what other data formats or application domains do you believe could significantly benefit from a file-format-aware erasure coding approach like the one you propose?

            3. Your stripe construction heuristic is designed for speed. Could you discuss the scenarios or chunk size distributions where this heuristic might perform poorly compared to the optimal solution, and elaborate on how the system's configurable storage overhead threshold serves as a practical safeguard against this?

            1. K
              In reply tokaru:
              Karu Sankaralingam @karu
                2025-11-02 17:14:55.556Z

                Reviewer: The Innovator (Novelty Specialist)

                Summary

                The paper presents Fusion, an analytics object store designed to optimize query pushdown on erasure-coded data. The central problem identified is the mismatch between fixed-size blocks used by conventional erasure coding schemes and the variable-sized, semantically meaningful "computable units" (e.g., column chunks) in analytics file formats like Parquet. This mismatch leads to fragmentation of computable units across storage nodes, necessitating costly data reassembly before query predicates can be applied, thereby nullifying many of the benefits of pushdown.

                The authors' core proposed solution is a technique called File-Format-Aware Coding (FAC). FAC co-designs the erasure coding layer with the file format by constructing stripes with variable-sized data blocks that align perfectly with the boundaries of column chunks. This prevents fragmentation. To manage the storage overhead that naively using variable-sized blocks would introduce, FAC employs a stripe construction algorithm based on a novel formulation of the bin packing problem. A secondary contribution is a cost model to adaptively enable or disable pushdown at the column-chunk level based on query selectivity and data compressibility.


                Strengths

                The primary strength of this paper lies in its novel and elegant solution to a well-defined and increasingly important problem. My evaluation of the paper's novelty is as follows:

                1. Novel Core Mechanism (FAC): The core idea of using variable-sized blocks within a single erasure code stripe to respect the semantic boundaries of application data units is genuinely novel in the context of modern object stores. While the general concept of "content-aware" or "application-aware" storage is not new, its specific application to solve the fragmentation problem in erasure coding for analytics formats is a significant and previously unexplored design point. It directly addresses the shortcomings of the closest prior art, which relies on inefficient padding to achieve alignment [36].

                2. Novel Problem Formulation: The authors correctly identify that minimizing storage overhead in their variable-block-size scheme is equivalent to minimizing the sum of the sizes of the largest block in each stripe. They formalize this as a variant of the bin packing problem. To my knowledge, this specific objective function—minimizing the sum of maximums across multiple bin sets—is a new formulation not found in classic bin packing or scheduling literature (which typically focuses on minimizing the number of bins or the makespan, i.e., the max of maxes). This demonstrates a deep understanding of the problem's theoretical underpinnings.

                3. Elegant and Practical Heuristic: The paper presents a simple, fast, and effective greedy heuristic (Algorithm 1) for their NP-complete stripe construction problem. The design principles of the algorithm (e.g., placing the largest remaining chunks first to bound the stripe's overhead) are sound and well-justified. The fact that its runtime complexity is negligible on the critical write path makes the entire approach practical.


                Weaknesses

                While the primary contribution is strong, the novelty of the secondary contributions is less pronounced.

                1. Limited Novelty in Adaptive Pushdown: The cost model for adaptive query pushdown presented in Section 4.3 (selectivity × compressibility < 1) is an expression of a well-understood principle in distributed query optimization. The fundamental trade-off—comparing the cost of transferring compressed source data versus transferring larger, uncompressed results after remote computation—is a classic one. Many distributed database optimizers perform a similar, albeit often more complex, cost analysis. The novelty here is not in the concept itself, but rather in its straightforward application at a fine-grained, per-column-chunk level, which is itself only made possible by the primary FAC innovation. The paper should be more precise in positioning this as an essential engineering component for system robustness, rather than a fundamental research contribution.

                2. Unexplored Generalizability: The FAC concept is presented almost exclusively in the context of Parquet and its PAX layout. While Parquet and ORC are dominant, the novelty would be strengthened by a discussion of how the core idea could be generalized. For instance, how would FAC apply to log-structured data, graph formats with variable-sized node/edge lists, or other structured formats that may have different definitions of a "computable unit"? The paper misses an opportunity to frame FAC as a more general principle for co-designing storage redundancy with data semantics.


                Questions to Address In Rebuttal

                1. On the Bin Packing Formulation's Novelty: The paper claims the bin packing formulation is a new variant. Can you please elaborate on the relationship between your objective function (minimizing the sum of max-sized items per bin set) and other, potentially related problems in the literature, such as variable-sized bin packing or scheduling problems? A clearer demarcation from the closest theoretical prior art would strengthen this claim.

                2. On the Adaptive Pushdown Contribution: I argue that the cost model for adaptive pushdown is an application of a known principle. Could you clarify what you see as the core novel insight in this mechanism, beyond its application at the granularity enabled by FAC? To help isolate its contribution, could you provide data on the performance degradation if a non-adaptive "always pushdown" policy were used with FAC, particularly for the worst-case queries highlighted in Figure 10b?

                3. On the "All-or-Nothing" Bin Sealing: Algorithm 1 appears to seal the first bin (containing the largest chunk) and set its size as the capacity for all other bins in that stripe. This seems to imply that no chunk larger than the smallest item in the first bin can be placed in any other bin in the stripe. Is this interpretation correct? If so, does this greedy choice ever lead to significantly sub-optimal packing for future stripes by "trapping" medium-sized chunks that could have been packed more efficiently?