On-Demand Videos

Real-time OLAP databases are optimized for speed and often rely on tightly coupled storage-compute architectures using disks or SSDs. Decoupled architectures, which use cloud object storage, introduce an unavoidable tradeoff: cost efficiency at the expense of performance. This makes them unsuitable for databases that need to provide low-latency, real-time analytics, especially the new wave of LLM-powered dashboards, retrieval-augmented generation (RAG), and vector-embedding searches that thrive only when fresh data is milliseconds away. Can we achieve both cost efficiency and performance?
In this talk, we’ll explore the engineering challenges of extending Apache Pinot—a real-time OLAP system—onto cloud object storage while still maintaining sub-second P99 latencies.
We’ll dive into how we built an abstraction in Apache Pinot to make it agnostic to the location of data. We’ll explain how we can query data directly from the cloud (without needing to download the entire dataset, as with lazy-loading) while achieving sub-second latencies. We’ll cover the data fetch and optimization strategies we implemented, such as pipelining fetch and compute, prefetching, selective block fetches, index pinning, and more. We'll also share our latest work about integration with open table formats like iceberg, and how we will continue to achieve fast analytics directly on parquet files by implementing all the same techniques that apply to tiered storage.

The data lake is a fantastic, low-cost place to put data at rest for offline analytics, but we've built it under the terms of a terrible bargain: all that cheap storage at scale was a great thing, but we gave up schema management and transactions along the way. Apache Iceberg has emerged as king of the Open Table Formats to fix this very problem.
Built on the foundation of Parquet files, Iceberg adds a simple yet flexible metadata layer and integration with standard data catalogs to provide robust schema support and ACID transactions to the once ungoverned data lake. In this talk, we'll build Iceberg up from the basics, see how the read and write path work, and explore how it supports streaming data sources like Apache Kafka™. Then we'll see how Confluent's Tableflow brings Kafka together with open table formats like Iceberg and Delta Lake to make operational data in Kafka topics instantly visible to the data lake without the usual ETL—unifying the operational/analytical divide that has been with us for decades.

Storing data as Parquet files on S3 is increasingly used not just as a data lake but also as a lightweight feature store for ML training/inference or a document store for RAG. However, querying petabyte- to exabyte-scale data lakes directly from cloud object storage remains notoriously slow (e.g., latencies ranging from hundreds of milliseconds to several seconds on AWS S3).
In this talk, we show how architecture co-design, system-level optimizations, and workload-aware engineering can deliver over 1000× performance improvements for these workloads—without changing file formats, rewriting data paths, or provisioning expensive hardware.
We introduce a high-performance, low-latency S3 proxy layer powered by Alluxio, deployed atop hyperscale data lakes. This proxy delivers sub-millisecond Time-to-First-Byte (TTFB)—on par with Amazon S3 Express—while preserving compatibility with standard S3 APIs. In real-world benchmarks, a 50-node Alluxio cluster sustains over 1 million S3 queries per second, offering 50× the throughput of S3 Express for a single account, with no compromise in latency.
Beyond accelerating access to Parquet files byte-to-byte, we also offload partial Parquet processing from query engines via a pluggable interface into Alluxio. This eliminates the need for costly index scans and file parsing, enabling point queries with 0.3 microseconds latency and up to 3,000 QPS per instance (measured using a single-thread)—a 100× improvement over traditional query paths.
.png)
ALLUXIO DAY XII 2022
April 28, 2022
Shawn Sun from Alluxio will present the journey of using Alluxio as the storage system for Kubernetes through Container Storage Interface (CSI) plugin and Alluxio CSI driver. This talk will cover the challenges we are facing with traditional setup in the AI/ML training jobs, and how Alluxio CSI driver manages to address them. It will also talk about a recent change to the driver that made it more sturdy and robust.
ALLUXIO DAY XII 2022
April 28, 2022
This talk will discuss the process and technical details behind a responsible vulnerability disclosure of an issue detected in Alluxio recently. I will share some of the lessons I’ve learned as a security researcher dealing with multiple open-source vendors and my thoughts about the actions organizations and projects should take to ensure successful vulnerability management and disclosure programs. Learn more about creating more secure software.
This presentation will include information about how Alluxio and NetApp StorageGRID helps enterprises accelerate the adoption of cloud and optimize their resource spend on a modern hybrid big data architecture. The conversation will cover use case and architecture info from a variety of enterprises and some of the high level technical details of how these business solutions are constructed.
ALLUXIO DAY X 2022
March 3, 2022
Chen Liang from Uber and Beinan Wang from Alluxio will present the practical problems and interesting findings during the launch of Alluxio Local Cache. Their talk covers how Uber’s Presto team implements the cache invalidation and dashboard for Alluxio’s Local Cache. Chen Liang will also share his experience using a customized cache filter to resolve the performance degradation due to a large working set.
ALLUXIO DAY X 2022
March 3, 2022
Within Alluxio, the master processes keep track of global metadata for the file system. This includes file system metadata, block cache metadata, and worker metadata. When a client interacts with the filesystem it must first query or update the metadata on the master processes. Given their central role in the system, master processes can be backed by a highly available, fault tolerant replicated journal. This talk will introduce and compare the two available implementations of this journal in Alluxio, the first using Zookeeper and the more recent version using Raft.
ALLUXIO DAY X 2022
March 3, 2022
In this talk, Lei Li and Zifan Ni share the experience of applying Alluxio in their AI platform to increase training efficiency at bilibili. The talk also includes technical architecture and specific issues addressed.
Data platform teams are increasingly challenged with accessing multiple data stores that are separated from compute engines, such as Spark, Presto, TensorFlow or PyTorch. Whether your data is distributed across multiple datacenters and/or clouds, a successful heterogeneous data platform requires efficient data access. Alluxio enables you to embrace the separation of storage from compute and use Alluxio data orchestration to simplify adoption of the data lake and data mesh paradigms for analytics and AI/ML workloads.
Join Alluxio’s Sr. Product Mgr., Adit Madan, to learn:
- Key challenges with architecting a successful heterogeneous data platform
- How data orchestration can overcome data access challenges in a distributed, heterogeneous environment
- How to identify ways to use Alluxio to meet the needs of your own data environment and workload requirements
ALLUXIO DAY IX 2022
January 21, 2022
Video: Presentation Slides: Industrial Bank's Alluxio Deployment from Alluxio, Inc.
ALLUXIO DAY IX 2022
January 21, 2022
ALLUXIO DAY IX 2022
January 21, 2022
ALLUXIO DAY VIII 2021
December 14, 2021
Feifei Cai & Hao Zhu from WeRide provide an overview of Alluxio + Spark use case, which has been deployed and running in production to accelerate auto data tagging in the autonomous driving development.
ALLUXIO DAY VIII 2021
December 14, 2021
This talk will introduce Apache Iceberg and its place in a modern and open data platform. It will cover the motivation for creating Iceberg at Netflix, as well as the data architecture that Iceberg makes possible.