Products
Blog

Make Multi-GPU Cloud AI a Reality
If you’re building large-scale AI, you’re already multi-cloud by choice (to avoid lock-in) or by necessity (to access scarce GPU capacity). Teams frequently chase capacity bursts, “we need 1,000 GPUs for eight weeks,” across whichever regions or providers can deliver. What slows you down isn’t GPUs, it’s data. Simply accessing the data needed to train, deploy, and serve AI models at the speed and scale required – wherever AI workloads and GPUs are deployed – is in fact not simple at all. In this article, learn how Alluxio brings Simplicity, Speed, and Scale to Multi-GPU Cloud deployments.

Alluxio's Strong Q2: Sub-Millisecond AI Latency, 50%+ Customer Growth, and Industry-Leading MLPerf Results
Alluxio's strong Q2 featured Enterprise AI 3.7 launch with sub-millisecond latency (45× faster than S3 Standard), 50%+ customer growth including Salesforce and Geely, and MLPerf Storage v2.0 results showing 99%+ GPU utilization, positioning the company as a leader in maximizing AI infrastructure ROI.
.png)
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
.jpeg)
How to Become a Contributor to Alluxio Open Source Project
This is a tutorial to guide a newbie to complete a new-contributor task and become an open-source contributor of the Alluxio project.
No items found.
.jpeg)
Self-serve Data Architecture with Presto and Alluxio Across Clouds
This article highlights synergy between the two widely adopted open-source projects, Alluxio and Presto, and demonstrates how together they deliver a self-serve data architecture across clouds.
Hybrid Multi-Cloud
.jpeg)
Alluxio and Apache Ranger Best Practices
As data stewards and security teams provide broader access to their organization’s data lake environments, having a centralized way to manage fine-grained access policies becomes increasingly important. Alluxio can use Apache Ranger’s centralized access policies in two ways: 1) directly controlling access to virtual paths in the Alluxio virtual file system or 2) enforcing existing access policies for the HDFS under stores.
No items found.

Using Consistent Hashing in Presto to Improve Caching Data Locality in Dynamic Clusters
Running Presto with Alluxio is gaining popularity in the community. It avoids long latency reading data from remote storage by utilizing SSD or memory to cache hot dataset close to Presto workers. Presto supports hash-based soft affinity scheduling to enforce that only one or two copies of the same data are cached in the entire cluster, which improves cache efficiency by allowing more hot data cached locally. The current hashing algorithm used, however, does not work well when cluster size changes. This article introduces a new hashing algorithm for soft affinity scheduling, consistent hashing, to address this problem.
Large Scale Analytics Acceleration

How to Set Up Monitoring System for Alluxio with Prometheus and Grafana in 10 Minutes
This blog will introduce how Tencent uses Prometheus and Grafana to set up monitoring system for Alluxio in 10 minutes.
No items found.
.jpeg)
Thousand-Node Alluxio Cluster Powers Game AI Platform A Production Case Study from Tencent
To provide model training with the best experience, Tencent has implemented a 1000-node Alluxio cluster and designed a scalable, robust, and performant architecture to speed up Ceph storage for game AI training. This blog will give you insight into how Alluxio has been implemented and optimized at Tencent.
Model Training Acceleration
.jpeg)
A Year with Alluxio Community 2021
2021 marked accelerated growth for the Alluxio Open Source Project. We could not be more grateful for what the community has achieved together in this past year. This blog provides a glimpse of the year long summary of our community growth.
Large Scale Analytics Acceleration
.jpeg)
Machine Learning Model Training with Alluxio: Part 3 - Benchmarking
This blog is the last one in the machine learning series. Our first blog introduced the what and why of our solution, and the second blog compared traditional and Alluxio solutions. This blog will demonstrate how to set up and benchmark the end-to-end performance of the training process.
Model Training Acceleration

Machine Learning Model Training with Alluxio: Part 2 - Comparable Analysis
This blog is the second in the machine learning series following the previous one, which discussed Alluxio's solution to improve training performance and simplify data management. With the help of Alluxio, loading data from cloud storage, training and caching data can be done in a transparent and distributed way as a part of the training process, thus improving training performance and simplifying data management. In this blog 2 of the series, we focus on comparing traditional solutions with Alluxio’s.
Model Training Acceleration
GPU Acceleration
.jpeg)
Machine Learning Model Training with Alluxio: Part 1 - Solution Overview
In this blog, we provide an overview of Alluxio's AI/ML model training solution. For more details about the reference architecture and benchmarking results, please refer to the full length whitepaper.
Model Training Acceleration
GPU Acceleration

Top Data Predictions for 2022
As more organizations advance their data revolution strategy, and run more diverse workloads on a wider variety of platforms across clouds and hybrid clouds, 2022 will see even more advances in AI, machine learning and analytic workloads and technologies and services to support them.
Data Migration
Hybrid Multi-Cloud
Your selections don't match any items.
