We just announced our amazing Q2 results with customer growth, sub-millisecond latency capability and record MLPerf Storage 2.0 results. Here is a quick summary of the announcement.
This quarter marked a phenomenal period for Alluxio as we launched Enterprise AI 3.7, our most advanced release yet, delivering sub-millisecond TTFB latency for AI workloads accessing cloud storage. The new 3.7 version of Alluxio AI achieves up to 45× lower latency than S3 Standard and 5× lower latency than S3 Express One Zone, while delivering impressive throughput of up to 11.5 GiB/s per worker node with linear scalability. Read more about what’s new with Alluxio AI 3.7 here: https://www.alluxio.io/blog/alluxio-ai-3-7-now-with-sub-millisecond-latency.
Our customer momentum has been exceptional, with over 50% growth in the first half of 2025 compared to the previous period. We've welcomed notable new customers, including Salesforce, Dyna Robotics, and Geely, spanning industries from tech and finance to e-commerce and media. These organizations are leveraging Alluxio's AI acceleration platform to enhance training throughput, streamline feature store access, and speed up inference workflows across hybrid and multi-cloud environments. Learn more about the collaboration of Alluxio with Salesforce in this white paper: https://www.alluxio.io/whitepapers/meet-in-the-middle-for-a-1-000x-performance-boost-querying-parquet-files-on-petabyte-scale-data-lakes.

Alluxio's leadership in AI infrastructure was further validated in the quarter by our outstanding MLPerf Storage v2.0 benchmark results. Our distributed caching architecture achieved exceptional GPU utilization rates, with 99.57% for ResNet50 and 99.02% for 3D-Unet, while delivering substantial throughput gains across diverse training and checkpointing workloads. These results underscore our core mission: keeping GPUs fed with data at the speed they require for maximum ROI on infrastructure investments.
As we head into the second half of our 2026 fiscal year, we're more energized than ever about the future of AI infrastructure. With enterprise demand for high-performance AI workloads accelerating rapidly and our technology proving its value at scale, we're positioned to help organizations unlock unprecedented speed and efficiency in their AI initiatives. The best is yet to come!
.png)
Blog

Learn about the new features in Alluxio AI 3.8 designed to eliminate two of the most painful bottlenecks in modern AI pipelines. Introducing Alluxio S3 Write Cache, which dramatically reduces object store write latency and improves write-heavy workload performance, and Safetensors Model Loading Acceleration that delivers near-local NVMe throughput for model weight loading

For write-heavy AI and analytics workloads, cloud object storage can become the primary bottleneck. This post introduces how Alluxio S3 Write Cache decouples performance from backend limits, reducing write latency up to 8X - down to ~4–6 ms for concurrent and bursty PUT workloads.

Oracle Cloud Infrastructure has published a technical solution blog demonstrating how Alluxio on Oracle Cloud Infrastructure (OCI) delivers exceptional performance for AI and machine learning workloads, achieving sub-millisecond average latency, near-linear scalability, and over 90% GPU utilization across 350 accelerators.