Alluxio today announced strong results for the second quarter of its 2026 fiscal year. During the quarter, the company launched Alluxio Enterprise AI 3.7, a major release that delivers sub-millisecond TTFB (time to first byte) latency for AI workloads accessing data on cloud storage.
Alluxio also reported new customer wins across multiple industries and AI use cases, including model training, model deployment, and feature store query acceleration. In addition, the MLPerf Storage v2.0 benchmark results underscored Alluxio’s leadership in AI infrastructure performance, with the platform achieving exceptional GPU utilization and I/O acceleration across diverse training and checkpointing workloads.
Key Features of Alluxio Enterprise AI 3.7
- Ultra-Low Latency Caching for Cloud Storage – Alluxio AI 3.7 introduces a distributed, transparent caching layer that reduces latency to sub-millisecond levels while retrieving AI data from cloud storage. It achieves up to 45× lower latency than S3 Standard and 5× lower latency than S3 Express One Zone, plus up to 11.5 GiB/s (98.7 Gbps) throughput per worker node, with linear scalability as nodes are added.
- Enhanced Cache Preloading – The Alluxio Distributed Cache Preloader now supports parallel loading, delivering up to 5× faster cache preloading to ensure hot data availability for faster AI training and inference cold starts.
- Role-Based Access Control (RBAC) for S3 Access – New granular RBAC capabilities allow tight integration with identity providers (OIDC/OAuth 2.0, Apache Ranger), controlling user authentication, authorization, and permitted operations on cached data.
Customer Momentum in H1 2025
The first half of 2025 saw record market adoption of Alluxio AI, with customer growth exceeding 50% compared to the previous period. Organizations across tech, finance, e-commerce, and media sectors have increasingly deployed Alluxio’s AI acceleration platform to enhance training throughput, streamline feature store access, and speed inference workflows. With growing deployments across hybrid and multi-cloud environments, demand for Alluxio AI reflects rapidly rising expectations for high-performance, low-latency AI data infrastructure. Notable customers added in the half include:
- Salesforce
- Dyna Robotics
- Geely
Substantial I/O Performance Gains Confirmed in MLPerf Storage v2.0 Benchmark
Alluxio’s distributed caching architecture underscores its commitment to maximizing GPU efficiency and AI workload performance. In the MLPerf Storage v2.0 benchmarks:
- Training Throughput
- ResNet50: 24.14 GiB/s supporting 128 accelerators with 99.57% GPU utilization, scaling linearly from 1 to 8 clients and 2 to 8 workers.
- 3D-Unet: 23.16 GiB/s with 8 accelerators, 99.02% GPU utilization, similarly scaling linearly.
- CosmoFlow: 4.31 GiB/s with 8 accelerators, utilizing 74.97%, nearly doubling performance when scaling clients.
- LLM Checkpointing
- Llama3-8B: 4.29 GiB/s read and 4.54 GiB/s write (read/write times: 24.44s and 23.14s).
- Llama3-70B: 33.29 GiB/s read and 36.67 GiB/s write (read/write times: 27.39s and 24.86s).
Availability
Alluxio Enterprise AI version 3.7 is available here: https://www.alluxio.io/demo
Like this:
Like Loading...
Related
This entry was posted on August 27, 2025 at 12:19 pm and is filed under Commentary with tags Alluxio. You can follow any responses to this entry through the RSS 2.0 feed.
You can leave a response, or trackback from your own site.
Alluxio Closes Strong Q2 with Customer Growth, Sub-Millisecond Latency Capability for AI Data & Record MLPerf Storage v2.0 Benchmark Results
Alluxio today announced strong results for the second quarter of its 2026 fiscal year. During the quarter, the company launched Alluxio Enterprise AI 3.7, a major release that delivers sub-millisecond TTFB (time to first byte) latency for AI workloads accessing data on cloud storage.
Alluxio also reported new customer wins across multiple industries and AI use cases, including model training, model deployment, and feature store query acceleration. In addition, the MLPerf Storage v2.0 benchmark results underscored Alluxio’s leadership in AI infrastructure performance, with the platform achieving exceptional GPU utilization and I/O acceleration across diverse training and checkpointing workloads.
Key Features of Alluxio Enterprise AI 3.7
Customer Momentum in H1 2025
The first half of 2025 saw record market adoption of Alluxio AI, with customer growth exceeding 50% compared to the previous period. Organizations across tech, finance, e-commerce, and media sectors have increasingly deployed Alluxio’s AI acceleration platform to enhance training throughput, streamline feature store access, and speed inference workflows. With growing deployments across hybrid and multi-cloud environments, demand for Alluxio AI reflects rapidly rising expectations for high-performance, low-latency AI data infrastructure. Notable customers added in the half include:
Substantial I/O Performance Gains Confirmed in MLPerf Storage v2.0 Benchmark
Alluxio’s distributed caching architecture underscores its commitment to maximizing GPU efficiency and AI workload performance. In the MLPerf Storage v2.0 benchmarks:
Availability
Alluxio Enterprise AI version 3.7 is available here: https://www.alluxio.io/demo
Share this:
Like this:
Related
This entry was posted on August 27, 2025 at 12:19 pm and is filed under Commentary with tags Alluxio. You can follow any responses to this entry through the RSS 2.0 feed. You can leave a response, or trackback from your own site.