Overview
There is an arms race between vendors who are scrambling to position their storage systems as the ideal solution to handle ever-increasing performance and scale problems for AI workloads.
This is not just a matter of which storage vendor delivers the greatest performance to accelerate data throughput to GPU servers. A greater problem is that as AI models become more complex and inferencing sessions need to handle much longer context lengths, workloads frequently saturate GPU server memory and main system RAM. The result is an increasing churn and I/O wait times for data transfers to external storage. This growing network traffic to external storage causes serious problems that directly impact GPU computing workloads and increase costs in both on-premises and cloud-based data centers.
Here’s the rub: Even with vendors touting their latest storage performance improvements or advocating that their proprietary client software be installed on GPU servers to enable intelligent caching, token optimization, or other workarounds, the core problem persists—data pathways between external storage and GPU servers remain bottlenecked by network bandwidth, regardless of the storage type.
Adding more intelligence to better prioritize which subsets of data should be routed to the GPUs is important and helps. But even with the best of these methods an ever-increasing volume of data still must traverse the network bottleneck from external storage to feed the GPUs, and that is the crux of the problem.
The good news is a far better solution already exists to overcome these issues, which is hidden in plain sight within every GPU server– both on-premises and in the cloud—running on standard distributions of the Linux operating system, and that is the underutilized local NVMe storage built directly into the GPU servers.
This paper provides an overview of how organizations of any size can see tangible benefits and cost savings by activating this idle capacity as a seamless, protected, and ultra-fast Tier 0 shared storage environment to dramatically accelerate GPU workloads and improve GPU utilization.
When activated, Tier 0 is by far the most performant storage type currently available for AI workloads. And when this capacity is aggregated across multiple GPU servers in a cluster, organizations can get far greater output from their GPU investments at significantly lower costs, both on-premises and in the cloud.
Hammerspace Tier 0 is by far the most performant shared storage type currently available for AI workloads, with benchmarks demonstrating Tier 0 can support 16.5x more GPUs than a Lustre solution with an equivalent number of client connections. And when this capacity is aggregated across multiple GPU servers in a cluster, organizations can get far greater output from their GPU investments at significantly lower costs, both on-premises and in the cloud.
More importantly, this innovation works with existing on-premises and/or cloud-based GPU (or CPU/ARM) servers from any vendor, by activating internal local NVMe capacity that has already been paid for.
This bears repeating: This ultra-fast local NVMe capacity – which is largely underutilized – is a sunk cost, since it is already included within the price of GPU servers in both on-premises and cloud-based deployments. The fact that this also doesn’t add any additional power requirements is also a net benefit for data centers who are near the limit of their available power.
And with Hammerspace, activating this local NVMe as Tier 0 shared storage is done leveraging intelligence built into the Linux kernel, and which is included in all standard distributions. This means that servers do not require installation of proprietary client software, or any other alterations as they are delivered from manufacturers. As will be outlined below, Tier 0 is a game-changing solution that immediately provides value for environments of all sizes, from a small cluster with only a few GPU servers, all the way to hyperscale environments with many thousands of GPUs.
Moreover, utilizing this local Tier 0 capacity reduces or can even eliminate the costs of adding extreme-performance networking infrastructure and/or additional external high-performance storage systems, all of which are very expensive ways of trying to lessen the impact of the network bottleneck when trying to keep the GPUs fully utilized.
This ultra-fast local NVMe capacity – which is largely underutilized – is a sunk cost, since it is already included within the price of GPU servers in both on premises and cloud-based deployments.
And with Hammerspace, Tier 0 delivers these benefits with an open, standards-based platform while also providing global multi-protocol file/object access, non-disruptive data orchestration, data protection, and other enterprise-class data services across a global namespace that unifies storage silos from any vendor in one or more locations, including hybrid cloud environments, accelerating time to value for AI investments.
The result is Hammerspace enables customers to dramatically simplify the task of making their existing infrastructure AI-ready, reducing the costs and friction of launching AI or other high-performance data initiatives even within their existing infrastructure.
Activating Tier 0: Local NVMe Storage within GPU Servers
Lurking within every GPU server is an untapped resource that can dramatically reshape GPU-computing workloads. Virtually all GPU servers–whether on-premises or in the cloud–run standard distributions of Linux that already include NFS and pNFS v4.2 performance capabilities Hammerspace has contributed to the open source community over the last eight-plus years. In addition, these GPU servers include at least two, and usually eight high-performance local NVMe drives. Some vendors offer servers that support up to 16 local NVMe drives.
Different manufacturers ship varying amounts and capacities of NVMe drives in their servers. Some allow the drive slots to be only partially populated, while others—including cloud-based GPU servers—include these drives by default.

Fig. 1: Stranded NVMe capacity in GPU servers is mostly underutilized in GPU clusters. Data needed by the GPUs must go across the network to/from external storage, which slows down workloads, and decreases GPU utilization.
The problem is that while some applications can partially leverage this local capacity as a high-performance cache or temporary scratch space—particularly in a single server or with very small training datasets—in clusters of two or more GPU servers or with larger datasets, this NVMe storage often becomes stranded, siloed capacity that is rarely fully utilized. The reasons for this are as follows:
- Capacity is siloed: Local NVMe drives are isolated within each server. GPU servers cannot share this capacity across the cluster.
- Lack of shared access: Without some way to bridge these drives to create a shared storage environment across servers, any data on local storage must be manually copied, adding overhead, latency, and operational complexity.
- Limited data protection: Because NVMe storage on GPU servers lacks built-in redundancy across a cluster, there’s a high risk of data loss or disruption if a server crashes, or is disrupted in other ways.
Large SSD Drives and Larger GPU Server Clusters Change the Equation
As SSD drive densities increase and the number of GPU servers in a cluster grows, the landscape changes drastically. As of March 2025, GPU server vendors typically ship servers with local NVMe capacities ranging from 30TB to 122TB per server. These are based upon SSD densities of between ~4 and ~8TB per drive in eight and 16-drive server configurations.
But drive densities are increasing rapidly, with 30TB and 60TB drives commonly available in enterprise-class SSDs. And manufacturer roadmaps are forecasting the availability of 122TB drives by Q4 2025.
This means that even with the smaller drive capacities that are shipped in servers today, an eight-server cluster can contain nearly a Petabyte (PB) of ultra-fast NVME storage in default configurations. And for configurable GPU servers, upgrading the same eight-server cluster to 30TB drives would increase the available capacity to 3.8PB.
Comparing the price/ performance per GB per second delivered to the GPUs, external storage is ~10x more costly than Tier 0.
What is game changing about the Tier 0 innovation is that this extreme performance is available within the existing GPU servers and power budget, without the need to add external high-performance storage systems with their associated switch infrastructure, rackspace, power overhead, and other costs.
On a cost-per-TB basis, maxing out the internal NVMe storage on GPU servers costs about half as much as purchasing the same capacity using external arrays. But price/performance comparisons tell the real story: When comparing the performance cost of delivering data to GPUs, external Tier 1 NVMe storage has a loaded cost per GB per second roughly 10x greater than internal Tier 0 storage, assuming a 200Gb/s network. This calculation of loaded costs for Tier 1 storage factors in the cost of the drives & enclosures, power/cooling, rackspace, 200Gb/s networking, etc. Increase the network to a more expensive 800Gb/s switch and the price/performance advantages of Tier 0 are even greater1.
As will be detailed below, the implications of activating such massive volumes of high-performance capacity—combined with the growing numbers of GPU servers being deployed within AI clusters— are prompting organizations to re-think their IT architectures and their investments in external high performance storage and networking infrastructure.
The Problem is the Network Bottleneck
The key point here is that any external storage, however fast, is only as fast as the speed of the network connection between the servers and storage. This is not news, and has always been the case in all IT environments.
But the implications of this with GPU-heavy workloads for AI training and inferencing are far more serious than in traditional enterprise IT use cases. Because the other reality is if faster pipes are needed to keep the GPUs fully utilized, the more expensive the network infrastructure overhead becomes. And the bottleneck problem only gets worse as data volumes explode with larger GPU clusters, and with the added traffic caused by more complex models and larger context lengths that saturate server system memory.
Even using expensive 400Gb/s InfiniBand switches and the fastest external storage available, writing data across the network takes ~2.5 times longer than writing directly to local NVMe Tier 0 storage in a GPU server equipped with eight drives. This delay means valuable GPU cycles can be wasted waiting for data to arrive. And that’s true even when caching or other token optimization techniques are in use.
If the network is 100Gb/s, it takes nearly ten times longer to write the same amount of data to external Tier 1 storage as it does when writing to Tier 0 within the servers.
There are variables inherent in these numbers that can optimize the data flows, such as with different protection methods and affinitization, which intelligently places frequently accessed data as close as possible to the GPU that needs it. But the point remains: any data that needs to move across the network is going to be slower than utilizing the local NVMe already built into the servers, as shown in the tables below.

In cloud-based instances the difference is even more dramatic: In AWS, 500GB checkpoints in a cloud-based eight-GPU H100 instance will write to local NVMe in ~4.5 seconds. If it is written to high-performance EBS (io2) storage in AWS, the same job will take ~139 seconds (~2.3 minutes), or over 31 times longer. Writing checkpoints via the more affordable AWS EBS (gp3) storage will take ~9 minutes, or nearly 124 times longer.
Microsoft Azure offers a very fast Ultra disk, which has speeds equivalent to an 80Gb/s on-premises network, which is very fast for cloud offerings. But this is still 11 times slower than leveraging Tier 0 within an Azure GPU server cluster.

Checkpointing use cases are particularly telling, since they are needed to save the model’s current state to protect long-running jobs against possible node failure. The problem is that in many existing training frameworks the GPUs must sit idle while waiting for the checkpoint to be written to storage. In cloud-based GPU workloads this idle time is particularly painful, since you’re paying for the GPUs whether they are being used or not.
Newer training frameworks address this issue with asynchronous checkpointing to minimize GPU downtime. But the performance differential between Tier 0 and external storage is not only about checkpointing. The same performance differences seen in the examples above also affect all other data traffic between the GPUs and external storage. And the volume of data can be far greater than a 500GB checkpoint.
Let’s be clear here: purchasing the most expensive external storage and 800Gb/s networking options can approach the throughput of Tier 0 NVMe storage for on-premises environments, albeit at the cost of several thousand dollars per network port, not to mention the added power requirements and other overheads.
But with Tier 0 this performance can be achieved with NVMe capacity that organizations have already paid for within their GPU servers, and using the network infrastructure they already own. Even with existing 100Gb/s networks that list for about $200 per port, your workloads will still move faster by utilizing Tier 0.
In large environments that need dozens or even hundreds of GPUs, the cost penalty of adding that many more expensive network ports and external storage units explodes to extreme scales.
Simply put, adding the substantial expense of external networking and storage to an AI project—especially when you already own higher-performing local NVMe storage in your GPU servers—will directly reduce the ROI of your AI investment. And in cloud-based or hybrid on-premises/cloud deployments, the impact is even greater.
With Tier 0, and especially with larger GPU clusters and drive capacities, you can get the best of both worlds: The highest performing storage available on the market to keep your GPUs fed and your workloads moving, while using the NVMe capacity and network infrastructure you already own.
How to Activate Tier 0 Capacity within GPU Servers
When considering Tier 0, the obvious first question is this: If activating local NVMe storage within CPU- and GPU-based server clusters provides such a dramatic performance and cost benefit, then why isn’t everyone already taking advantage of it? Why is the storage industry focused primarily on the performance of external storage systems, or on creating proprietary client software that must be installed on the servers to optimize which data they feed to the GPUs?
The simple answer is that it requires the capabilities that the Hammerspace Data Platform uniquely provides to safely and effectively activate local NVMe capacity as ultra-fast Tier 0 shared storage across a cluster.

Fig. 2: Hammerspace activates the stranded capacity within the GPU servers, so that it is a seamless Tier 0 within a multi-tier, multi-vendor storage environment. Hammerspace data orchestration automates data placement and data protection, to maximize GPU efficiency.
Hammerspace has created a data platform that dramatically simplifies adapting existing infrastructure to become AI-ready, by eliminating storage silos and automating cross-platform data orchestration. Without requiring customers to install any client software, or alter their existing infrastructure, Hammerspace creates a high-performance parallel file system to provide unified file/object access across multiple storage types from any vendor, including local NVMe within servers, remote sites, and the cloud.
For the scale and performance needed for hyperscale AI workloads, Hammerspace uniquely leverages the parallelism of Linux standard pNFS v4.2 with Flex Files to automate data orchestration to the GPUs at extreme performance levels. As a standard that Hammerspace contributed to the Linux kernel years ago, the pNFSv4.2 client is already installed in every data center on the planet, and has been included in virtually all Linux distributions for over five years.
This has enabled Hammerspace to achieve the linear scalability needed by Meta, for example, to power its Llama 2, 3 and now Llama 4 large language models (LLMs) across many thousands of GPUs both in onpremises and cloud-based data centers.
And Hammerspace achieves this performance without needing to route data through proprietary controller nodes, or requiring customers to install proprietary client software on existing user computers and application servers, and without the need to alter existing servers and storage from any vendor.2
The Keys to Unlocking Tier 0
1. Data protection & orchestration
To activate and protect local NVMe storage across a cluster of GPU servers, Hammerspace aggregates this capacity into a seamless global namespace that can comprise multiple tiers of any storage type from any vendor, including across multiple sites and clouds.
But spanning all the storage types in a global file system is not enough. The key Hammerspace superpower that makes Tier 0 effective is the ability to automate non-disruptive file-granular data orchestration between Tier 0 and any other class of storage, both on-premises and remote.
Unlike the simple policies used by HSMs (hierarchical storage management) or other point solutions that are typically one-dimensional commands with limited options, Hammerspace enables administrators to create comprehensive service-level Objectives that can be finely-tuned based upon business logic, such as data durability, data placement, workflow automation, and other data services.
Multiple Objectives can be customized to accommodate a virtually unlimited number of business rules, use cases, and workflows, and may apply to all or intelligently-selected subsets of data across any storage type. The Objectives may be applied conditionally to determine with file/object granularity how data, directories, and/or buckets should be managed across any storage type from any vendor, including the cloud.
In the example above where a checkpoint file is written to local Tier 0, a service-level Objective would be set in Hammerspace to define how that checkpoint is to be protected, the level of data durability that is required, where it will be placed, etc.
For example, as soon as the file is written to Tier 0 Hammerspace would asynchronously copy or move that file downstream to an object store or any other target. Or in the case of clusters with larger Tier 0 capacities, the file could also be replicated to other nodes in the cluster for rapid recovery if needed. As many instantiations of the file as may be required for performance and/or durability can be determined with simple declarative Objectives.
With Hammerspace, these are not forked file copies that need to be wrangled later. Rather, they are instantiations of the same file, with shared file metadata that is globally accessible to users/applications and managed by Hammerspace across the entire unified namespace.
But spanning all storage types in a global file system is not enough. The key superpower is the ability to automate non-disruptive file-granular data orchestration between storage types and locations.
2. Data orchestration activates Tier 0 for all workloads
But Hammerspace data orchestration is not only used to protect checkpoint files. Leveraging the extreme performance Hammerspace delivers with pNFSv4.2 with Flex Files, Hammerspace Objectives can also automate any data movement into and out of Tier 0 to parallelize workloads across the GPU cluster.
Remember that the “east-west” inter-node networking speeds between the servers in a cluster is much faster than the “north-south” networking between the servers and external storage. This means that Hammerspace’s ability to stage data forward into Tier 0 close to the compute nodes, plus its ability to leverage pNFSv4.2 to parallelize I/O to accelerate it even more, together result in multiple ways that activating Tier 0 provides a net benefit to any high-performance workload.
Additionally, since networks are bi-directional, Hammerspace can simultaneously move checkpoint files out of a Tier 0 (GPU node sending) to external storage or to another Tier 0 node in the background without impacting the bandwidth needed to feed the GPUs reading training data (GPU node receiving) from the same Tier 0.
3. Where are the GPUs? Enabling burst workloads to cloud-based GPU clusters
A significant problem for organizations trying to launch AI initiatives is the availability of on-premises GPUs. In addition, even if GPUs are available to purchase, the planned usage or available data center power envelope may not make owning the GPUs the best choice. Which is why many organizations are using third-party providers offering GPUs-as-a-service, or are renting cloud-based GPUs. But both of these options are complicated, and require manually pushing files/objects to and from remote resources, and managing these workloads in yet another silo.
Hammerspace dramatically simplifies this, by extending its global file system and data orchestration between on-premises data centers and one or more clouds. It only takes a few minutes to spin up another Hammerspace cluster in any of the major public cloud providers, which can be scaled to whatever size is needed for the job.
Once a new Hammerspace instance is activated in the cloud, within minutes it automatically synchronizes file system metadata with the existing on-premises Hammerspace cluster. From then on, any cloud-based storage, including Tier 0 storage within cloud-based GPU clusters, becomes a seamless extension of the on-premises Hammerspace global namespace.
Hammerspace Objectives can now orchestrate data to and from the cloud automatically, including Tier 0 within the cloud-based GPUs just as is done between any other storage class in one or more on-premises data centers. This multi-site capability can bridge as many as 16 on-premises and cloud-based Hammerspace clusters to create a single unified, high-performance namespace that can span existing or new storage from any vendor in a truly global data environment.
Hammerspace eliminates these issues, extending its global file system and data orchestration between on premises data centers and one or more cloud resources.

Fig. 3: The Hammerspace Global Data Platform activates Tier 0 within GPU servers, with unified file/object access and data orchestration globally across on-prem and cloud-based resources.
Users and applications see the file system via standard SMB, NFS, and S3 mount points as before, without needing to install client software or to alter their workflows. Hammerspace handles data placement and other data services, plus automates workflows in the background across all silos, all without interruption to users or applications.
And to maximize cost savings, when burst workloads are finished the cloud-based Hammerspace cluster can be immediately decommissioned to reduce cloud compute fees. Orchestration of data will have already happened transparently as a background operation, ensuring the output data generated by the GPUs is automatically protected and placed wherever it is needed, in any storage type or location.
4. Shared standards-based file/object access
As mentioned above, Hammerspace requires no proprietary client software to be installed on users’ computers or application servers, and requires no alterations to existing storage. Users may not even be aware that Hammerspace is there. Both end-users and applications simply see shares in the global file system via standard NFS, SMB and/or S3 protocols exactly as they did before, except now they have unified access to the global namespace of all data, across all storage types, silos, and locations.
Only the data that users and applications have permissions to see is visible to them via shares across the entire global file system. Applications that expect to see S3 buckets, see S3 buckets including any or all data on any storage. And applications that expect to see a file hierarchy via NFS or SMB shares see that same data presented in a standard file/folder structure.
Which storage type the data happens to be on at the moment is totally governed by the Hammerspace servicelevel Objectives determined by your business rules, and are completely transparent to users and applications. In fact, Hammerspace incorporates a key feature called Live Data Mobility that it contributed into the Linux kernel as part of pNFSv4.2 with Flex Files so that even files that are open and actively being read or written can be orchestrated across storage tiers transparently as a background operation. There is no interruption to application or user workflows during the movement of live files, even to different storage types.
Hammerspace’s ability to provide universal access via standard protocols to all data across any storage, plus the ability to orchestrate data transparently at a file/object-granular level in the background anywhere are both critical in AI workloads. This is particularly the case for organizations with large quantities of unstructured data that they want to activate. Rather than needing to migrate whole volumes to a new repository, with Hammerspace only the specific files/objects that are needed for the job are orchestrated to the compute resources, again without creating forked copies or interrupting other users or applications.
Gone are the days of having to perform costly migrations and pushing data copies into net-new repositories, which themselves become new isolated silos. This dramatically simplifies adapting your existing infrastructure to become fully AI-ready.
Applications that expect to see S3 buckets, see S3 buckets including any or all data on any storage. Applications that expect to see a file hierarchy via NFS or SMB shares see that same data presented in a standard file/folder structure.
What about HCI and CPU-based servers?
Although the focus of this paper is on GPU-based workloads that are most urgently needed for AI use cases, the fact is that CPU-based servers often ship with the same ultra-fast local NVMe drives that can be activated as Tier 0. This is also true for low-power ARM CPUs.
These server types are often confused with Hyperconverged Infrastructure (HCI), which typically combine compute and storage in a single unit. HCI systems are predominantly used for structured data, VMs, and other block-oriented I/O patterns that are not suitable for CPU-intensive unstructured data workloads. And while HCI systems can use distributed storage architectures and other tiering solutions, their focus on block-oriented use cases make them unsuitable for processing unstructured data.
But structured data only makes up about 20% of enterprise datasets, and according to industry analysts, is growing at only about 20% per year. On the other hand, unstructured file/object data makes up about 80% of all enterprise data and is growing annually at an estimated 65%.
There are numerous CPU-intensive use cases that need large volumes of unstructured data, and that can take advantage of the same performance and cost improvements of Tier 0 as the GPU-computing workloads outlined above. Among them are other AI use cases, which include CPU-based AI inferencing, predictive analytics, or machine-learning pipelines dealing directly with unstructured datasets.
However, beyond AI use cases there are numerous other CPU-based application workflows that demand high performance and that can directly benefit from the ultra-low latency and enhanced throughput provided by activating Tier 0 on local NVMe drives within CPU servers.
Some of these use cases include:
- Real-time analytics and data streaming
- Rapid processing of log data, sensor inputs, or clickstream data to generate insights instantly.
- High-performance search and indexing
- Fast indexing and search across large-scale document repositories or datasets
(e.g., Elasticsearch, Splunk).
- Fast indexing and search across large-scale document repositories or datasets
- Video and media processing
- Low-latency editing, transcoding, rendering, and content delivery requiring quick, frequent data access.
- Financial trading and modeling
- Real-time risk analysis, algorithmic trading, and financial modeling needing immediate access to large datasets.
- Genomics and bioinformatics
- Rapid processing of genome sequencing data, imaging data, or bioinformatics workloads demanding near-instantaneous data retrieval.
The benefits for CPU-based Tier 0 include the same advantages highlighted in the GPU-based examples above:
- The lower cost of activating ultra-high performance local NVMe drives within the servers vs. going over expensive networks to external arrays,
- The significant boost in performance and lower latency that Tier 0 makes possible compared with external storage, which directly benefits these applications.
Additional Hammerspace performance advantages for CPU-based workloads
In addition to leveraging the extreme performance made possible with pNFSv4.2 with Flex Files and activating Tier 0 within CPU-based servers, Hammerspace takes advantage of its tight integration with standard Linux to provide another direct performance advantage for CPU-based workloads.
HCI systems and some other vendor solutions use a polling-based approach to process I/O. This is a user-space solution by which the CPU is constantly kept busy checking (polling) for incoming data. This polling occurs constantly, even when there is no data present, and consumes a significant amount of CPU resources. The result of this excessive load on the CPUs is increased latency, and degraded performance.
Hammerspace takes advantage of standard NFS, which is Linux kernel-based, to accelerate I/O with an IRQ approach (IRQ=interrupt requests) that significantly lowers CPU overhead and reduces latency.
Unlike the constant CPU load needed by polling-based approaches, the kernel-based IRQ approach used by Hammerspace means the CPU is alerted only when the data actually arrives, enabling far more efficient event-driven processing.
The result is faster, lower-latency I/O and more efficient CPU usage compared to user-space solutions that rely on CPU-intensive polling loops.
On the horizon: Coming soon to a Linux distribution near you
Another key innovation that Hammerspace has recently contributed upstream into Linux—and which will directly improve performance of local NVMe in Tier 0 even more—is an NFS Protocol Bypass in the Linux kernel.
Known in the Linux community as LOCALIO, this enhancement adds intelligence to the NFS clients and servers to detect when they are running on the same host. This eliminates unnecessary latency of going through the NFS protocol stack in the kernel to loop back to the local file system and storage. The feature was implemented in a robust way that works even if the NFS client and server are running in containers.
LOCALIO was initially released in Linux kernel 6.12 in November 2024, with additional enhancements in kernel 6.14 which was released in March, 2025.
In benchmark testing, preliminary results show that bypassing the NFS protocol has a huge impact on reducing CPU utilization within the server. In addition, testing has so far shown a marked boost in read performance within Tier 0 when using LOCALIO. Early testing also indicates we’ll see write performance improvements, although how much is still being quantified. This testing is on-going based upon the latest kernel updates, and will be documented when results are ready.
Combined with other upcoming enhancements to increase affinitization of data placement across the GPU cluster, the LOCALIO feature has already shown that it will provide significant performance benefits even over and above what Tier 0 delivers today.

Fig. 4: Hammerspace engineering developed LOCALIO and contributed it into the Linux kernel to further increase performance of local NVMe in GPU and CPU-based servers.
Although LOCALIO was included in Linux 6.12, which was released as the Long-Term Support (LTS) version in November 2024, it takes some time for enterprise Linux distributions to adopt new releases. But the demand for this feature is significant, and the most recent release of Red Hat Enterprise Linux (RHEL10) in May 2025 was based upon kernel 6.12, and also included the kernel 6.14 enhancements to LOCALIO.
It will take some time before organizations adopt to RHEL10 and other distributions as they upgrade to the latest Linux kernel, but when they do so the improvements Hammerspace has contributed will already be there waiting for them to take advantage of. This will further improve the dramatic performance and efficiency improvements available now with Tier 0.
Standards are the key
Hammerspace has a long history of contributing key performance enhancements into Linux, with over 2,400 features and patches pushed into the open source kernel in recent years. Trond Myklebust, the Hammerspace CTO, has been the Linux NFS client kernel maintainer for more than 20 years, and in that role has ensured that work Hammerspace does adheres to the standards maintained in the community.
Hammerspace Senior Principal Software Engineer Mike Schnitzer is also a Linux kernel maintainer, responsible for the upstream Linux kernel’s Device Mapper (DM) subsystem. And Tom Haynes, another Hammerspace software engineer, is a co-author of RFC 8435, which defined the pNFS Flexible File Layout in 2018, the key feature of pNFS that was adopted into the v4.2 spec in 2019.
This focus on open standards aligns with Hammerspace’s vision to overcome the limitations of proprietary vendor silos and vendor lock-in for enterprise data environments, while maintaining compatibility with existing infrastructure. On the user/application side, customers never have to install proprietary software on their application servers, or alter their user’s computers. And on the storage side, Hammerspace is compatible with any storage, from any vendor.
As importantly, the extreme performance improvements Hammerspace has engineered into standard Linux—and which are part of all major distributions—are also tightly integrated within Hammerspace software, which is uniquely positioned to take advantage of them.
Summary
Organizations in all industries and of all sizes both public and private are grappling with how to retool their IT infrastructures to accommodate the explosive promise of AI. Gone are the days where enterprise IT was a mostly linear proposition, with a one-to-one ratio between applications and data, where unstructured data would gradually age out until it lay dormant in an archive.
While organizations generally realize they need an AI strategy, the return on investment in AI initiatives is far less certain, particularly for those who are not among the hyperscale elite. But a quick scan of the industry trade press often would lead you to think that the only way to launch an AI project is by creating a second AI-focused IT architecture alongside the infrastructure they already own.
Such solutions promote purchasing net-new high-performance storage repositories and expensive highspeed networking, or modifying customer’s application servers with proprietary client software, plus multiple other expensive and vendor-locked solutions.
All of these offerings perpetuate the problems of the past, where proprietary data silos fragment access and hinder efficient utilization of data. Such an approach makes this process unnecessarily complex, leads to the problems of data gravity and inefficient use of resources, and most importantly—substantially higher costs in both CAPEX and OPEX.
Hammerspace has already proven that its standards-based approach dramatically simplifies the task of enabling organizations to make their existing IT infrastructure AI-ready. For organizations embarking on or expanding their AI journey, this means AI-focused resources can now be additive and tightly integrated with existing IT infrastructure, rather than needing to be an isolated net-new silo.
The Hammerspace Data Platform has demonstrated this vision in real-world environments that span onpremises and multi-cloud by enabling unified file/object access, and high-performance data orchestration in organizations of all sizes, including scaling to extreme volumes and performance levels across multiple on-premises and hybrid cloud sites, in all verticals including HPC and hyperscale AI environments.
For organizations this means AI-focused resources can now be additive and tightly integrated with existing IT infrastructure, rather than an isolated net-new silo.
These capabilities and this real-world experience are what make Tier 0 a solution that is having a gamechanging impact on high-performance workloads for both GPU- and CPU-based computing.
Hammerspace enables organizations of all sizes in any industry to take advantage of the unprecedented performance boost Tier 0 achieves, surpassing the fastest external storage types available on the market. And this is matched only by the incredible cost savings possible by activating this underutilized extremeperformance storage tier.
The fact that this activates a performance tier that you may already own is icing on the cake.
This is the Hammerspace innovation.