SNIA Developer Conference September 15-17, 2025 | Santa Clara, CA
Provide the audience with a clear description of the CDL behavior model. Provide the audience with the performance improvements demonstrated by the implementation of CDL in the field Provide the audience with understanding of the state of support for CDL in the SW and HW ecosystem.
A new set of NVM Express® (NVMe®) and Open Compute Project® (OCP) features is revolutionizing the virtualization landscape, enabling a new SSD-supported virtualization ecosystem. This presentation will explore these innovative features and their potential applications in host systems. We will describe an example virtual machine (VM) setup and discuss how the features can be utilized together to create a robust, secure, and performant virtualized environment. Specifically, we will cover the use of SR-IOV to expose individual functions with child controllers, Exported NVM Subsystems for building virtualized subsystems, and OCP’s security extensions to Caliptra for maintaining security. We will also delve into advanced topics, including Quality of Service (QoS) parameter setting across varied VMs with differing Service Level Agreement (SLA) processes, and live migration of VMs from one SSD to another. The presentation will highlight how Tracking Allocation Status and Granularity facilitates live migration. Finally, we will touch on extended SSD-supported virtualization examples, including single-port vs dual-port variations, AI enablement through direct GPU access, storage use cases, application container use cases, and extensible integration to Flexible Data Placement. These examples may utilize subsets of the features in novel configurations or explore entirely new applications.
Research on Shingled Magnetic Recording (SMR) disks began around 2008, with the first ever consumer-grade SMR (Drive Managed SMR) disk released in 2013, followed by HM-SMR (Host Managed SMR) disk in 2014 and HA-SMR (Host Aware SMR) disk in 2016. This talk reflects on 18 years of SMR evolution —covering physical layouts, filesystems, garbage collection algorithms, device drivers, and simulators. Furthermore, the talk will also discuss how SMR disks integrated with data storage solutions like RAID and deduplication, including real-world use cases of SMR disks by hyperscalers. Given this perspective, we will ask: are we really chasing a mirage with SMR disks when it comes to AI workloads? With the much-anticipated arrival of the first consumer-grade Heat Assisted Magnetic Recording (HAMR) disk—and more expected in the coming years—this talk will explore how lessons learned from the past 18 years of SMR technology can help pave a smoother transition to HAMR drives, particularly for AI workloads.
The state of Fibre Channel is exciting as we embark on the development of 256GFC (aka Gen 9 Fibre Channel)! With 128GFC products coming out at the end of this year, the Fibre Channel community is feverishly working on the next generation speeds. Now is the time to step back a bit and view the Fibre Channel roadmap. In this session, we discuss the technical challenges associated with the creation of 256GFC products as well as explore the features of Gen 8 Fibre Channel. Come explore the excitement around FC-SP-3 "Autonomous In-flight encryption" (AIE) that provide a plug-n-play solution for end-to-end security! Take a look at the expansion of Fabric Notifications to produce automated responses to disruptions in the network. Let's see where Fibre Channel is taking you.
Since EDSFF was created, the biggest complaint about EDSFF has been that there are too many form factors. While this was by design, EDSFF still has more factors than what was supported previously in Enterprise and Datacenter application. So introducing E2 as a new EDSFF form factor in the market is obviously going to get a lot of scrutiny. The goal of this presentation is to discuss the motivation behind creating the EDSFF E2 form factor, why existing form factors could not meet this need, and why the EDSFF E2 form factor ended up with the specific dimensions.
HPC and AI workloads require processing massive datasets and executing complex computations at exascale speeds to deliver time-critical insights. In distributed environments where storage systems coordinate and share results, communication overhead can become a critical bottleneck. This challenge underscores the need for storage solutions that deliver scalable, parallel access with microsecond latencies from compute clusters. Caching can help reduce communication costs when implemented on either servers or clients. Servers, in this context, refer to the data servers that provide file system and object store functionalities, while clients denote the storage clients running on compute nodes in HPC/AI clusters that access and retrieve data from these servers. However, server-side caching is limited by the fixed memory and network bandwidth of individual servers. Traditional client-side caching, on the other hand, is typically node-local which limits data reuse across the cluster and often results in redundant caching efforts, leading to inefficiencies and duplicated data. Furthermore, without a shared global view, synchronizing caches consistently across nodes becomes challenging, further diminishing their overall effectiveness. Global Distributed client-side caching over high-speed interconnects is attractive because it leverages the higher aggregate resources—such as DRAM, local SSDs, network bandwidth, and RDMA capabilities—available across the client nodes, scaling independently of the number of server nodes. However, fully realizing these benefits demands an efficient caching framework underpinned by carefully tuned policies to manage these valuable resources. In this presentation, we detail the design and implementation of an efficient, distributed client-side caching framework that addresses these challenges.
Command Duration Limits (CDL) is a QoS protocol for SCSI and ATA HDDs that provides the host with a model of traffic classes and command execution policies that enable a drive to optimize execution of consumed commands. The standard has a two dimensional model. CDL defines Duration Limits by which the host can set general priority for classes of commands and it defines Duration Limit Policies to further guide the device on the desired behavior of commands as they approach the limits of the Duration Limits. CDL has been adopted as SCSI and ATA standards, and implemented by drives, the Linux kernel, and the SAS hardware ecosystem. It is also being considered for adoption for NVMe HDDs. In this talk, we will present the CDL behavioral model, and an overview of the standard as implemented in SCSI and ATA, as well as some thoughts on how this model would be adapted to NVMe. We will also present real world results of CDL’s implementation in the field, showing examples of its benefits on command tail latencies, and also the performance trade-offs involved in maximizing IOPs per TB in today’s ever higher capacity HDDs.