Cloud service providers are pushing the boundaries of storage capacities to meet the performance requirements of big data workloads and expanding data volumes. Intel and AliCloud collaborated on Cloud Storage Acceleration Layer (CSAL) to address NAND density and scale challenges. CSAL uses Intel Optane SSDs to improve system performance while reducing costs scaling QLC NAND value. In this presentation, we will show and analyze the real world I/O behaviors from big data services of Alibaba Cloud to demonstrate predictable performance using CSAL.
Computational storage is well known for its data movement reduction, leading to higher application performances. The lower power consumption is another benefit which is often highlighted, but a full system life cycle analysis is recommended to get the full picture of the energy used and its related environmental impact.
In this talk, a detailed carbon footprint analysis will be described for a standard system and a computational storage based system. That will help in understanding the computational storage benefits in term of carbon footprint, and will provide some opportunities to pave the way for a roadmap of next generation eco-designed computational storage devices.
LEARNING OBJECTIVES
Data protection is one of the key aspects of a stateful application. In case of containerized deployment scenarios, this is even more crucial considering the ephemeral nature of containers.
This session throws light on key aspects of container data protection, different kinds of challenges involved here. It will also cover the activities going on in this area, solutions being worked out to bridge the gaps
Data insight is a pivotal area for any technology today. Comprehensive storage monitoring can provide critical health information for preventive and predictive maintenance of various solutions. As the application deployments see an exponential growth towards container environments, container storage monitoring is inevitable. This session will provide the current state of container storage observability, the key challenges in storage monitoring in container deployments (esp Kubernetes), and also the trends. The session will also discuss potential solutions and suggestions based on open source projects for the container observability framework for heterogeneous storage vendors. The session will provide the technical architecture for a Kubernetes native observability solution and discuss the possibilities.
LEARNING OBJECTIVES
Since its inception, the Storage Performance Development Kit (SPDK) project has demonstrated that it's possible to build efficient storage systems capable of tens of millions of IOPS. This talk will describe the techniques and principles used to achieve over 80M IOPS at an amazing average latency of just 25 microseconds. We will also discuss the workloads, test methodologies, and automation tools that allow for continuously measuring and improving performance of NVMe, NVMe-oF and Vhost implementations.
LEARNING OBJECTIVES
Ansible has made automation frameworks to seamlessly function with multi-cloud and hybrid cloud environments. With auto-scaling and configuration management, Ansible-based automation platforms help CI/CD developers with network compliance, SaaS operations, and end-to-end legacy systems modernization. The challenge is to manage the Docker and Containers for Hybrid and Multi-Cloud environments, which are increasingly gaining preference because of their flexibility, performance, and economy. How can one ensure an automation framework for application deployment and configuration management for DevOps teams working with remote cloud instances and niche digital architectures? This paper aims to answer these questions by demonstrating Ansible’s contribution to the different phases of the CI/CD pipeline. The paper will also go to lengths to describe Ansible Automation Mesh, where its playbooks deal with configuration management automation to ensure rapid development and zero downtime for Hybrid and Multi-cloud environments.
LEARNING OBJECTIVES
In the modern era of multi-core systems, utilizing cores is a tedious job. Synchronization and communication among processors involve high cost. Software transaction memory systems (STMs) addresses these issues and provide better concurrency in which programmer need not have to worry about consistency issues. Another advantage of STMs is that they facilitate compositionality of concurrent programs with great ease. Different concurrent operations that need to be composed to form a single atomic unit is achieved by encapsulating them in a single transaction.
We introduce an innovative STM system to exploit multi-cores as multi-version object based STM (MVOSTM) which works on higher level operations such as insert, delete instead of read, write operations; ensures compositionally by making the transactions atomic; harness the greater concurrency while maintaining multiple versions; and reduces the number of aborts further. We have used MVOSTM for the list and hash-table data structure as list-MVOSTM and HT-MVOSTM. Experimental evaluation shows that list-MVOSTM and HT-MVOSTM outperform state-of-the-art STMs and incur negligible aborts.
LEARNING OBJECTIVES
During the last 10 years, a new communication protocol to interface with SSDs has been specified from the ground-up, allowing to fully exploit the levels of parallelism and performances of all-flash storage: NVMe, and NVMe-oF. NVMe-oF promises to enable the performances of direct attached all-flash storage with the flexibility and TCO savings of shared storage. Implementing NVMe-oF support in enterprize storage arrays allows to combine NVMe protocol performance with the reach feature set of modern storage arrays and enables modern storage arrays to meet growing customer demands.
LEARNING OBJECTIVES
Data science, machine learning (ML), and artificial intelligence have exploded in popularity in the last few years, with companies building out dedicated ML teams. Kubeflow is an OpenSource, ML toolkit for Kubernetes which provides useful components resolving problems in multiple areas. In this hands on session we will discuss about how you can make a scalable architecture for ML training and inference at scale using a distributed storage in a backend and Kubeflow.
LEARNING OBJECTIVES
Crimson is a project to refactor ceph osd based on the shared-nothing seastar framework. The current Crimson SeaStore is mainly targeting NAND SSD because of their high performance-price ratio and is heavily optimized for random read and sequence write patterns. This presentation will extend the SeaStore framework to support multiple tiers of devices, mainly byte addressable random access devices such as optane Persistent Memory (PMEM), and random block access devices such as optane SSD.
PMEM has the following usage models: 1) As transient cache which has nearly the same speed and latency of DRAM but has a much larger size; 2) As persistent storage for metadata which is updated synchronously like memory; 3) As a hot device tier for both metadata and data. In the extended SeaStore framework, each kind of device has its own device manager to arrange the usage of the device: 1) NAND SSD uses segment manager which partition device into segments and uses a cleaner to reclaim space; 2) Optane SSD uses random block device manger which uses bitmap to allocate space; 3) and PMEM uses its own manager which also uses bitmap to allocate space. Some common paths with different devices are shared, such as transaction commit and delta processing, with adjustments needed to adapt to the specialized I/O patterns.
LEARNING OBJECTIVES
This paper discusses similarities and differences between Persistent Memory and NAND Flash storage technologies.
Media specific characteristics like byte address ability, access modes, and cost determine where Persistent Memory is better fit than NAND Flash Storage. With adoption of serial and fabric attached protocols like CXL, Gen-Z and OpenCAPI, Persistent Memory can be packaged in same form-factor as NVMe drives, and it can reuse existing NVMe management infrastructure. This helps OEM to develop common chassis architecture which can house both Memory and NAND Flash SSDs in same slots.
This paper highlights few of the ideas where Persistent Memory and NAND Flash drives become physically replaceable.
LEARNING OBJECTIVES
Some things can’t scale in the cloud. When you are trying to get all the performance out of your systems for SaaS and IaaS instances and RAID 0 seems like the option. What do you do when you have 60 servers go down due to an SSD failure? In this session, you will learn about new breakthrough data protection technology for SSDs that gives you better performance during a drive rebuild than you can get from RAID 0 while getting RAID 5 like data protection.
LEARNING OBJECTIVES
On the domain controller side we got a lot of updates recently:
- Updated Heimdal
- Working with the latest MIT Kerberos
On the member server side we fixed some critical bugs and have plans for future improvements how a file server can avoid as much domain controller interaction as possible.
LEARNING OBJECTIVES
Compute Express Link™ (CXL™) is an open industry-standard interconnect offering coherency and memory semantics using high-bandwidth, low-latency connectivity between the host processor and devices such as accelerators, memory buffers, and smart I/O devices. CXL technology is designed to address the growing needs of high-performance computational workloads by supporting heterogeneous processing and memory systems for applications in Artificial Intelligence, Machine Learning, communication systems, and High-Performance Computing. These applications deploy a diverse mix of scalar, vector, matrix, and spatial architectures through CPU, GPU, FPGA, smart NICs, and other accelerators. The CXL 2.0 specification introduces support for switching, memory pooling, and persistent memory – all while preserving industry investments by supporting full backward compatibility. Based on member feedback, the CXL Consortium technical working groups published Engineering Change Notices (ECN) for the CXL 2.0 specification, primarily to enhance performance, reliability, software interface, and testability while offering design simplification. The presentation will also provide a high-level sneak peek of the latest advancements in the CXL 3.0 specification development, its new use cases and industry differentiators.
LEARNING OBJECTIVES
Data persistence on CXL is an essential enabler toward the goal of instant-on processing. DRAM class performance combined with non-volatility on CXL enables a new class of computing architectures that can exploit these features and solve real-world bottlenecks for system performance, data reliability, and recovery from power failures. New authentication methods also enhance the security of server data in a world of cyberattacks.
LEARNING OBJECTIVES
Enabling, sustaining and ensuring the highest possible performance along with continued application availability is a major goal for all mission critical applications to meet the demanding business SLA’s., all the way from on-premises to VMware Hybrid Clouds.
Please join this session to learn how performance of business-critical Oracle workloads can be accelerated via Persistent Memory technology using Intel Optane DC PMM in App Direct mode backed Oracle 21c Persistent Memory Filestore on VMware vSphere platform, thereby achieving those stringent business goals and enhancing performance.
LEARNING OBJECTIVES
Modern persistent-memory (PM) file systems perform well in benchmark settings, when the file system is freshly created and empty. But after being aged by usage, as will be the normal mode in practice, their memory-mapped performance degrades significantly. This paper shows that the cause is their inability to use 2MB hugepages to map files when aged, having to use 4KB pages instead and suffering many extra page faults and TLB misses as a result.
We introduce WineFS, a novel hugepage-aware PM file system that largely eliminates this effect. WineFS combines a new alignment-aware allocator with fragmentation-avoiding approaches to consistency and concurrency to preserve the ability to use hugepages. Experiments show that WineFS resists the effects of aging and outperforms state-of-the-art PM file systems in both aged and un-aged settings. For example, in an aged setup, the LMDB memory-mapped database obtains 2× higher write throughput on WineFS compared to NOVA, and 70% higher throughput compared to ext4-DAX. When reading a memory-mapped persistent radix tree, WineFS results in 56% lower median latency than NOVA.
LEARNING OBJECTIVES
Introduced in 2021, Fabric Notifications developed into a key resiliency solution for storage networks. Providing the ability to respond to events impacting the Fabric, Fabric Notifications enhances the overall user experience when deploying traditional Fibre Channel SANs. In this session, the improvements in the user experience are profiled for solutions using Linux, AIX, and PowerPath. The latest developments in both standards and market solutions are also provided as an update to this exciting new technology.
LEARNING OBJECTIVES
The Solid State Drive (SSD) capacity and performance is increasing over the years and due to this thermal management in SSD is becoming very challenging. As SSD works harder to meet growing demands, it consumes more power and consequently generates more heat. Apart from designing the SSDs to tolerate higher temperature, another effective solution to address the overheating issue is to have a good thermal management. The thermal throttling mechanism is thus fundamental for SSDs. A thermal sensor is used in the drive to monitor the temperature and temperature details can be extracted via S.M.A.R.T. command.
The idea of this paper is to Study thermal impact of different types of IOs, chunk size in multi core SSD with different density, Die Size and NAND Type (SLC & TLC) as parameter. This study will help us in understanding & improve the thermal throttling policy.
LEARNING OBJECTIVES
The traditional methods of managing enterprise disk drives leverage multiple machine learning models which include its inherent bias or uncertainties. However, recent development and growing interest of academia and industry in Conformal Prediction has further led to newer perspectives on solving the existing problems with the machine learning models. Conformal prediction (CP) is a set of algorithms devised to assess the uncertainty of predictions produced by a machine learning model.
In this talk, we will first introduce Conformal Prediction, followed by a comparative analysis of solving the problems (using the traditional machine learning method and conformal prediction framework). We will discuss three issues of disk drive management using conformal classification – 1) disk drive failure estimation, 2) inventory planning of individual storage drives, and 3) mechanism for ranking based dynamic hot-sparing.
To conclude, we will discuss in the application section, how the framework can be used for ransomware detection and designing an efficient disk drive scrubbing mechanism.
LEARNING OBJECTIVES
This talk will in detail describe the symbolic link race condition that led to the Samba CVE-2021-20316. Symlink races lead to CVEs all over the place, even the secure programming language Rust was recently bitten by it with CVE-2022-21658. Samba solved the problem with a major development effort over many months, however the solution comes with a performance penalty. In the best tradition of "Make it right first, make it fast", work is ongoing to refactor Samba's internals to not only get back the speed of the insecure path handling, but potentially go beyond that and moreover add proper SMB2 symlink error handling. The talk will describe Samba's current solution, the reasons why we did not solve it "the right way" from the beginning, and our way out of this.
LEARNING OBJECTIVES
This talk is going to give an overview of recent changes in the Samba fileserver and an outlook on the development roadmap. Recent development has been mainly focusing on security resulting in the release of Samba 4.15 last year.
Looking forward there are many things the Samba fileserver development team has on its todo list and this presentation will give a first hand insight into the making of the next Samba versions.
LEARNING OBJECTIVES
As new SSD architectures are achieving 1M IOPS for Random Operations, a “one in a million” event happens every second. This puts the term Quality of Service (QoS) as significant data point for SSD performance benchmarks and how they impact application performance.
LEARNING OBJECTIVES
Infrastructure offload based around NVMe-oF can deliver performance of direct attached storage with the benefits and composability of shared storage. Storage Performance Development Kit (SPDK) is a set of drivers and libraries uniquely positioned to demonstrate how projects like Infrastructure Programmer Development Kit (IPDK) can provide vendor agnostic high performance and scalable framework. This session will discuss how the SPDK NVMe-oF target has evolved to enable infrastructure offload.
LEARNING OBJECTIVES
Accessing files securely and efficiently matters, and over the past year many improvements have been made in Linux for accessing files remotely via SMB3.1.1. This has been a great year for cifs.ko with the addition of new SMB3.1.1 features and optimizations. It continues to be the most active network/cluster file system on Linux. And now with the addition of a kernel server to Linux (ksmbd), there are multiple Linux server options (Samba and ksmbd).
Improvements to performance have been made by adding support for handle leases (deferred close), better optimizing multichannel, and by changes to read ahead caching, and directory and metadata caching and also signing improvements have been made. Offline caching (fscache) has been rewritten and improved, and support for the Witness protocol (server notification about key events like server moving), and security has improved with support for the strongest encryption, and more recently the exciting work on QUIC. This presentation will go through the features added over the past year to the Linux client (and kernel server) and demonstrate how they help common scenarios, from accessing the cloud (like Azure) to accessing Samba, Windows, Macs and the new Linux kernel server (ksmbd).
LEARNING OBJECTIVES
The SNIA Swordfish™ specification for the management of storage systems and data services is an extension of the DMTF Redfish® specification. Together, these specifications provide a unified approach for the management of servers and storage in converged, hyper-converged, hyperscale, and cloud infrastructure environments.
If you’ve heard about this latest wave in storage management, but are looking for the latest updates to the functionality of Swordfish, this presentation is for you. The speaker will catch up on the standard’s most recent developments, which include:
LEARNING OBJECTIVES
Over the past decade IT organizations have discovered the benefits of moving applications and storage to the cloud. Recent trends point to an increase of on premises hybrid cloud solutions. Cloud-native applications running in containers has driven the need for high-performance persistent storage that is both easy to deploy and manage.
In this 20-minute presentation, Fibre Channel expert Mark Jones will explore the role that Fibre Channel plays in merging the benefits of a cloud-based infrastructure with the advantages of an on-premises hosting solution for gains in reliability, performance, and scalability – all with cost reduction. With a focus on Kubernetes, Mark will explain how these best practice environments can be combined for the next generation of application deployments.
LEARNING OBJECTIVES
Digital information storage in DNA is not a new idea, and the recent developments in the synthesis of DNA have significantly reduced the price per base pair. The current price per base-pair remains $0.07, far from commercial implementation compared to HDD or SSD based storage cost. Earlier work focuses on designing a data encoding method for DNA-Based storage, error correction code, genomic compression, enzymatic method of synthesizing nucleotides in the lab, and few works on encrypting the data. Unfortunately, little work has been done to reduce the amount of nucleotide required for digital information representation, detect malicious tampering of digital data, and choose the deduplication method as compression doesn’t help much for genomic data.
This talk will address the challenges encountered and present proposed solutions for reducing the storage cost for DNA-Based storage per unit of digital information, which uses similarity-based variable-length deduplication with delta encoding).
We also discuss storing archival storage metadata in Blockchain, making data immutable and tamper-proof (smart contract for similarity-based deduplication) and eliminating a single point of failure (decentralized meta-data storage in Blockchain). To conclude, we will present an evaluation of considering the Polysaccharide (Sugar) as one of the archival storage media and comparative analysis with DNA-based storage.
LEARNING OBJECTIVES
Demand for higher bandwidth to access data in a secure and reliable way keeps growing. This has resulted in a transition toward the most recent implementation of Fibre Channel technology operating at 64G speed. Despite being fully backward compatible, the new solution introduces key changes from previous generations. In this session we discuss the new modulation schema, the enhanced FEC algorithm, the modified link bring up sequence and other technical considerations.
LEARNING OBJECTIVES
This proposed talk will present a design solution that improves the speed performance and cost efficiency of B+ tree by leveraging computational storage drives (CSDs) with built-in transparent compression. As the most widely used indexing data structure, B+ tree powers almost all the relational database management systems today. Nevertheless, it is well known that B+ tree has two major shortcomings: low storage efficiency and high write amplification. The arrival of CSDs with built-in transparent compression straightforwardly mitigates the storage efficiency issue for B+ tree. This proposed talk will show that CSDs can also be leveraged to significantly reduce the B+ tree write amplification and hence improve its speed performance. The key is to exploit the fact that CSDs with built-in transparent compression allows data management software employ sparse data structure without sacrificing the true physical storage cost. Experimental results show that the developed solution can reduce the B+ tree write amplification by over 10x.
LEARNING OBJECTIVES
For the past three decades, PCI-SIG® has delivered a succession of industry-leading PCI Express® (PCIe®) specifications that remain ahead of the increasing demand for a high-bandwidth, low-latency interconnect for compute-intensive systems in diverse market segments, including data centers, artificial intelligence and machine learning (AI/ML), high-performance computing (HPC) and storage applications.
In early 2022, PCI-SIG released the PCIe 6.0 specification to members, doubling the data rate of the PCIe 5.0 specification to 64 GT/s (up to 256 GB/s for a x16 configuration). To achieve high data transfer rates with low latency, PCIe 6.0 technology adds innovative new features like Pulse Amplitude Modulation with 4 levels (PAM4) signaling, low-latency Forward Error Correction (FEC) and Flit-based encoding. PCIe 6.0 technology is an optimal solution to meet the demands of Artificial Intelligence and Machine Learning applications, which often require high data bandwidth, low latency transport channels.
This presentation will explore the benefits of PCIe 6.0 architecture for AI/ML workloads and its impact on next-generation cloud data centers. Attendees will also learn about the potential AI/ML use cases for PCIe 6.0 technology.
LEARNING OBJECTIVES
Just as with mining in the real world, the people who mine cryptocurrency use powerful equipment to increase their chance of finding valuable resources. But all that power comes at a cost: namely, massive energy consumption. However, next generation blockchains such as the Chia Network could change all that. Unlike other cryptocurrencies, you don't mine Chia with a processor, graphics card, or an ASIC miner. Instead, you farm Chia with storage space, which is where hard drives and solid-state drives come in, using the Serial Attached SCSI (SAS) infrastructure to connect them. In this presentation, Jonmichael Hands, vice president, Storage Business Development, Chia Network, will explain the “proof of space and time” (storage capacity) concept that is the foundation of the Chia network, and why SAS is a great technology for farming Chia. He will also discuss the difference between Chia plotting and farming and how they are carried out with the use of storage devices.
The presentation will be moderated by Cameron Brett, president of the SCSI Trade Association (STA) and director of Enterprise and Cloud Storage Marketing at KIOXIA America. SAS is the leading storage interface used in enterprise servers and storage deployments in data centers.
LEARNING OBJECTIVES
NVMe and SNIA are both working on standards related to Computational Storage. The question that is continually asked is are these efforts are compatible or at odds with each other. The truth is that many of the same people are working on both of these standards efforts and are very interested in ensuring that they work together as opposed to conflicting with each other. This presentation will discuss how the two standards efforts go hand in hand, the aspects of the SNIA API that support the NVMe efforts, and the NVMe efforts to share common architectural structures with those defined in SNIA. As part of this discussion, a lexicon of terminology used in both organizations will be presented and the decoder ring that allows you to understand one document in terms of the other in spite of some differences in names used.
LEARNING OBJECTIVES
The next generation of automobiles moves to the adoption of PCIe for data communications in vehicles, and the JEDEC Automotive SSD solution enables a high performance, high reliability solution for this shared centralized storage. Features such as SR/IOV highlight the requirements of these computers on wheels with multiple SoC functions for vehicle control, sensors, communications, entertainment, and artificial intelligence.
LEARNING OBJECTIVES
Computational storage is well known for its data movement reduction, leading to higher application performances. The lower power consumption is another benefit which is often highlighted, but a full system life cycle analysis is recommended to get the full picture of the energy used and its related environmental impact.
In this talk, a detailed carbon footprint analysis will be described for a standard system and a computational storage based system. That will help in understanding the computational storage benefits in term of carbon footprint, and will provide some opportunities to pave the way for a roadmap of next generation eco-designed computational storage devices.