Up until now, much of the apparent innovation of IoT has been centered around devices, platforms, ancillary technology areas like Edge, 5G, and the new use cases they create. Yet, the true value of IoT is in the data. IoT and its ecosystem is continuously and steadily creating new data silos making it imperative for enterprises to analyse and handle this data much more effectively than ever. IoT application areas such as smart cities, smart retail, connected devices, smart homes and mobility are demanding changes in the way this voluminous data is analysed.
With data also comes the security, storage, governance, privacy aspects, and the need to realize its full potential. Furthermore, IoT being a mesh of devices, machinery, firmware and nodes, its critical to ensure that the data is adequately transferred, stored and processed. This keynote will touch upon the fundamental challenges to IoT data management, and the nuances of Storage and New age Datacenter technology to support the ongoing IoT revolution.
This presentation will provide insights on how the data explosion has changed the protection landscape with IOT devices, cloud, data staying for longer times, and the challenges these bring out with edge devices, data centers etc.
Driven by the demand for fast response time to users when challenged with huge data sets, Computational Storage has emerged as a fast growing data center infrastructure tool, especially with the transition to PCIe Flash. Adding intelligence direct to Flash storage devices is a simple idea that can save expensive and time consuming data movement to the host CPU while parallelizing compute across storage drives. This presentation will discuss the technology and market demands, application use cases and standardization efforts that are propelling Computational Storage to be a leading Flash based solution for modern database, big data/analytics, CDNs and emerging AI/ML workloads.
a. Understand need of metric data analysis, advantages, challenges and possible solutions for efficient data service management.
b. Understand solutions through open source data management tools like Open SDS.
This presentation is about development of DApp (Decentralized Application) on top of blockchain as a service. One such Application could be secure, shared ledger for data management operations.
a. Why and where Blockchain is needed?
b. How DApp can be developed on Blockchain?
c. Current Limitations of Blockchain
a. Dependable parameters for Neural network analysis for IO profiling
b. Tuning Hyper Parameters to improve classification
c. Tuning and or calibrating IO staging strategies in Storage Controller
In today’s competitive business environment, storage management providers are continually working towards improving the business value to their customers. The enterprises are deploying large scale distributed storage subsystems to cater high workload demands. The challenge for the storage management services is proactively finding performance bottlenecks, health checks, notify risks and prevent before they occur. The prior knowledge of the known issues from other customer storage deployments for correlation is a challenge. To address these challenges, they are implementing solutions based on the Artificial Intelligence including Machine Learning (& Deep Learning) which require large data sets to derive insights and run predictive analytics. The data from different customer deployments will give even better predictive insights. It requires architectural changes to the storage management services to deploy on the cloud designed by cloud native services those are reliable and auto scalable. The cloud native services such as kubernetes cluster, docker, lambda functions, object storage, elastic search, API gateway services and NoSQL e.g. dynamodb/cassandra for data lake are helping to manage storage infrastructure seamlessly. The cloud based AI services such as Amazon Sagemaker/ IBM Watson/MS Azure ML are used for integrating with data lake to run predictive analytics. The experience of addressing different storage management challenges using cloud native services will be shared.
a. Inevitable need of Artificial Intelligence (AI) application in storage management
b. Understanding and choosing the right set of cloud native services
c. Future of storage management service architecture
Standards prevent vendor lock-in and allow the simple development of cross-cloud applications that will revolutionize computing in the same way as the development of the Internet. Under the charter of DoT and Meity, the Cloud Computing Innovation Council of India is working with TSDSI (Telecommunications Standards Development Society, India) to develop cloud standards for India that will be useful in the Indian context. This effort is slated to be complete by March 2020. The talk will give an overview of the methodology used as well as the current state of the storage standards.
Many organizations, big and small, are moving their data into the cloud. New startups, (even banks) have adopted the model of starting off by keeping their entire data in the cloud. However, many do not properly assess its security implications. In this presentation, I will start by explaining different cloud architectures and the kind of security they offer. I will then explain the new attack vectors that have come up due to cloud storage and computing. I will give examples of how attackers have targeted cloud users and their data in the cloud. In the end, I will give my checklist on how data should be stored securely in the cloud.
a. Understand various cloud architectures and compare their level of security offered
b. Learn about recent attacks on Cloud infrastructure and how they could have been avoided
c. Get a checklist of how data can be stored securely in the cloud
Storage tiering is a feature that dynamically moves data between different types of storage to meet space, performance and cost requirements. Storage tiering policies place the most frequently accessed data onto the highest performance storage and less frequently accessed or old age data onto low performant or cheaper storage that at any given point of time one copy of data is present across all the storage tiers.
a. Introduction of storage tiering
b. Introduction of how deduplication works
c. Challenges with Storage tiering in dedupe environments
a. Understand GenZ technology
b. Understand data transfer between different Gen-Z components
c. Benefit of GenZ to ecosystem and next generation of computing
Workshop on SNIA Swordfish basic web client and Swordfish API emulator.
a. Installation, usage and contribution to Swordfish basic web client
b. Installation, usage and contribution to Swordfish API emulator
c. Other SNIA swordfish tools
Internet of things has revolutionized the industry but it has given a big challenge to data center solutions. The rapid generation of massive amounts of data and it's processing requires data centers to be highly scalable and reliable. We will take a look at how this problem can be solved by enhancing the data center policies and storage tiering to optimize overall performance.
a. Internet of Things
b. Data center problems due to massive amount of IoT data
c. Data center policies and storage tiering to optimize IoT data handling
a. Memory design challenges in handheld devices
b. Tool design for memory architecture exploration
c. Potential uses of NV memory in handheld devices
In this presentation, a brief introduction of new interconnect standard CCIX will be presented, followed by storage use case acceleration achievable with CCIX. CCIX is a new class of interconnect focused on emerging acceleration applications such as machine learning, network processing, storage off-load, in-memory data base and 4G/5G wireless technology. The standard allows processors based on different instruction set architectures to extend the benefits of cache coherent, peer processing to a number of acceleration devices including FPGAs, GPUs, network/storage adapters, intelligent networks and custom ASICs. FPGA acceleration has become a de-facto standard for obtaining the high computational throughput in present times. Storage domain can take advantage of FPGA acceleration for compute intensive tasks like compression, encryption etc. Through CCIX interconnect FPGA accelerators can act as coprocessors to host, sharing data structures seamlessly, with reduced data transfer latency.
Presentation shall discuss two use cases 1. Memory expansion model over CCIX 2. Storage with compute offload. Details on programming model for realizing the true potential of CCIX interconnect will be presented.
a. CCIX interconnect relevance & Introduction
b. Role of Coherent accelerators in compute intensive tasks
c. Programming model for realizing the true potential of CCIX interconnect
NFS-Ganesha is a user-mode file server for NFS (v3, 4.0, 4.1, 4.1 pNFS, 4.2) and for 9P from the Plan9 operating system. It can support all these protocols concurrently. The project was started around 2009 and it got well matured over past few years and includes participation including CEA, IBM, Red Hat. There are a lot protocol specific features added including LABELED NFS, Delegations to nfs-ganesha layer. The is workload specific changes made to nganesha layer which includes async op and non blocking io's. There are other projects like storhaug which integrates the nfs-ganesha to ctdb and so on.
a. NFS Protocols 4 and higher
b. NFS-Ganesha Project
c. Highly available NAS solution for NFS-ganesha
Evolution of cloud based technologies have revolutionized modern IT environment, while posing new challenges. Rapid generation of data, and variety of mechanisms to maximize its value have resulted in new cloud offerings rapidly. Since there can't be one solution to all the problems, its obvious that industry is inclined toward multi-cloud environment. In this talk we'll discuss about an approach to make the better use of multi-cloud environment.
NVMe over Fabrics protocol (NVMe-oF) is designed to connect and scale NVM storage over a network with minimum latency overhead. However, existing NVMe-oF solutions cannot harness the low latency benefits of next-generation storage technologies, since these solutions utilize expensive CPU resources and require protocol conversion (NVMe-oF to NVMe) that adds overhead. This can be solved by developing an Ethernet SSD with native NVMe-oF protocol support. In this talk we examine the design of NVMe-oF based Ethernet SSD's and the use-cases they are enabling.
Learning Outcomes
a. Overview of NVMe-oF protocol and existing solutions
b. Design of NVMe-oF Ethernet SSD controller
c. Use cases
Over the past decade, Fibre Channel (FC) technology has been the norm when it comes to data centers. With storage applications such as data backup, replication and disaster recovery, it was used to connect storage to servers. FC backbone uses highest speed of 64 Gigabits per second (32G), which explicitly requires leased lines consisting of FC switches and bridges to be connected to Storage, and implicitly adding more cost to the infrastructure. Using latest technology of 100G RDMA (Remote Direct Memory Access), it is possible to achieve transfer speeds of 100 Gbps (Gigabits per second) with lower latency on the existing Ethernet switches and routers over long distances up to 300kms. Will focus on how 100G Ethernet could replace the FC back-bone for storage connectivity with High Availability (HA) and Disaster Recovery (DR).Basically explains the front (client-facing data ports ) and backend support (Storage and HA interconnect) on 100G infra with Storage platforms .
a. RoCE over Long distance
b. Cluster and HA on Ethernet N/w
c. What is RoCEv2
Also we are giving the opportunities which can bring in with the adoption of Optane in Android space in terms of new features. The analysis data is very exciting and the cost vs performance matrix is also interesting.
Learning Outcomes
a. Performance impacts in Android OS storage space
b. Use cases & advantages of Optane in Android/Storage space
c. NVMe storage capabilities and its future expansion possibilities
The new NVMe SSD interfaced can be connected across a Fabric. In fact it can be connected across lots of different fabrics: Ethernet (3 approaches), Fibre Channel, InfiniBand, and PCIe to date. Data Centers want to share storage readily among multiple compute nodes and be able to perform clustering, failover, and other system-wide operations at NVMe SSD speeds. NVMe over Fabrics (NVMe-oF) is the solution. This talk will describe the technology in its many forms. Describe use cases, for both Enterprise and cloud, where it is being applied. Then finish with potential future directions it is heading.
a. NVMe over Fabrics
b. Multiple versions of NVMe over Fabrics
c. Future NVMe over Fabrics direction
Learning Outcomes
a. NVMe over Fibre Channel architecture and its advantage. NVMe over Fibre Channel internals and how it is one of the most appropriate host attach protocol for fast storage media.
b. Performance comparison of NVMeFC Vs SCSI over Fibre Channel. Comparative analysis of performance results on various metrics like response time, IOPs and server CPU utilization using various workloads.
c. Customer benefit on adoption of NVMe over Fibre Channel. Detailing out benefits like low capital investment, addition of new workloads and ease of migration and administration.
Learning Outcomes
a. Move a step ahead from single dimensional analysis of storage performance management from considering just stored-time-series data to multi-dimensional analysis
b. Understand how data-science is helping evolve storage management
c. Understand the challenges of Ethernet storage management domain
a. Understand how the clustered file systems are falling short for container workloads
b. Introduction to kubernetes's various approach in consuming Persistent storage
c. Learn pro's & con's using which a parallel, clustered filesystem in containerized environment
Learning Outcomes
a. Understand CSI in Kubernetes
b. Understand GoCSI framework
c. Learn how to develop CSI drivers for Kubernetes
Learning Outcomes
a. Blockchain concepts
b. Business promises of Blockchain and Storage together
c. Blockchain use cases for the storage domain
Reliable estimation of prediction confidence remains a significant challenge in machine learning. We usually expect past performance to indicate future performance. When we deal with risk-sensitive systems – where the cost of a bad decision can be very high, and prediction accuracy is not the only objective; we need a multidimensional perspective about the forecast models. So, if the user is given a confidence of each new predictions made by the model, then a more meaningful action can be taken.
In this talk we will discuss about Conformal Prediction Framework and how it can be leveraged across various machine learning algorithms used in Storage Industry (e.g. Disk Drive Failure detection and storage demand forecasting). Furthermore, as an example, we will describe how this framework can be translated to time-series, classification and regression problem which will give a confidence (indication of the quality of each prediction) and credibility (filter mechanism with which we can “reject” certain predictions).
Learning Outcomes
a. Conformal Prediction framework for reliable time-series forecasting and application in regression and classification problems
b. Fundamentals of On-line learning approach and handling concept drift
Learning Outcomes
a. A general direction on tools and application available to achieve results.
b. Gain insightful information on how analytics can be performed.
c. Datacenter/Storage Admin, C-level management and other stakeholders can take key decisions based on actionable information obtained.
Learning Outcomes
a. What are the main challenges running workload on containers?
b. How do we solve the problem of dynamic requirement of persistent storage for containers?
c. How do we take application consistent backup of different workloads in Container orchestrator platform?