Aadhaar is a digital identity platform for 1.2 billion residents of India. The presentation will discuss the technology and business process innovations involved in the creation of Aadhaar. The role of Aadhaar as a key component of the India Stack with the goal of transforming India into a digital nation where all services can be availed in a paperless, cashless and presence less manner will also be discussed.
With the ever-growing data churn, companies are in need of newer ways to utilize their storage more efficiently than ever. Digital transformation is here to stay resulting in huge volumes of digital data, addressing Bigdata issues, having an integrated data management and cloud strategy to support scale of businesses. This put together calls for a strategy that plans, integrates and manages the storage infrastructure to be ‘transformation ready’. The talk will touch upon crucial factors in Storage and Data management that range from Datacenter elements, data movement across the technology stack, up until backup, archival and business continuity carefully highlighting the challenges and opportunities at hand. The talk will also summarize the most astounding trends in traditional as well as software-defined storage that were registered so far, and carve a depiction about what we should expect in 2018 and beyond.
In today’s technology-driven world, demand for computing is increasing on an almost daily basis. To meet this increasing demand, the implementation of consolidated servers and data centers is growing rapidly. However, computing and storage components consume more than 40% of energy requirements in addition to power distribution and cooling equipment.
According to the server energy model:
Pt = Pf + Pv
Where Pt = Total power consumption
Pf = Fixed power consumption by memory modules, disks, IO resources
Pv = Variable power consumption by CPUs (different power requirements at different operational frequencies)
Learning Objectives
A comparative study was conducted on multiple dimensions of parallel I/O streams using user space(SPDK) NVMe as well Linux Kernel space NVMe drivers. We compared IOPS, latency, throughput for sequential & random read / writes on ARM64 (ThunderX2) and x86 (Intel skylake) storage servers using latest NVMe PCIe SSDs. We also studied the effect of varying I/O queue depth and number of CPU cores. Latency for submission, completion, peak and average latencies are compared for ARM64 and x86 platforms. The purpose of the case-study is to provide information on the current state of the SPDK NVMe and Linux Kernel NVMe driver performance which can then be used by designers to architect their products.
Learning Objectives
Learning Objectives
Learning Objectives
Learning Objectives
The SNIA’s Scalable Storage Management Technical Work Group (SSM TWG) has created and published an open industry standard specification for storage management that defines a customer centric interface for the purpose of managing storage and related data services. This specification builds on the DMTF’s Redfish specification using RESTful methods and JSON formatting.
This presentation shows how Swordfish extends Redfish, details Swordfish concepts and talks about CSDL and JSON schema formats and ODATA protocol for modelling resources.
Pre-conference learning
www.snia.org/forums/smi/swordfish
2017 presentation at SDC India DOWNLOAD
Internet of Things is enabling traditional and newer operational workflows to be digitized for continuous measurement. The measured data is useful in predicting the operational failures/inefficiencies and take actions before they occur. The aggregation of measured data from millions of devices is overwhelming the public network bandwidth. Also, there are privacy and data sovereignty concerns as far as data on the move is concerned. Thus, most of the IoT platforms (Hyperscalers) offer edge (device side) footprint and enable data services at the edge that were available only in cloud few years back. The edge reduces the latency and provides local analytics for quicker actions.
In this talk, we present the implications on storage architecture as IoT data pipelines keep evolving from client server to distributed architecture to cater millions of devices; geographically spread. Also we discuss how data latency, privacy, sovereignty and need for governing massive amounts of data are driving the newer storage constructs.
Ozone brings in a new storage paradigm in Hadoop called object storage. It will co-exist with HDFS to provide file store and object store functionality in the same Hadoop cluster. Ozone will also solve the scalability and small file problem of HDFS, where users can now store trillions files in Ozone and access them as if they are on HDFS. Ozone plugs into existing Hadoop deployments seamlessly and programs like Hive and Spark work without any modifications. This talk looks at the architecture, reliability, and performance of Ozone. In this talk, we will also explore Hadoop Distributed Storage Layer, a block storage layer that makes this scaling possible, and how we plan to use the Hadoop Distributed Storage Layer for scaling HDFS. We will demonstrate how to install an Ozone cluster, how to create volumes, buckets and keys, how to run Hive and Spark against HDFS and Ozone file systems using federation, so that users don’t have to worry about where the data is actually stored. Ozone SDK will also be covered in this talk. In other words, a full user primer on Ozone will be part of this talk.
Learning Objectives
Pre-conference learning
SNIA Webcast: File vs Block vs Object Storage
Learning Objectives
Pre-conference learning
Pre-conference learning
For decades, architecting data security solutions revolved around the idea of building a fortress around the data. This is called perimeter-centric security.
This talk is aimed at introducing and discussing in detail about data-centric security. Data-centric security is about securing data without artificial physical / infrastructure boundaries. That is, instead of securing the applications (in-use), endpoints (at-rest) & network (in-motion) infrastructure that use, store & move data respectively, data-centric security embeds security controls within the data itself.
Learning Objectives
Pre-conference learning
Oxymoron: Computing on Encrypted Data - Srinivasan Narayanamurthy @SDC India 2017
Learning Objectives
The emerging field of Genomics Medicine requires physicians, data scientists and researchers to analyze huge amounts of genomics data quickly. This poses challenges on the backend infrastructure including the storage. In this talk, we present the genomic workload characteristics, its requirement on the backend storage sub-systems and how an composable infrastructure approach based on scale out file system can enable IT architects to customize deployments for varying functional and performance needs.
Learning Objectives
Pre-conference learning
With the current hyperscale datacenters, managing multi-vendor storage hardware using one simple user friendly tool is the datacenter admins desire. Server and storage Industries are trying to solve this common problem by providing a standard way of storage management. DMTF and SNIA have attempted to standardize the storage management using CIM and SMI-S standards for a decade. Now DMTF and SNIA have reviewed the lessons we learnt in a decade and have come up with Redfish and Swordfish. A simplified and easy to implement and use standards for the next generation of storage management. In addition to the standard based storage management, below are the common ask on the next generation of storage management.
In the last couple of years, the case for HPC in the cloud is growing stronger. But still, the HPC industry lies far behind enterprise IT in its willingness to outsource computational power. One of its reason being storage - as none of the built-in storage solutions available across the public cloud providers are suitable for applications with high bandwidth requirements.
The proposed presentation is aimed to discuss in detail different architectures (weighing Pro’s & Con’s) that can be used to build a reliable parallel filesystem in the cloud (showcasing AWS, IBM Spectrum Scale as an example) and data lifecycle techniques that help reduce OPEX cost by effectively managing parallel filesystem in the cloud.
Learning Objectives
More than a decade old data architecture isn’t enough for today’s data-driven businesses which are heavily dependent on AI/ML/DL. As the enterprises begin to operationalize these AI/ML workflows, they would need to optimize on storage I/O performance to feed to massively parallel GPU based compute. With growing IoT footprint, data management and AI/ML based compute challenges span across from edge, to the core and to the cloud. In this talk we propose a need for a modern data engineering and management pipeline to address the above challenges. Specific learning objectives being, how some of the existing data engineering workflows need to be re-thought, which includes dynamic data indexing, access pattern aware data layout etc. The talk would also cover other emerging data engineering challenges like data reduction and data quality assessment with specific focus on edge/core vs. cloud. The talk would also bring out any ongoing research towards addressing the mentioned challenges.
Learning Objectives
Learning Objectives
Learning Objectives