the future of data infrastructures the future of data...

13
Prepared for: Flash Memory Summit Santa Clara Convention Center August 6-9, 2018 Western Digital Corporation The Future of Data Infrastructure Technology Brief Western Digital’s Software Composable Infrastructure

Upload: others

Post on 05-Jun-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

Prepared for:

Flash Memory Summit

Santa Clara Convention Center

August 6-9, 2018

Western Digital Corporation

The Future of Data Infrastructure

Technology Brief

Western Digital’s Software Composable Infrastructure

Page 2: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

1

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

SAFE HARBOR - DISCLAIMERS Forward-looking Statements

This document contains forward-looking statements that involve risks and uncertainties, including, but not limited to, statements regarding our product and technology positioning, the anticipated benefits of our new technologies and transitioning into software composable infrastructures. Forward-looking statements should not be read as a guarantee of future performance or results, and will not necessarily be accurate indications of the times at, or by, which such performance or results will be achieved, if at all. Forward-looking statements are subject to risks and uncertainties that could cause actual performance or results to differ materially from those expressed in or suggested by the forward-looking statements.

Additional key risks and uncertainties include the impact of continued uncertainty and volatility in global economic conditions; actions by competitors; difficulties associated with go-to-market capabilities and transitioning into software composable infrastructures; business conditions; growth in our markets; and pricing trends and fluctuations in average selling prices. More information about the other risks and uncertainties that could affect our business are listed in our filings with the Securities and Exchange Commission (the “SEC”) and available on the SEC’s website at www.sec.gov, to which your attention is directed. We do not undertake any obligation to publicly update or revise any forward-looking statement, whether as a result of new information, future developments or otherwise, except as otherwise required by law.

Contents1. Introduction ………………………………………………………………………………………………… 2

2. Today’s Data Center Challenges …………………………………………………………………… 2

3. Diversity of Applications and Workloads ……………………………………………………… 3

4. Hyper-Converged Infrastructures ………………………………………………………………… 3

5. Need for Composable Infrastructure …………………………………………………………… 4 Open Standards are Essential for SCIs ……………………………………………………………… 5How an SCI Works …………………………………………………………………………………………… 5Benefits of an SCI …………………………………………………………………………………………… 6Components of an SCI …………………………………………………………………………………… 7

6. Western Digital’s Commitment …………………………………………………………………… 8 Product Options …………………………………………………………………………………………… 8Open Management API …………………………………………………………………………………… 8Driving the Ecosystem …………………………………………………………………………………… 9

7. Initial TCO Results ………………………………………………………………………………………… 9

8. Looking Ahead ………………………………………………………………………………………………10

Page 3: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

2

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

“Data analytics has become animportant part of this new [data]look enabling organizations to extract insights that could lead to better business decisions, efficiencies in productivity and automation, or unexpected value.”

IntroductionOur data-centric world is evolving. Big data is getting bigger and faster. Fast data is getting faster and bigger. High-performance applications are increasing in diversity and being fueled by artificial intelligence, machine learning, autonomous vehicles, smart cities and cars, virtual and augmented reality, and hyper-analytics. This has also created an increase in workflow complexities. As such, the task of orchestrating and managing the data infrastructure at scale has never been more challenging. Big data is being accessed by multiple applications over longer periods of time and extracted from storage repositories for use by fast data applications. The need to ‘spin up’ or ‘spin down’ application instances quickly, without disrupting concurrent applications, is required.

Western Digital provides insights into the architectural shifts underway to ensure that performance, scale and efficiency can be quickly and easily provisioned and optimized for the most demanding data environments. Through the convergence of emerging technologies, open standards, innovation and a growing ecosystem, the future is in composable infrastructures.

Today’s Data Center ChallengesEvolving from a record or a log of events, recordings or measurements, data is now taking on a whole new look driven by the four Vs (volume, velocity, variety and value). With this shift, data analytics has become an important part of this new look enabling organizations to extract insights that could lead to

better business decisions, efficiencies in productivity and automation, or unexpected value.

Data is no longer generated from just applications. It comes from diverse sources (e.g. mobile devices, production machines and sensors, video surveillance systems, IoT and IIoT devices, healthcare monitors and wearables, to name a few). It is being generated in both large-scale data centers at the core, as well as from remote and mobile sources at the edge of the network. To process and analyze this data for immediate or future use, fast data and big data environments are rising quickly.

Big data are very large datasets that need to be analyzed, and through computations and algorithms, unmask trends, patterns and associations between disparate data groups. The analysis can deliver valuable insights derived from ‘what happened’ to ‘what will likely happen’ (predictive analysis), and can make new connections that help achieve better outcomes. This analysis is based on data captured from the past. The ability to respond to events as they happen is also required, paving the way for fast data environments that process and transform data as it is captured. These applications leverage the algorithms derived from big data to provide real-time decisions and results.

Data analysis leads to actionable insights

Page 4: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

3

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

“What is required for today’s data-centric architectures are capabilities that enable more control over the blend of resources that each application needs so that optimized levels of processing, storage and network bandwidth can be scaled independent of one another.”

Applications within an enterprise change constantly throughout the day requiring businesses to be agile in their decision-making. They need the flexibility to provision IT resources quickly to meet application demands when and where they are required. The ability to react to the expanding diversity of big and fast data applications is vital and requires the allocation of more compute, storage and network resources at scale, and in real time across the enterprise.

Diversity of Applications and WorkloadsAs big data and fast data environments proliferate, they are breaking the boundaries of traditional infrastructures and system architectures. The ‘general-purpose’ technologies that have served so well in the past are reaching their limits of scalability, performance and efficiency. When data-intensive workloads are supported by general-purpose architectures, a uniform ratio of resources are typically used to address all compute processing, storage and network bandwidth requirements.

This ‘one size fits all’ approach is not effective for these diverse data-intensive workloads.

What is required for today’s data-centric architectures are capabilities that enable more control over the blend of resources that each application needs so that optimized levels of processing, storage and network bandwidth can be scaled independent of one another. These extreme data-intensive workloads that drive analytics, machine learning, artificial intelligence and smart systems demand an infrastructure that is both flexible and composable.

Hyper-Converged Infrastructures

Given the challenges associated with general-purpose architectures (fixed resource ratios, underutilization and overprovisioning), converged infrastructures (CIs) emerged delivering a preconfigured package of hardware resources in a single system. The compute, storage and networking components are discrete and managed through software. CIs have progressed into hyper-converged infrastructures (HCIs) where all of the hardware resources are virtualized, delivering software-defined computing, storage and networking.

Though HCIs combine compute, storage and network resources into a single virtualized system, it is not without inefficiencies. For example, the scalability limits are defined by the processor, and access to resources are made through the processor. So, to add more resources, such as storage, memory or networking, HCI architectures deliver additional processors even if compute capacity is not required.

Page 5: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

4

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

Additionally, when data is moved to a public cloud for archival in an HCI model, there is a lack of performance predictability due to noisy neighbors.

With the need to optimize infrastructure provisioning to meet specific application requirements, data center architects are trying to build flexible infrastructures using inflexible building blocks. A recent market survey1 of over 300 mid-sized and large enterprise IT users determined the following:

• Only 45% of total available storage capacity is provisioned

• Only 45% of compute hours and storage capacity are utilized

• Only 30% agile in the time it takes to provision new or change compute or storage instances

• Only 40% confident in ability to meet stated infrastructure SLAs and its associated components

The fixed building block approach cannot achieve the level of flexibility and predictable performance needed in today’s data center as the disaggregated HCI model needs to be enabled and become easily composable through software tools based on an open application programming interface (API).

Need for Composable InfrastructuresA composable infrastructure (aka software composable infrastructure or SCI) is a data center architectural framework whose physical compute, storage and network fabric resources are treated as services. As such, high-density compute, storage and network racks use software to create a virtual application environment that provides whatever resources the

application needs in real-time to achieve the optimum performance required to meet workload demands. It is an emerging market category sometimes referred to as ‘infrastructure as code,’ and goes beyond simply converging or hyper-converging IT resources into a single integrated unit by optimizing these resources to improve business agility.

In an SCI, virtual servers are created out of independent resource pools comprised of compute, storage and network devices, rather than partitioned resources that are hardwired as HCI servers. Thus, the servers can be provisioned and re-provisioned as needed, under software control, to suit the demands of particular workloads as the software and hardware components are tightly integrated.

“[In an SCI] the servers can beprovisioned and re-provisioned as needed, under software control, to suit the demands of particularworkloads as the software andhardware components are tightly integrated.”

1 From the IDC report entitled, “Quantifying Data Center Inefficiency: Making a Case for Composable Infrastructure.”

Today’s data-centric data center architectures

Data-Centric Data Center Architecture

StorageCompute

Data

I/O

Networking

Page 6: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

5

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

Open Standards are Essential for SCIs:

Utilizing infrastructures with fixed levels of resources, server utilization (as of 2016) ranged between 6 and 15 percent for non-virtualized servers, and about 30 percent for virtualized ones. With an API on the composing software, an application could request whatever resources are needed, delivering real-time server reconfigurations on-the-fly, without human intervention, and a step toward the self-managing data center.

Through a single management API, an SCI can integrate operational silos, logically pool resources and automatically compose and recompose blocks of disaggregated IT services to effortlessly meet the changing needs of an application. This enables IT personnel to provision workloads just as quickly and efficiently as public cloud providers, but in a secure, scalable on-premises private cloud

environment. When more compute, storage or networking resources are needed, one simply changes the virtual server configuration for that application from a computer. If a virtual server dies, the boot device can be moved to a new virtual server. The infrastructure is not only flexible, but composable, reducing underutilization and overprovisioning while creating a more agile, cost-effective data center environment.

Therefore, in order for the SCI initiative to be successful, it must be backed by an ecosystem to create those differentiated products for it to succeed and evolve. An open management API is essential for

success as it provides developers with public access to software and code for development, as well as the set of requirements that govern how an application communicates and interacts with another.

How an SCI Works:

An SCI resides on-premises in the data center. Since compute, storage and network are treated as services, the ‘composable’ part refers to its ability to make those resources available on-the-fly, depending on the needs of each physical or virtual application. Therefore, a management layer within the API is required to not only discover and access the resource pool, but also to ensure that the right resources are in the right place when they are needed.

In this infrastructure, if the workload needs additional compute or storage resources, a few keyboard clicks adds more to the server. Once the workload has completed its tasks, the resources can then be returned to the shared pool for use by other applications. This composition and re-composition under software control is fast and without physical involvement or reconfiguring the equipment. Resources are no longer trapped in separate silos eliminating ‘stranded storage’ and underutilized CPUs.

“Through a single management API, an SCI can integrate operational silos, logically pool resources and automatically compose and recompose blocks of disaggregated IT services to effortlessly meet the changing needs of an application.

Page 7: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

6

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

Benefits of an SCI:

The main benefit of SCIs is their ability to disaggregate compute, storage and network fabric resources into shared resource pools that can be available for on-demand allocation with an end result that vastly improves server utilization, performance, flexibility and agility in the data center. And, since the SCI works in conjunction with physical infrastructures already in place, is very cost-effective since resources are not wasted or the time it takes to deploy new applications. SCIs are also ideal for enabling SLA-driven architectures for compute and storage where various workloads are automatically assigned resources that have sufficient performance and capacities to meet application-specific SLAs.

The key benefits of SCIs can be summated into three categories and include its ability to: (1) Automate Agility; (2) Improve Efficiency; and (3) Improve Performance Predictability.

1. Automate Agility SCIs enable the access and management of a resource pool that can be routed dynamically on demand to where they are needed the most and are logically pooled so that administrators do not have to physically configure hardware to support a specific application. As such, SCIs can adapt more quickly to changing workloads and provision them more purposefully in support of a range of big and fast data applications.

2. Improve Efficiency SCIs improve economic efficiency by enabling better utilization of compute and storage resources. By reducing overprovisioning of resources, it improves power consumption, cooling, associated costs, etc. It also enables resources to be expanded as needed so organizations can buy smaller amounts of storage upfront, for example, and expand the footprint gradually and evenly over time, based on need.

3. Improve Performance Predictability With the ability to allocate resources properly to meet application demand, less wasted or unused IOPS (input/output operations per second) occur as resources perform exactly as expected and are available when required. This reduces the variability in business and/or operating processes and execution as the lower the variance, the more consistent the outcome. Unpredictability in performance makes it difficult to establish normal operating parameters, costs become difficult to forecast, and profits and long-term operating expenses become difficult to calculate.

“The key benefits of SCIs can be summated into three categories and include its ability to automate agility, improve efficiency and improve performance predictability.”

API

APIEnterpriseApplication

MobileDevice

Sensors

Wearable Device

TV

Games

Web AppsCar

APIs enable workloads to be provisions quickly and efficiently

Page 8: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

7

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

“NVMf promises to deliver the lowest end-to-end latency from application to storage enabling SCIs to provide the data locality benefits of direct-attached storage (low latency and high-performance), while delivering agility and flexibility by sharing resources throughout the enterprise.”

In a nutshell, SCIs can automatically and effortlessly compose and recompose blocks of disaggregated IT services to meet the changing needs of an application.

Components of an SCI:

There are two main components that enable an SCI. The first is software to dynamically provision and manage the resources, while the second is a network protocol to disaggregate the resources and make them shareable amongst multiple applications and servers. In the case of software, the IT resources can be dynamically provisioned and managed through an open API without physically touching hardware.

A networked protocol is equally important to an SCI enabling compute and storage resources to be disaggregated from the server and available to other applications. Connecting compute or storage nodes over a fabric is important as it enables multiple paths to the resource. Emerging as the frontrunner network protocol for SCI implementations is NVMe™-over-Fabric (also denoted NVMf, NVMef, NVMeoF or NVMe-oF™). Regardless of the acronym, NVMf promises to deliver the lowest end-to-end latency from application to storage enabling SCIs to provide the data locality benefits of direct-attached storage (low latency and high-performance), while delivering agility and flexibility by sharing resources throughout the enterprise.

Historically, Non-Volatile Memory Express™ (NVMe) technology is a streamlined, high performance, low latency interface that utilizes an architecture and set of protocols developed specifically for persistent flash memory technologies. It increases non-volatile memory storage performance in locally-attached PCIe-based servers and SSDs by eliminating the DAS bottlenecks associated with legacy HDD interfaces and SCSI command stacks. SSDs based on PCIe NVMe are well-suited for applications that require extremely fast data transfer rates, a significant number of I/O operations performed per second, and short access times when delivering data.

The NVMe protocol has been extended beyond local attached server applications by providing the same performance benefits across a network through the NVMe-over-Fabrics specification. This specification enables flash devices (such as PCIe SSDs and storage arrays) to communicate over networks, (such as InfiniBand™ or Converged Ethernet), delivering the same high performance, low latency benefits as local attached NVMe. As NVMe was designed for PCIe as a server attachment for storage, NVMf extends NVMe for use on network fabrics where the primary advantage is scale. While PCIe supports a limited number of servers and devices, there is virtually no limit to the number of servers that can share NVMf storage or the number of NVMf storage devices that can be shared.

As networks can be slow, narrow in focus or too expensive, the most cost-effective bandwidth is within servers, which is why so much storage locally resides inside of them. PCIe-based storage devices are helping to drive this migration because the interface protocol enables more data lanes to be added, which in turn, delivers fast I/O performance with very low latency built into every server. And, since almost every device supports a PCIe interface, there are no drivers to install, making it an effective interface for composable infrastructures.

NVMf extends NVMe for use on network fabrics*

*NVM Express, Inc.

Page 9: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

8

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

“[Western Digital] will unveil its new OpenFlex™ architecture that delivers breakthrough levels of scalability, efficiency and performance for big data and fast data applications, enabling a new era in NVMf-based SCIs.”

Western Digital’s CommitmentWith market leadership in developing innovative products, systems and solutions based on NVMe technology, Western Digital is further investing in the NVMf standard to enable a composable SLA-driven architecture for compute and storage covering flash-based media, hard drives and eventually memory. At Flash Memory Summit (August 6-9, 2018), the company will unveil its new OpenFlex™ architecture that delivers breakthrough levels of scalability, efficiency and performance for big data and fast data applications, enabling a new era in NVMf-based SCIs. Western Digital brings the essential building blocks, technological advancements, tools, resources and contacts to enable and help accelerate market adoption and will provide an open management API and storage abstraction layer to accelerate time-to-value.

Product Options:

Western Digital will initially offer two composable storage options - a high-performance, low latency fabric device for mission-critical applications and data, and a high-capacity fabric device for tiering, data protection and disaster recovery. The breakthrough news is that Western Digital has enabled disks to be addressed within an NVMf implementation for the first time so that all data center storage can be addressed in a similar manner. This new NVMf architecture is a huge step toward the software-defined data center where resources are assigned to applications without regard to where they are physically located. This is the essence of the composable infrastructure where compute, storage and network resources can be virtually configured and treated as a service for a specific application without the need for physical configuration.

For flash-based high-performance, mission-critical applications and data, the company will offer its OpenFlex F3000 Series fabric device which is a hot-swappable NVMe flash device that is slated to support capacities ranging from 12.8TB2 to 61.4TB. The initial implementation will be RDMA over Converged Ethernet (RoCE) which is ubiquitous in the data center and cost-effective to deploy. Support for InfiniBand will be forthcoming. Up to ten (10) OpenFlex F3000 Series fabric devices can be housed in the OpenFlex E3000 fabric storage platform which is a separate, but required 3U10 enclosure that can deliver capacity points ranging from 128TB up to 614TB based on the number and type of devices installed.

For data tiering, data protection and disaster recovery, the company will also offer its OpenFlex D3000 Series fabric device. The D3000 is projected to support capacities up to 168TB and hot-swappable for high availability.

Open Management API:

As part of the initial OpenFlex launch, Western Digital is developing an open API to dynamically provision and manage the resources 2 Projected specifications, subject to change. As used for storage capacity, one petabyte (PB) = one quadrillion bytes, one terabyte (TB) = one trillion bytes and one gigabyte (GB) = one billion bytes. Total accessible capacity varies depending on operating environment.

Above: OpenFlex F3000 Series fabric device and 3U enclosure for mission critical apps and data

Below: OpenFlex D3000 Series fabric device and 1U enclosure for data tiering, data protection and disaster recovery

Page 10: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

9

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

allowing flash storage, for example, to be shared by multiple servers in the SCI architecture. Resources can be scaled independent of each other. Capacity can be expanded at will to adapt to changing workloads. The result is the ability to enable heterogeneous compute environments, improve capacity utilization and serve multiple applications with a common pool of high-speed storage. Based on the DMTF RedFish® API (designed to provide secure management for data center infrastructures), Western Digital will make publicly available its open management Kingfish™ API to enable and drive vendor neutral solutions.

Driving the Ecosystem

The OpenFlex SCI initiative will require a complete ecosystem to support it in order to thrive. Included will be infrastructure and system architects, HW/SW/FW developers, contributors, test and validation groups, solutions and application developers, compute/storage/network fabric vendors, system integrators, and users. The company is committed to advancing SCI architectures in data centers globally for use in both mission critical applications and for archival and other data protection schemes. Western Digital is committed to supporting the broad ecosystem of data center software and hardware vendors and will showcase the OpenFlex platform to the industry consortium at FMS 2018.

Initial TCO ResultsTo justify the investment required to migrate over to an NVMf-based SCI platform, the ecosystem will want assurances that this infrastructure strategy will ultimately provide significant business value, a savings in total cost of ownership (TCO) and a return on investment (ROI). In a head-to-head comparison of equipment, components and associated costs required in a traditional HCI versus Western Digital’s new OpenFlex NVMf-based SCI, a 34 percent cost savings can be achieved if all of the equipment and components are purchased upfront3.

However, as the OpenFlex architecture delivers the proper blend of IT resources for on-demand workload allocation, not all of the equipment and components will need to be purchased in one fell swoop. In fact, by implementing a ‘buy-as-you-grow’ strategy, up to a 40 percent savings can be achieved over a traditional HCI, assuming the following:

• 50 to 60 percent of the hardware is purchased upfront

• One server is purchased every quarter for three years

• One OpenFlex F3000 Series fabric device is purchased annually

Network fabrics based on NVMf enable separate resources to be added over time and scale them independent of each other. As such, the ‘buy-as-you-grow’ model enables expenses to be delayed over

“In a head-to-head comparison of equipment, components and associated costs required in a traditional HCI versus Western Digital’s new OpenFlex NVMf-based SCI, a 34 percent cost savings can be achieved if all of the equipment and components are purchased upfront—by implementing a ‘buy-as-you-grow’ strategy, up to a 40 percent savings can be achieved.”

3 Assumptions include: (1) the HCI resources are not fully utilized; (2) CPU cores have 22 percent unused capacity; (3) DRAM has at least 13 percent of unused capacity; (4) storage capacity is at least 20 percent underutilized. The TCO values are based on internal analysis using those assumptions and public component pricing as of July 2018.

Western Digital will make publicly available its open management Kingfish API

Page 11: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

10

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

time, and with that, users can take advantage of lower dollar per gigabyte storage pricing over time, as well as the costs associated with running equipment and components that are underutilized, or worse, not being used.

Other areas of cost-savings when comparing an OpenFlex architecture to a traditional HCI can include:

• Reducing the number of storage racks from two to one

• Utilizing more scaled up servers within the rack to saturate the ports

• Utilizing multi-node servers versus individual servers

• Reducing the number of servers by almost 50 percent

• Upgrading the CPUs and DRAM with more cores and capacity

• Utilizing two 100Gb switches versus four 25Gb switches

• Reducing expenses in power and cooling as a result of the CPUs running faster and more efficiently

• Reducing expenses associated with networking given the efficiencies of NVMf

With Western Digital’s OpenFlex architecture, resources can more easily be added as needed enabling a ‘buy-as-you-grow’ model that delivers an extra savings over three years.

Looking AheadBig data in the core and fast data at the edge have exceeded the capabilities of traditional infrastructures and architectures especially relating to scalability, performance and efficiency. As general-purpose infrastructures

are supported by a uniform ratio of resources to address all compute processing, storage and network bandwidth requirements, they are no longer effective for diverse data-intensive workloads.

Western Digital’s OpenFlex architecture is truly an open composable NVMf platform that disaggregates compute, storage and network resources. Data center architects, cloud service providers, systems integrators, software defined storage partners, OEMs, etc. can now deliver storage and compute services with greater economics, agility, efficiency and simplicity at scale, while enabling dynamic SLAs across workloads. Starting at the rack level, with federated management across multiple fabrics and sites, the OpenFlex architecture is based on the company’s composable vision of:

“The ‘buy-as-you-grow’ model enables expenses to be delayed over time, and with that, users can take advantage of lower dollar per gigabyte storage pricing over time, as well as the costs associated with running equipment and components that are underutilized, or worse, not being used.”

Page 12: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNOLOGY BRIEF

11

Technology Brief | The Future of Data Infrastructure ©2018 Western Digital Corporation or its affiliates. All rights reserved.

• Leveraging an industry leader in both disk and flash storage to deliver a scalable, modular set of NVMf storage devices based on a common interface

• Offering an open API and storage abstraction layer to accelerate time to value

• Reducing the complexities, inefficiencies and performance bottlenecks of current storage offerings to deliver optimum cost and performance

• Enabling SCIs from next-generation container technologies

Today’s data-centric architectures require more control over the blend of resources that each application needs so that optimized levels of processing, storage and network bandwidth can be delivered and scaled independent of one another. These extreme workloads demand an infrastructure that is both flexible and composable. Western Digital is in a leading position to drive NVMf-based SCI technology and market adoption forward, and together with an ecosystem, create environments for data to thrive — capturing, preserving, accessing and transforming it to full potential. When data thrives, our people, communities and planet thrive as well, all through the power, potential and possibilities of data. ■

“Western Digital is in a leading position to drive NVMf-based SCI technology and market adoption forward, and together with an ecosystem, create environments for data to thrive.”

Page 13: The Future of Data Infrastructures The Future of Data ...innovation.wdc.com/downloads/OpenFlex-Tech-Brief.pdfbeyond simply converging or hyper-converging IT resources into ... (as

TECHNICAL BRIEF

Technical Brief | The Future of Data Infrastructures ©2018 Western Digital Corporation or its affiliates. All rights reserved.

About Western Digital

Western Digital creates environments for data to thrive. The company is driving the innovation needed to help customers capture, preserve, access and transform an ever-increasing diversity of data. Everywhere data lives, from advanced data centers to mobile sensors to personal devices, our industry-leading solutions deliver the possibilities of data. Western Digital® data-centric solutions are marketed under the G-Technology™, HGST, SanDisk®, Upthere™ and WD® brands.

Western Digital, the Western Digital logo, G-Technology, SanDisk, Upthere, WD, Kingfish, the Kingfish logo, and OpenFlex are registered trademarks or trademarks of Western Digital Corporation or its affiliates in the U.S. and/or other countries. Hewlett Packard Enterprise is a trademark of HP Hewlett Packard Group LLC. InfiniBand is a trademark of System I/O, Inc. The NVMe and NVMe-oF word marks are trademarks of NVM Express, Inc. Redfish is a registered trademark of the Distributed Management Task Force (DMTF), Inc. All other marks are the property of their respective owners. © 2018 Western Digital Corporation or its affiliates. All rights reserved.

WTC01-EN-US-0818-01

API Application Programming Interface that features a set of subroutine definitions, protocols, tools and clearly defined methods for communicating with various components. In the case of SCIs, storage and other IT resources can be dynamically provisioned and managed through an open API without physically touching hardware. Open means publicly available to developers.

Container A controlling element for an application instance that runs within a type of virtualization scheme called container-based virtualization. The container incorporates the application software, as well as instructions on how to self-provision the required resources for the application using virtualization or an SCI.

HCI Hyper-Converged Infrastructure is a data center architectural framework that combines storage, compute and networking into a single system to reduce data center complexities and increase scalability. These platforms include a hypervisor for virtualized computing, software-defined storage and virtualized networking, and run on standard off-the-shelf servers. Multiple nodes can be clustered together to create pools of shared compute and storage resources.

Noisy Neighbors When data is moved into a public cloud, performance issues may occur from other data being on the same cloud. Public clouds provide multitenant storage, so the sharing of resources also includes competing for the same dedicated servers that reside in the cloud provider’s data center.

NVMe Non-Volatile Memory express is a streamlined, high-performance, low latency interface that utilizes an architecture and set of protocols developed specifically for persistent flash memory technologies.

NVMe-of (aka NVMf and NVMeof)

The NVMe protocol has been extended beyond local-attached server applications and across a network through the NVMe-over-Fabrics specification, and enables flash devices (such as PCIe SSDs and storage arrays) to communicate over networks, such as InfiniBand and Converged Ethernet, delivering the same high-performance, low-latency benefits as local-attached NVMe.

Performance Predictability

In a traditional storage infrastructure, applications compete for resources from a fixed number of IOPS and the devices perform exactly as expected. In an SCI, there is no wasted or unused IOPS as they are available when needed, guaranteeing performance without intervention.

Provisioning The process of assigning storage capacity to servers, computers, virtual machines or any other computing device.

RDMA Remote Direct Memory Access is when memory from one system or computer is provided to another without involving either operating systems, enabling high-throughput, low latency networking. Common RDMA implementations include the Virtual Interface Architecture, RDMA over Converged Ethernet (RoCE), InfiniBand, Omni-Path and internet Wide-Area RDMA Protocol (iWARP).

Redfish API The Distributed Management Task Force (DMTF) Redfish is a standard API designed to deliver simple and secure management for converged, hybrid IT and the Software Defined Data Center (SDCC). Both human-readable and machine-capable, Redfish leverages common internet and web services to expose information throughout the data center.

Key Definitions