the new imperative in the data center with workload centric networking
TRANSCRIPT
#ATM16
The new imperative in the data center with workload-centric networkingSimon Huang Data Center TME managerHP EnterpriseMarch 2016 @ArubaNetworks |
2#ATM16
Agenda
Transformation with Hybrid IT
Data Center Fabric Technologies
Network Design Examples
Fulfill Cloud First Architecture with Data Center portfolio
2
3#ATM16
Transformation with Hybrid IT
4#ATM16
Helping customers transform to a hybrid infrastructureImprove the user experience with accelerated the delivery of apps & services
• Increase availability to gracefully support business growth and unpredictable
• Reduce time to service to enhance business agility
• Enhance customer choice for freedom to build networks that suite business requirements
• Lower total cost of ownership to gain competitive edge in the market
Transformto a hybrid
infrastructure
Enableworkplace
productivity
Protectyour digitalenterprise
Empowerthe data-drivenorganization
5#ATM16
Open Standards Open APIs Open Ecosystem Open Source
Committed to Open Networking in the Data Center
Altoline (OpenSwitch)FlexFabric (Comware)
6#ATM16
Use Gartner Data Center MQ as sales tactic
Gartner rates HPE as most OPEN network vendor“Organizations should consider HPE for any data center network solution, especially those looking for open and standards-based solutions.”
2015 Data Center Networking MQ
7#ATM16
HPE Cloud FirstData Center Fabric Technologies
7
8#ATM16
Types of FabricsLayer 2 or Layer 3 and more
Layer 2 STP and STP variants TRILL / SPB Proprietary protocols (IRF,
Fabricpath, VCS, MLAG) VXLAN (something in
between)
Layer 3 OSPF / BGP Better scalability Predictable path Easy to troubleshoot
9#ATM16
STP Variants
– STP– rPVST+, RSTP, MSTP
– Load-share across the available links on a per VLAN basis
– Root Bridge (RB) election - Who remembers this?– Lowest bridge ID (priority)– If bridge IDs are all the same lowest MAC address– HPE Comware devices reverse this (highest priority is elected RB)
– In PVST the lowest priority will become the root bridge for a VLAN
10#ATM16
HPE Cloud First Network Design Examples
10
11#ATM16
Example 1: Traditional IRF L2 Fabric
• Centralized IRF L2/L3 core switches
• IRF L2 access switches
• Loop free L2 fabric
• STP should still be enabled to prevent any possible loops in future
Bare Metal Servers
L2 Access Switches
IRF L2/L3 Core Switches40G Link
aggregation & 802.1Q trunks
WAN / Internet
L2 External Zone IRF Switches
Load Balancers
Firewalls
WAN Routers
L2 Fabric
13#ATM16
Example 2: L2 Fabric with control plane (SPB/TRILL)
• SPB/TRILL removes STP while still providing a loop free L2 network for east/west traffic• Distributed control plane (No single point of failure)• Architecture Neutral (leaf to leaf or spine leaf)• Lack of control plane interoperability L3 Gateway
Layer Routers
Servers/VMs
10/40G interconnects L2 TRILL / SPB
Fabric
14#ATM16
What is a fabric?– Interconnected network nodes that look like a “fabric” when viewed collectively from a distance
– Spine leaf/ CLOS /Two-Tier /Ethernet Fabric /Network Fabric are interchangeable and commonly used terms
– For most it’s a scalable, high performing, resilient network architecture.
15#ATM16
Example 3: CLOS Fabric• CLOS (physical) network architecture provide edge/core multi-tier design• Each leaf switch is connected to all spine switches• Customers may choose to deploy a 2 spine fabric (2 x 40G uplinks) and expand to 4+ spines (4 x
40G uplinks or more) when they require additional bandwidth • Protocol independent (STP/TRILL/SPB/L3) over the physical fabric
Leaf Switches
Spine Switches
4 Spine CLOS Fabric2 Spine CLOS Fabric
Spine Switches
Leaf Switches
16#ATM16
Example 4: L3 Fabric
• Removal of STP• Usage of standards based routing protocol• Shortest path routing• Provides Equal Cost Multi Pathing (ECMP)• Traffic engineering
Servers/VMs
Default gateway for each subnet
terminates at access switch
Multiple /30 L3 subnets
L2
Routing protocol
L3 EMCP routes
10/40G interconnects
17#ATM16
Example 5: Network Virtualization Enabled Fabric• Network Virtualization enabled fabric provides virtual to physical networking bridging capabilities• Underlay network fabric can be deployed either as L2 / L3• Most devices do not need to be VXLAN Tunnel End Points (VTEP) or VXLAN capable (e.g. SPINEs
only need to route traffic between VTEPs) • L3 fabric recommended if network virtualization software supports different subnets between VTEPs• Hardware VTEPs used to bridge between virtual and physical networks
Unicast Layer 3 Underlay Network
VXLAN
Software VTEP
Physical Servers, WAN Routers/ Firewalls
172.16.2.10/24
VM2: 172.16.2.2/24
HPE Hardware
VTEPs
Network Virtualization Controller
OVSDB / NETCONF
18#ATM16
Example 6: L2 over L3 VXLAN overlay Fabric
• Loop free L2 fabric without STP or network virtualization controllers
• VXLAN overlay tunnels provides link load sharing across the physical fabric
• High performance multicast forwarding in the underlay (compared to multicast forwarding using the overlay)
Servers/VMs
Leaf/EoR Layer (L2 VTEPs)
Spine/Core Layer (L2 VTEPs)
L2 Fabric
Access/ToR Layer
L3 Gateway Layer
POD 1 POD 2 POD 3 POD 4 POD …
Overlay VXLAN tunnels
Underlay L3 Network
19#ATM16
Example 7: Multiple Fabrics• Network fabrics can be deployed as different availability zones to limit failure domains• L2 network connectivity between VMs/Servers are provided within an availability zone• L3 used between different availability zones • Border leaf switches used to connect out of an availability zone
Border LeafIGP: OSPF
EGP: BGP AS#65001
IGP: OSPFEGP: BGP AS#65002
Availability Zone#1 (10.1.0.0/16)
Availability Zone#2 (10.2.0.0/16)
Core Network: BGP AS#65100
Border Leaf
Example with multiple availability zones
L3 L3
20#ATM16
Example 8: Inter Data Center Fabric Connectivity• VXLAN and HPE EVI technologies can be combined for active/active inter-DC L2 DCI over L3
WAN• EVI provides additional benefits such as:
− Active/Active VRRP default gateways for VMs− STP failure domain local to each DC− Improves WAN utilization by dropping unknown frames and providing ARP suppression− Optimized Server/Client Traffic Paths
• VXLAN based DCI must avoid full networking “interconnect”
Virtual Overlay VXLAN tunnels
Physical Underlay Network
Active Data Center (DC) 1
L2 or L3
Virtual Overlay VXLAN tunnels
Physical Underlay Network
Active Data Center (DC) 2
L2 or L3WAN L1/L2
network
EVI tunnel
21#ATM16
IP Fabric
Example 9: Storage & Converged Fabric• When storage connectivity is required, converged both FC/Ethernet or FCoE/IP with universal
access• Spilt out to separate IP/SAN fabrics from leaf/ToR switches:
IRF
FCoE Storage
VN_PortVN_Port
VF_Port
FCoE Storage
5900CP FCF
IRF
FC Storage
N_PortN_Port
F_Port
FC Storage
5900CP FCF
Server/CNA
Server/CNA
Server/CNA
Server/HBA
Native FC Ethernet FCoE
IP Fabric
SAN Fabric A
SAN Fabric B
SAN Fabric A
SAN Fabric B
VF_Port
F_Port
22#ATM16
Which Fabric deployment to choose?• It depends on customer requirements!
• L2 between access/core layer in the DC: Used in the DC mainly because it L2 connectivity is required by the applications across racks, L2 loops are typically more difficult to troubleshoot compared to L3
• Traditional IRF L2 fabric works and is mature
• L3 between access/core layer in the DC: Main benefit is STP is no longer required to prevent loops, L3 routing protocol is used, easier to troubleshoot for network OPs
• Decision for L2/L3 between access/core layer in the DC really depends on application network requirements.
• VXLAN over L3 fabric in the DC provides L3 fabric benefits while providing L2 network connectivity for applications
• Is there a requirement to support storage traffic FC/FCoE and/or iSCSI?
23#ATM16
DC Fabrics
Multiple options
Layer 2 Layer 3
STP (MSTP, rPVST+)
IRF
TRILL/SPB-m
OSPF
BGP
VXLAN
• HPE has comprehensive fabric capabilities
• HPE offers standard based “open” fabric
• Final design depends on customer requirements
• Typical products:− Spine = 7900/ 12900− Leaf = 5900/ 5930/ 7900 / 12900
24#ATM16
Guidance currently in FFRA
• DC Architecture Guidance
25#ATM16
Fulfill the Architecture with Data Center portfolio
25
26#ATM16
HPE Data Center – Built to Win
High Performance DC Fabric
High Density / High Performance / Highly Scalable / Highly Resilient
Composable Infrastructure
“Instantiating open, complex networks and associated policies in minutes vs. weeks”
Zero Touch Provisioning- DC Fabric
IMC Platform (ZTP / Fabric Manager)DevOps (Python, Ansible …)
Hybrid Cloud Integration Helion/Openstack/CSA integration
27#ATM16
HPE Data Center FlexFabric for Spine/Leaf Deployment
Modular network OS with Intelligent Resilient Fabric
1/10/40GbE L2/L3 and converged switches25GbE/100GbE Roadmap
HP
E IM
C M
anag
emen
t Core switchesSpine
HPE Comware Network OS
L2/L3 IPv4/v6MPLS/VPLS
VXLAN
Top of Rack (TOR)
Leaf switches
High Density 10/40/100GbE switches
HP
E T
echn
olog
y S
ervi
ces
HP
E C
onsu
lting
Ser
vice
s
SDN
28#ATM16
HPE Insertion Point
10G/40G Spine10/40G Top-of-Rack/Leaf High Density 100GbE Spine
Lead with HPE FlexFabric Modular 5930LAN/SAN/MACSEC/VXLAN Ready
HPE FlexFabric 7910/7904The power of the 129xx
@ smaller scale
HPE FlexFabric 12900EWhen density is everything
29#ATM16
Data Center Core SwitchingFlexFabric 129xxE FlexFabric 79xxLarge/Very Large Core Compact Modular Core
Same switching architecture (CLOS/VoQ)Spine/Leaf (deployment architecture)
HPE FlexFabric Core Switches At-a-Glance
All Platforms SDN Ready with OF 1.3 support and IPv6 Ready Phase2
30#ATM16
HPE Modular 5930
Universal PortsLAN/SAN convergence 1/10GbE and 4/8G FC
OpenFlow 1.3SDN automation and central management
L2 encryption (MACSEC)End-to-End Security
Maximum performance
Quick set up and upgrades
Lower energy and utility costs
Investment protection
VXLAN in hardwareOverlay Ready − HPE DCN/NSXNSX Certification soonCloud enabled Comware 7
HPE 5930-2SLOT-2QSFP ChassisHPE 5930-4SLOT Chassis Hot Swappable Modules
31#ATM16
HPE FlexFabric Access Switches At-a-Glance
FlexFabric Top-of Rack, Access Blade Switches
PortfolioModular
5930 & 5930
5900 5900CP 5700 6127XLG Moonshot-45XGc
Highlights40G ToR
Converged LAN/SAN
MACSEC (HW)
10/40GbE ToR Converged LAN/SAN ToR 1/10GbE ToR
HPE BladeSystemInterconnect
Comware based
HPE BladeSystemInterconnect
Comware based
32#ATM16
Capturing the transition from 10 to 25 Gb/s
SYS
FlexFabric5930Se ries
PN JG996A SN AFK 6740
3.2 Tbps ASIC – 32x100G QSFP28
SYS
FlexFabric5950Series
2.0 Tbps ASIC – 48x25G SFP28 and 8x100G QSFP28 Switch
PN JG996A SN AFK6740
SYS
FlexFabric5930Series
PN JG996A SN AFK6740 PN JG996A SN AFK6740
3.2 Tbps ASIC – 4xSlot
2016 – Server transitioning from 10 Gb/S driving an inflection point for 100 GbE in Data Center networking
Leaf : XGS/Tomahawk5950 (32 * 100G, 48 * 25G + 8 * 100G) – 1RUModular 5950 (10/25/40/100G + 16/32G Converged Ports) – 2RU
Spine: ARAD+/Jericho/Jericho+TCAM12902E/12904E/12908E/12916E (3/6/12/21RU)
33#ATM16
IMC Value PropositionThe foundation of a complete DC management, monitoring and orchestration solution
Lead with HPE Intelligent Management Center Platform
FCAPS Ready
Consistent visibility
Actionable data
Flexible deployment Remediation
Accessible management
Security driven
34#ATM16
Orchestrate network fabrics
Complete the SDN architecture with management
Accelerates deployment of services and applications
•Unified SPB, TRILL, IRF fabric management
•Manages across geo dispersed locations (DCI/EVI)
•VMware vMotion playback•Unified DCB, FCoE mgt
•Configuration, monitoring & policy mgt for all SDN layers
•OpenFlow switch management•SDN controller performance management
•One application for managing SDN and traditional environ.
• “Just right” network services tuned to business needs
•Simplifies provisioning, monitoring of applications
•Eliminates manual provisioning
•Easy drag and drop UI
IMC Orchestration for Data Center
35#ATM16
Industry leading energy efficiency
Built the cloud on HPE Converged InfrastructureA customizable RA built on our industry-leading infrastructure portfolio
35
The data center standard delivering optimal performance, expandability, and availability for your workload
HPE ProLiant Gen9 servers
Affordable, intelligent, and resilient flash with unmatched performance, flexibility, and ease of management
HPE 3PAR StoreServ storage
High-density, ultra-low-latency, top-of-rack (ToR) switches designed to deliver optimal throughput and availability
HPE FlexFabric Switch series
Powered by Intel® Xeon ® Processor E5-2600 v3 product family
“Future proof” design
Agile infrastructure management
All-flash at nearly the same cost as HDDs
Industry-leading scale and density
Architected for 99.9999% availability
Simplified management for lower TCO
Low latency, high-performance data switching
Business resilience and high availability
36#ATM16
HPE Data Center Switching Positioning
• Integrated, fully featured• Enterprise data centers• Comware
• Disaggregated• Web-scale data centers• Choice of NOS
HPE FlexFabric HPE Altoline
Worldwide SDN market growing to over $8 billion by 2018
37#ATM16
HPE Network Virtualization SolutionsIndustry’s most complete portfolio - addressing varied/complex use cases
37
Virtual Cloud Network HPE-VMware NSX Distributed Cloud Network
Enhanced Neutron Networking
Open source, DevOps Virtualized VMW Enterprises Service Provider/Telcos
De-risk the journey with HPE Trusted Network Transformation Services
DCNVCN
Multi DC NFV
Powered by HPE
38#ATM16
39#ATM16
Join Aruba’s Titans of Tomorrow force in the fight against network mayhem. Find out what your IT superpower is.
Share your results with friends and receive a free superpower t-shirt.
www.arubatitans.com
Thank you