+ All Categories
Home > Technology > Extreme Spine Leaf Design

Extreme Spine Leaf Design

Date post: 12-Apr-2017
Category:
Upload: jeff-green
View: 116 times
Download: 0 times
Share this document with a friend
1
Multi-Rate 1,2.5,5,10 Gigabit Edge PoE++ Multi-Rate Spine Leaf Design (10,25,40,50,100 Gigabit) X440-G2 (L3 - Value 1G to 10G) PoE Fiber DC Policy SummitStack-V (WITHOUT any additional license required). Upgradeable 10GbE (PN 16542 or 16543). Policy built-in (simplicity with multi-auth). EXOS 21.1 or higher Value with Automation First Extreme Switch to support Cloud Value X460-G2 (Advanced L3 1-40G) Multirate Option PoE Fiber DC Policy Fit The Swiss Army Knife of Switches Half Duplex ½ & ½ 3 Models This is where: 10G on existing copper Cat5e and Cat6 extend the life of the installed cable plant. Great for 1:N Convergence. X620 (1OG Copper or Fiber) Speed Next Gen Edge Lowered TCO via Limited Lifetime Warrantee Wallplate AP AP + Camera Outdoor Wave 2 Multi-Gigabit Wireless High Density -pack or Wedge Facebook Extreme Support XoS Platform Config L2/L3 Analytics Any OS Any Bare Metal Switch Policy Disaggregated Switch CAPEX or OPEX (you choose)? Reduced Risk (just witness or take action) Time is the critical Factor with XYZ Account Services... Infrastructure Business model Ownership Considerations Management Location 32 x 100Gb 64 x 50Gb 128 x 25Gb 128 x 10Gb 32 x 40Gb 96 x 10GbE Ports (via4x10Gb breakout) 8 x 10/25/40/ 50/100G 10G Next Gen: Spine Leaf X670 & X770 - Hyper Ethernet Common Features Data Center Bridging (DCB) features Low ~600 nsec chipset latency in cut through mode. Same PSUs and Fans as X670s (Front to back or Back to Front) AC or DC. X670-G2 -72X (10GbE Spine Leaf) 72 10GbE X670-48x-4q (10GbE Spine Leaf) 48 10GbE & 4 QSFP+ QSFP+ 40G DAC Extreme Feature Packs Core Edge AVB OpenFlow Advance Edge 1588 PTP MPLS Direct Attach Optics License Extreme Switches include the license they normally need. Like any other software platform you have an upgrade path. QSPF28 100G DAC Thin & Crunchy XoS Platform with one track of software. Speed with Features (Simple). Metro Functionality like ATM or SONET Flexible Horizontal or Vertical stacking Purposed for Broadcom (ASICs) So What, Who cares? Deliver XYZ Account, the value of HP with the feature function of Cisco. XYZ Account Business Value Why Extreme? Summit Summit Policy delivers automation.. Thick & Chewy Know and control the who, what, when, where and the user experience across your XYZ Account Network. Control with insight... Why Enterasys? XYZ Account Strategic Asset Custom ASICs S & K Series Chantry Motorola Air Defense So What, Who cares? Flow Based Switching Simplicity w Policy Wired and Wireless 100% insourced support Today you get both Control So What, Who cares? Fit Speed Unique Value Unique Control Summit G2 Yesterday - Cabletron Changed the game w Structured wiring (remember Vampire taps, Coax ethernet ect.) Today - Extreme Delivers Structured networking Policy Summit Who? Where? When? What device? How? Quarantine Remediate Allow Authentication NAC Server Summit Netsite Advanced NAC Client Joe Smith XYZ Account Access Controlled Subnet Enforcement Point Network Access Control This is where if X + Y, then Z... LLDP-MED CDPv2 ELRP ZTP If user matches a defined attribute value ACL QoS Then place user into a defined ROLE A port is what it is because? This is where you easily Identify the impact and Source of Interference Problems. Detailed Forensic Analysis Device, Threats, Associations, Traffic, Signal and Location Trends Record of Wireless Issues Network Trend Analysis Historical Analysis of Intermittent Wireless Problems Performance Trends a Spectrum Analysis for Interference Detection Real-time Spectrograms Proactive Detection of Application Impacting Interference Visualize RF Coverage Real-time RF Visualizations Proactive Monitoring and Alerting of Coverage Problem ADSP for faster Root Cause Forensic Analysis for SECURITY & COMPLIANCE. Event Sequence Classify Interference Sources Side-by-side Comparative Analysis Air Defense Application Experience Full Context App App Analytics App Stop the finger-pointing Application Network Response. Flow or Bit Bucket Collector 3 million Flows Sensors X460 IPFix 4000 Flows (2048 ingress, 2048 egress) Sensor PV-FC-180, S or K Series (Core Flow 2/ 1 Million Flows) Flow-based Access Points From the controller (8K Flows per AP or C35 is 24K Flows) Flows Why not do this in the network? 10110111011101110 101101110111011101 6 million Flows Business Value Context BW IP HTTP:// Apps Platform Automation Control Experience Solution Framework Is your network faster today than it was 3 years ago? Going forward it should deliver more, faster, different X430-G2 (L2 - 1G to 10G) PoE Distribute content from a single source to hundreds of displays Ethernet as a Utility (PoE) Injectors Up to 75 Watts XYZ Account Data Center Chassis V Spline Fabric Modules (Spine) I/O Modules (Leaf) Spine Leaf Proven value with legacy approach. Can not access Line cards. No L2/l3 recovery inside. No access to Fabric. Disaggregated value... Control Top-of-Rack Switches L2/L3 protocols inside the Spline Full access to Spine Switches No EGO, Complexity or Vendor Lock-in). Fat-Tree Clos / Cross-Bar Traditional 3-tier model (Less cabling). Link speeds must increase at every hop (Less predictable latency). Common in Chassis based architectures (Optimized for North/South traffic). Every Leaf is connected to every Spine (Efficient utilization/ Very predictable latency). Always two hops to any leaf (More resiliency, flexibility and performance). Friendlier to east/west traffic (The uplink to the rest of the network is just another leaf). The XYZ Account handshake layer : This is where convergence needs to happen – LAN/SAN, FCoE, ETS. Stop or allow whatever you can (Efficient Multicasting). Virtualization happens with VXLAN and VMotion (Control by the overlay). N plus one fabric design needs to happen here (Delivers simple no vanity future proofing, No-forklift migrations, interop between vendors and hit-less operation). This is where, a Fabric outperforms the Big Uglies ONE to ONE: Spine Leaf The XYZ Account Ethernet Expressway Layer: deliver massive scale... This is where low latency is critical, switch as quickly as you can. DO NOT slow down the core keep it simple (Disaggregated Spline + One Big Ugly Elastic Capacity - Todays XYZ Accounts spines are tomorrows leafs. Dial-in the bandwidth to your specific needs with the number of uplinks. Availability - the state of the network is kept in each switch; no single point of failure. Seamless XYZ Account upgrades, easy to take a single switch out of service. (Cloud Fabric) Disaggregation Spine Leaf Legacy Challenges: Complex/Slow/Expensive Scale-up and Scale out Vendor lock-in Proprietary (HW, SW) Commodity Fabric Modules (Spine) I/O Modules (Leaf) Spline (Speed) Active - Active redundancy fn(x,y,z) The next convergence will be collapsing datacenter designs into smaller, elastic form factors for compute, storage and networking. This is where, you can never have enough. Customers want scale made easy. Hypervisor integration w cloud simplicity. L2 L3 L2 L3 L2 L3 L2 L3 L2 L3 Start Small; Scale as You Grow This is where, you can simply add a Extreme Leaf Clusters Each cluster is independent (including servers, storage, database & interconnects). Each cluster can be used for a different type of service. Delivers repeatable design which can be added as a commodity. XYZ Account Spine Leaf Cluster Cluster Cluster Egress Scale Ingress Active / Active VM VM VM RR RR BGP Route-Reflector RR iBGP Adjacency This is where VXLAN (Route Distribution) This is where Why VxLAN? It Flattens network to a single tier from the XYZ Account end station perspective. All IP/BGP based (Virtual eXtensible Local Area Network). Host Route Distribution decoupled from the Underlay protocol. VXLANs goal is allowing dynamic large scale isolated virtual L2 networks to be created for virtualized and multi- tenant environments. Route-Reflectors deployed for scaling purposes - Easy setup, small configuration. Traffic Engineer “like ATM or MPLS” UDP Start Stop UDP UDP Use Existing IP Network VM VM VM VM VM VM VM VM VTEP VTEP VM VM VM VM VM VM VM VM VM VM VM VM VM VM VM App 1 App 2 App 3 Intel, Facebook, OCP Facebook 4-Post Architecture - Each leaf or rack switch has up to 48 10G downlinks. Segmentation or multi-tenancy without routers. Each spine has 4 uplinks – one to each leaf (4:1 oversubscription). Enable insertion of services without sprawl (Analytics for fabric and application forensics). No routers at spine. One failure reduces cluster capacity to 75%. (5 S's) Needs to be Scalable, Secure, Shared, Standardized, and Simplified. Network (Fit) Overlay Control The XYZ Account the VxLan forwarding plane for NSX control: This is where logical switches span across physical hosts and network switches. Application continuity is delivered with scale. Scalable Multi-tenancy across data center. Enabling L2 over L3 Infrastructure - Pool resources from multiple data centers with the ability to recover from disasters faster. Address Network Sprawl with an VXLAN overlay. Deeper Integration with infrastructure and operations partners, integrations, and frameworks for IT organizations. Vmware NSX (Control Plane) Management Plane deliver by the NSX Manager. Control Plane NSX Controller Manages Logical networks and data plane resources. Extreme delivers an open high performance data plane with Scale NSX Architecture and Components CORE CAMPUS 10Gb Aggregation High Density 10Gb Aggregation 10Gb/40Gb Aggregation High Density 25Gb/50Gb Aggregation 100Gb Uplinks 100Gb Uplinks Server PODs 770 / 870 Spine Data Center Private Cloud vC-1 vC-2 vC-N This is where XYZ Account must first it must have the ability to scale with customer demand, delivering more than just disk space and processors. Scale – XYZ Account must have be able the to seamlessly failover, scale up, scaled down and optimize management of the applications and services. Flexibility - The infrastructure XYZ Account must have the ability to host heterogeneous and interoperable technologies. Business - The business model costs might be optimized for operating expenses or towards capital investment. Cloud Computing (Control Plane) (On-Premise) Infrastructure (as a Service) Platform (as a Service) Storage Servers Networking O/S Middleware Virtualization Data Applications Runtime Storage Servers Networking O/S Middleware Virtualization Data Applications Runtime You manage Managed by vendor Managed by vendor You manage You manage Storage Servers Networking O/S Middleware Virtualization Applications Runtime Data Software (as a Service) Managed by vendor Storage Servers Networking O/S Middleware Virtualization Applications Runtime Data Public Private MSP F A B R I C This is where Azure ExpressRoute lets XYZ Account create private connections between Azure datacenters and XYZ Account infrastructure on or off premises. ExpressRoute connections don't go over the public Internet. They offer more reliability, faster speeds, and lower latencies, and higher security than typical Internet connections. XYZ Account can transfer data between on-premises systems and Azure can yield significant cost benefits. XYZ Account can establishing connections to Azure at an ExpressRoute location, such as an Exchange provider facility, or directly connect to Azure from your existing WAN network, such as a multi-protocol label switching (MPLS) VPN, provided by a network service provider Microsoft Assure (Control Plane) Cloud The key impact of this model for the customer is a move from managing physical servers to focus on logical management of data storage through policies. Compute Storage Data Center Architecture Considerations Compute Cache Database Storage Client Response 80% North-South Traffic Oversubscription : upto 200:1 (Client Request +Server Response = 20% traffic). Inter-rack latency: 150 micros. Lookup Storage = 80% traffic. Scale: Up to 20 racks (features Non- blocking 2 tier designs optimal). VM VM VM VM Purchase "vanity free" This is where.. Open Compute might allow companies to purchase "vanity free". Previous outdated data center designs support more monolithic computing. Low density X620 might help XYZ Account to avoid stranded ports. Availability - Dual X620s can be deployed to minimize impact to maintenance. Flexibility of the X620 can offer flexibility to support both 1G and 10G to servers and storage. One RACK Design Closely coupled Nearly coupled Loosely coupled Shared Combo Ports 4x10GBASE-T & 4xSFP+ 100Mb/1Gb/10GBASE-T The monolithic datacenter is dead. Servers Storage Summit Management Switch Summit Summit Storage Management Servers Open Compute - Two Rack Design This is where, XYZ Account can reduce OPEX and leverage a repeatable solution. With the spline setup, XYZ Account can put redundant switches in the middle and link each server to those switches. Fewer Hops between Servers - The important thing is that each server is precisely one hop from any other server. Avoid Stranded ports – Designs often have a mix of fat and skinny nodes. If XYZ Account deploys a 48-port leaf switches many configurations might have anywhere from 16 to 24 stranded ports. Two RACK Servers Storage Summit Management Switch Summit Summit Storage Management Servers Storage Summit Management Switch Summit Summit Typical spline setup Open Compute : Eight Rack POD Design This is where Typical spline setup : Eight Rack POD Leaf Spine Servers Storage Summit Management Switch Summit Summit Storage Servers Storage Summit Management Switch Summit Summit Storage Servers Storage Summit Management Switch Summit Summit Storage Servers Storage Summit Management Switch Summit Summit Storage Servers Storage Summit Management Switch Summit Summit Storage Servers Servers Storage Summit Management Switch Summit Summit Storage
Transcript
  • Mult

    i-R

    ate

    1,2

    .5,5

    ,10 G

    iga

    bit

    Ed

    ge

    PoE

    ++

    Mult

    i-R

    ate

    Spi

    ne L

    ea

    f D

    esi

    gn (

    10,2

    5,4

    0,5

    0,1

    00 G

    iga

    bit

    )

    X440-G2 (L3 - Value 1G to 10G)

    PoE

    Fiber

    DC

    Policy

    SummitStack-V (WITHOUT any

    additional license required).

    Upgradeable 10GbE (PN 16542 or 16543).

    Policy built-in (simplicity with multi-auth).

    EXOS 21.1 or

    higher

    Value with Automation

    First Extreme

    Switch to support

    Cloud Value

    X460-G2 (Advanced L3 1-40G) Multirate Option

    PoE

    Fiber

    DC

    Policy

    Fit The Swiss Army Knife of Switches Half Duplex

    & 3 Models

    This is where: 10G on

    existing copper Cat5e

    and Cat6 extend the

    life of the installed

    cable plant. Great for

    1:N Convergence.

    X620 (1OG Copper or Fiber)

    Speed Next Gen Edge

    Lowered TCO via

    Limited Lifetime Warrantee

    Wallplate AP

    AP + Camera

    Outdoor Wave 2

    Multi-Gigabit

    Wireless

    High Density

    -pack or Wedge

    Facebook

    Ex

    tre

    me

    Su

    pport

    XoS

    Platform

    Config L2/L3

    Analytics

    Any OS

    Any Bare Metal Switch

    Policy

    Disaggregated Switch

    CAPEX or OPEX (you choose)?

    Reduced Risk (just witness or take action)

    Time is the critical Factor with XYZ Account Services...

    Infrastructure

    Business model

    Ownership

    Considerations

    Management

    Location

    32 x 100Gb

    64 x 50Gb

    128 x 25Gb

    128 x 10Gb

    32 x 40Gb

    96 x 10GbE Ports

    (via4x10Gb breakout)8 x 10/25/40/

    50/100G

    10G

    Next Gen: Spine Leaf

    X670 & X770 - Hyper Ethernet

    Common Features

    Data Center Bridging (DCB) features

    Low ~600 nsec chipset latency in cut through mode.

    Same PSUs and Fans as X670s (Front to back or Back to

    Front) AC or DC.

    X670-G2 -72X (10GbE Spine Leaf) 72 10GbE

    X670-48x-4q (10GbE Spine Leaf) 48 10GbE & 4 QSFP+

    QSFP+

    40G DAC

    Extreme Feature Packs

    Core

    Edge

    AVB

    OpenFlow

    Advance

    Edge

    1588 PTP

    MPLS

    Direct Attach

    Optics License

    Extreme Switches

    include the license

    they normally need.

    Like any other

    software platform

    you have an

    upgrade path.

    QSPF28

    100G DAC

    Thin & Crunchy

    XoS Platform with one track of software.

    Speed with Features (Simple).

    Metro Functionality like ATM or SONET

    Flexible Horizontal or Vertical stacking

    Purposed for Broadcom

    (ASICs)

    So What, Who cares?

    Deliver XYZ Account, the

    value of HP with the feature

    function of Cisco.

    XYZ Account Business Value

    Wh

    y E

    xtre

    me

    ?

    Summit

    Summit

    Policy delivers automation..

    Thick & Chewy

    Know and control

    the who, what, when, where and the user

    experience across your XYZ Account

    Network.

    Control with insight...

    Wh

    y E

    ntera

    sys?

    XYZ Account Strategic Asset

    Custom ASICs

    S & K Series

    Chantry

    Motorola

    Air

    Defense

    So What, Who cares?

    Flow Based Switching

    Simplicity w Policy

    Wired and Wireless

    100% insourced support

    Today you get both

    Control

    So What, Who cares?

    Fit

    Speed

    Unique

    Value

    Unique

    Control

    Summit G2

    Yesterday - Cabletron Changed the game w Structured wiring

    (remember Vampire taps, Coax ethernet ect.)

    Today - Extreme Delivers Structured networking

    Policy

    Summit

    Who?

    Where?

    When?

    What device?

    How?

    QuarantineRemediateAllow

    Authentication

    NAC Server

    Summit

    Netsite

    Advanced

    NAC Client

    Joe Smith

    XYZ Account

    Access

    Controlled

    Subnet

    Enforcement

    Point

    Network

    Access

    Control

    This is where

    if X + Y, then Z...

    LLDP-MED

    CDPv2

    ELRP

    ZTP

    If user

    matches a

    defined

    attribute

    value

    ACL

    QoS

    Then place

    user into a

    defined ROLE

    A port is what it is because?This is where you easily Identify

    the impact and Source of

    Interference Problems.

    Detailed Forensic Analysis

    Device, Threats, Associations,

    Traffic, Signal and Location

    Trends

    Record of Wireless Issues

    Network Trend Analysis

    Historical Analysis of

    Intermittent Wireless

    Problems

    Performance Trends a

    Spectrum Analysis for

    Interference Detection

    Real-time Spectrograms

    Proactive Detection of

    Application Impacting

    Interference

    Visualize RF Coverage

    Real-time RF Visualizations

    Proactive Monitoring and

    Alerting of Coverage Problem

    ADSP for faster Root Cause Forensic

    Analysis for SECURITY & COMPLIANCE.

    Event

    Sequence

    Classify

    Interference

    Sources

    Side-by-side

    Comparative

    Analysis

    Air Defense

    App

    lic

    ati

    on

    Expe

    rie

    nc

    e

    Fu

    ll C

    onte

    xt

    App

    App

    Analytics

    App

    Stop the

    finger-pointing

    Application Network Response.

    Flow or Bit

    Bucket

    Collector

    3 million Flows

    Sensors

    X460 IPFix 4000 Flows

    (2048 ingress, 2048 egress)

    Sensor PV-FC-180, S or K Series (Core

    Flow 2/ 1 Million Flows)

    Flow-based Access Points

    From the controller (8K Flows

    per AP or C35 is 24K Flows)

    Flows

    Why not do this in the

    network?

    10110111011101110 101101110111011101

    6 million Flows

    Business ValueContext BW IP HTTP:// Apps

    Platform Automation Control Experience Solution Framework

    Is your network faster today than

    it was 3 years ago? Going forward

    it should deliver more, faster,

    different

    X430-G2 (L2 - 1G to 10G)

    PoE

    Distribute content

    from a single source

    to hundreds of displays

    Ethernet as a Utility

    (PoE)

    Injectors

    Up to 75

    Watts

    XYZ Account Data CenterXYZ Account Data Center

    Chassis V Spline

    Fabric Modules (Spine)

    I/O Modules (Leaf)

    Spine

    Leaf

    Proven value with legacy approach.

    Can not access Line cards.

    No L2/l3 recovery inside.

    No access to Fabric.

    Disaggregated value...

    Control Top-of-Rack Switches

    L2/L3 protocols inside the Spline

    Full access to Spine Switches

    Chassis V Spline

    Fabric Modules (Spine)

    I/O Modules (Leaf)

    Spine

    Leaf

    Proven value with legacy approach.

    Can not access Line cards.

    No L2/l3 recovery inside.

    No access to Fabric.

    Disaggregated value...

    Control Top-of-Rack Switches

    L2/L3 protocols inside the Spline

    Full access to Spine Switches

    No EGO, Complexity or Vendor Lock-in).

    Fat-Tree

    Clos / Cross-Bar

    Traditional 3-tier model (Less cabling).

    Link speeds must increase at every hop (Less

    predictable latency).

    Common in Chassis based architectures (Optimized

    for North/South traffic).

    Every Leaf is connected to every Spine (Efficient

    utilization/ Very predictable latency).

    Always two hops to any leaf (More resiliency,

    flexibility and performance).

    Friendlier to east/west traffic (The uplink to the

    rest of the network is just another leaf).

    No EGO, Complexity or Vendor Lock-in).

    Fat-Tree

    Clos / Cross-Bar

    Traditional 3-tier model (Less cabling).

    Link speeds must increase at every hop (Less

    predictable latency).

    Common in Chassis based architectures (Optimized

    for North/South traffic).

    Every Leaf is connected to every Spine (Efficient

    utilization/ Very predictable latency).

    Always two hops to any leaf (More resiliency,

    flexibility and performance).

    Friendlier to east/west traffic (The uplink to the

    rest of the network is just another leaf).

    The XYZ Account handshake layer:

    This is where convergence needs to happen LAN/SAN, FCoE, ETS. Stop or allow

    whatever you can (Efficient Multicasting).

    Virtualization happens with VXLAN and VMotion (Control by the overlay).

    N plus one fabric design needs to happen here (Delivers simple no vanity future proofing,

    No-forklift migrations, interop between vendors and hit-less operation).

    This is where,

    a Fabric outperforms the Big Uglies

    ONE to ONE: Spine Leaf

    The XYZ Account handshake layer:

    This is where convergence needs to happen LAN/SAN, FCoE, ETS. Stop or allow

    whatever you can (Efficient Multicasting).

    Virtualization happens with VXLAN and VMotion (Control by the overlay).

    N plus one fabric design needs to happen here (Delivers simple no vanity future proofing,

    No-forklift migrations, interop between vendors and hit-less operation).

    This is where,

    a Fabric outperforms the Big Uglies

    ONE to ONE: Spine Leaf

    The XYZ Account Ethernet Expressway Layer: deliver massive scale...

    This is where low latency is critical, switch as quickly as you can. DO NOT slow down

    the core keep it simple (Disaggregated Spline + One Big Ugly

    Elastic Capacity - Today s XYZ Account s spines are tomorrow s leafs. Dial-in the

    bandwidth to your specific needs with the number of uplinks.

    Availability - the state of the network is kept in each switch; no single point of failure.

    Seamless XYZ Account upgrades, easy to take a single switch out of service.

    (Cloud Fabric) Disaggregation

    Spine

    Leaf

    Legacy Challenges:

    Complex/Slow/Expensive

    Scale-up and Scale out

    Vendor lock-in

    Proprietary (HW, SW)Commodity

    Fabric Modules (Spine)

    I/O M

    odules (Lea

    f)

    Fabric Modules (Spine)

    I/O M

    odules (Lea

    f)

    The XYZ Account Ethernet Expressway Layer: deliver massive scale...

    This is where low latency is critical, switch as quickly as you can. DO NOT slow down

    the core keep it simple (Disaggregated Spline + One Big Ugly

    Elastic Capacity - Today s XYZ Account s spines are tomorrow s leafs. Dial-in the

    bandwidth to your specific needs with the number of uplinks.

    Availability - the state of the network is kept in each switch; no single point of failure.

    Seamless XYZ Account upgrades, easy to take a single switch out of service.

    (Cloud Fabric) Disaggregation

    Spine

    Leaf

    Legacy Challenges:

    Complex/Slow/Expensive

    Scale-up and Scale out

    Vendor lock-in

    Proprietary (HW, SW)Commodity

    Fabric Modules (Spine)

    I/O M

    odules (Lea

    f)

    Spline (Speed)

    Chassis V Spline

    Fabric Modules (Spine)

    I/O Modules (Leaf)

    Spine

    Leaf

    Proven value with legacy approach.

    Can not access Line cards.

    No L2/l3 recovery inside.

    No access to Fabric.

    Disaggregated value...

    Control Top-of-Rack Switches

    L2/L3 protocols inside the Spline

    Full access to Spine Switches

    No EGO, Complexity or Vendor Lock-in).

    Fat-Tree

    Clos / Cross-Bar

    Traditional 3-tier model (Less cabling).

    Link speeds must increase at every hop (Less

    predictable latency).

    Common in Chassis based architectures (Optimized

    for North/South traffic).

    Every Leaf is connected to every Spine (Efficient

    utilization/ Very predictable latency).

    Always two hops to any leaf (More resiliency,

    flexibility and performance).

    Friendlier to east/west traffic (The uplink to the

    rest of the network is just another leaf).

    The XYZ Account handshake layer:

    This is where convergence needs to happen LAN/SAN, FCoE, ETS. Stop or allow

    whatever you can (Efficient Multicasting).

    Virtualization happens with VXLAN and VMotion (Control by the overlay).

    N plus one fabric design needs to happen here (Delivers simple no vanity future proofing,

    No-forklift migrations, interop between vendors and hit-less operation).

    This is where,

    a Fabric outperforms the Big Uglies

    ONE to ONE: Spine Leaf

    The XYZ Account Ethernet Expressway Layer: deliver massive scale...

    This is where low latency is critical, switch as quickly as you can. DO NOT slow down

    the core keep it simple (Disaggregated Spline + One Big Ugly

    Elastic Capacity - Today s XYZ Account s spines are tomorrow s leafs. Dial-in the

    bandwidth to your specific needs with the number of uplinks.

    Availability - the state of the network is kept in each switch; no single point of failure.

    Seamless XYZ Account upgrades, easy to take a single switch out of service.

    (Cloud Fabric) Disaggregation

    Spine

    Leaf

    Legacy Challenges:

    Complex/Slow/Expensive

    Scale-up and Scale out

    Vendor lock-in

    Proprietary (HW, SW)Commodity

    Fabric Modules (Spine)

    I/O M

    odules (Lea

    f)

    Spline (Speed)

    Active - Active redundancy

    fn(x,y,z) The next convergence will be collapsing

    datacenter designs into smaller, elastic form

    factors for compute, storage and networking.

    This is where, you can never have enough.

    Customers want scale made easy.

    Hypervisor integration w cloud simplicity.

    L2L3

    L2L3

    L2L3

    L2L3

    L2L3

    L2L3 L2

    L3L2

    L3

    L2L3

    L2L3

    Active - Active redundancy

    fn(x,y,z) The next convergence will be collapsing

    datacenter designs into smaller, elastic form

    factors for compute, storage and networking.

    This is where, you can never have enough.

    Customers want scale made easy.

    Hypervisor integration w cloud simplicity.

    L2L3

    L2L3

    L2L3 L2

    L3

    L2L3

    Start Small; Scale as You Grow

    This is where, you can simply add

    a Extreme Leaf Clusters

    Each cluster is independent

    (including servers, storage,

    database & interconnects).

    Each cluster can be used for

    a different type of service.

    Delivers repeatable design

    which can be added as a

    commodity.

    XYZ Account Spine

    Leaf

    Cluster Cluster Cluster

    Egress

    Scale

    Ingress

    Active / Active

    VM

    VMVM

    Start Small; Scale as You Grow

    This is where, you can simply add

    a Extreme Leaf Clusters

    Each cluster is independent

    (including servers, storage,

    database & interconnects).

    Each cluster can be used for

    a different type of service.

    Delivers repeatable design

    which can be added as a

    commodity.

    XYZ Account Spine

    Leaf

    Cluster Cluster Cluster

    Egress

    Scale

    Ingress

    Active / Active

    VM

    VMVM

    RR RR

    BGP Route-ReflectorRR

    iBGP Adjacency

    This is where

    VXLAN (Route Distribution)

    This is where Why VxLAN? It Flattens network to a single

    tier from the XYZ Account end station

    perspective.

    All IP/BGP based (Virtual eXtensible Local

    Area Network). Host Route Distribution

    decoupled from the Underlay protocol.

    VXLAN s goal is allowing dynamic large

    scale isolated virtual L2 networks to be

    created for virtualized and multi-

    tenant environments.

    Route-Reflectors deployed for scaling

    purposes - Easy setup, small configuration.

    Traffic Engineer like ATM or MPLS

    UDP

    Sta

    rt

    Stop

    UDP UDP

    Use Existing IP Network

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    Traffic Engineer like ATM or MPLS

    UDP

    Sta

    rt

    Stop

    UDP UDP

    Use Existing IP Network

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VMVTEP VTEP

    RR RR

    BGP Route-ReflectorRR

    iBGP Adjacency

    This is where

    VXLAN (Route Distribution)

    This is where Why VxLAN? It Flattens network to a single

    tier from the XYZ Account end station

    perspective.

    All IP/BGP based (Virtual eXtensible Local

    Area Network). Host Route Distribution

    decoupled from the Underlay protocol.

    VXLAN s goal is allowing dynamic large

    scale isolated virtual L2 networks to be

    created for virtualized and multi-

    tenant environments.

    Route-Reflectors deployed for scaling

    purposes - Easy setup, small configuration.

    Traffic Engineer like ATM or MPLS

    UDP

    Sta

    rt

    Stop

    UDP UDP

    Use Existing IP Network

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VMVTEP VTEP

    Dense 10GbE

    Interconnect using

    breakout cables,

    Copper or Fiber

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VMVM

    VM

    VM

    VM

    VM

    App 1

    App 2

    App 3

    Dense 10GbE

    Interconnect using

    breakout cables,

    Copper or Fiber

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VMVM

    VM

    VM

    VM

    VM

    App 1

    App 2

    App 3

    Intel, Facebook, OCP

    Facebook 4-Post Architecture - Each

    leaf or rack switch has up to 48 10G

    downlinks. Segmentation or multi -tenancy

    without routers.

    Each spine has 4 uplinks one to each

    leaf (4:1 oversubscription).

    Enable insertion of services without

    sprawl (Analytics for fabric and

    application forensics).

    No routers at spine. One failure

    reduces cluster capacity to 75%.

    (5 S's) Needs to be Scalable, Secure,

    Shared, Standardized, and Simplified.

    Dense 10GbE

    Interconnect using

    breakout cables,

    Copper or Fiber

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VMVM

    VM

    VM

    VM

    VM

    App 1

    App 2

    App 3

    Intel, Facebook, OCP

    Facebook 4-Post Architecture - Each

    leaf or rack switch has up to 48 10G

    downlinks. Segmentation or multi -tenancy

    without routers.

    Each spine has 4 uplinks one to each

    leaf (4:1 oversubscription).

    Enable insertion of services without

    sprawl (Analytics for fabric and

    application forensics).

    No routers at spine. One failure

    reduces cluster capacity to 75%.

    (5 S's) Needs to be Scalable, Secure,

    Shared, Standardized, and Simplified.

    Network (Fit)

    Active - Active redundancy

    fn(x,y,z) The next convergence will be collapsing

    datacenter designs into smaller, elastic form

    factors for compute, storage and networking.

    This is where, you can never have enough.

    Customers want scale made easy.

    Hypervisor integration w cloud simplicity.

    L2L3

    L2L3

    L2L3 L2

    L3

    L2L3

    Start Small; Scale as You Grow

    This is where, you can simply add

    a Extreme Leaf Clusters

    Each cluster is independent

    (including servers, storage,

    database & interconnects).

    Each cluster can be used for

    a different type of service.

    Delivers repeatable design

    which can be added as a

    commodity.

    XYZ Account Spine

    Leaf

    Cluster Cluster Cluster

    Egress

    Scale

    Ingress

    Active / Active

    VM

    VMVM

    RR RR

    BGP Route-ReflectorRR

    iBGP Adjacency

    This is where

    VXLAN (Route Distribution)

    This is where Why VxLAN? It Flattens network to a single

    tier from the XYZ Account end station

    perspective.

    All IP/BGP based (Virtual eXtensible Local

    Area Network). Host Route Distribution

    decoupled from the Underlay protocol.

    VXLAN s goal is allowing dynamic large

    scale isolated virtual L2 networks to be

    created for virtualized and multi-

    tenant environments.

    Route-Reflectors deployed for scaling

    purposes - Easy setup, small configuration.

    Traffic Engineer like ATM or MPLS

    UDP

    Sta

    rt

    Stop

    UDP UDP

    Use Existing IP Network

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VMVTEP VTEP

    Dense 10GbE

    Interconnect using

    breakout cables,

    Copper or Fiber

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VMVM

    VM

    VM

    VM

    VM

    App 1

    App 2

    App 3

    Intel, Facebook, OCP

    Facebook 4-Post Architecture - Each

    leaf or rack switch has up to 48 10G

    downlinks. Segmentation or multi -tenancy

    without routers.

    Each spine has 4 uplinks one to each

    leaf (4:1 oversubscription).

    Enable insertion of services without

    sprawl (Analytics for fabric and

    application forensics).

    No routers at spine. One failure

    reduces cluster capacity to 75%.

    (5 S's) Needs to be Scalable, Secure,

    Shared, Standardized, and Simplified.

    Network (Fit) Overlay Control

    The XYZ Account the VxLan forwarding plane for NSX control:

    This is where logical switches span across physical hosts and network switches. Application

    continuity is delivered with scale. Scalable Multi-tenancy across data center.

    Enabling L2 over L3 Infrastructure - Pool resources from multiple data centers with the ability to

    recover from disasters faster.

    Address Network Sprawl with an VXLAN overlay. Deeper Integration with infrastructure and

    operations partners, integrations, and frameworks for IT organizations.

    Vmware NSX (Control Plane)

    Management Plane deliver

    by the NSX Manager.

    Control Plane NSX Controller

    Manages Logical networks

    and data plane resources.

    Extreme delivers an open

    high performance data

    plane with Scale

    NSX Architecture and Components

    CORE

    CAMPUS

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    X870-32c

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    10GbAggregation

    High Density 10Gb

    Aggregation

    10Gb/40GbAggregation

    High Density 25Gb/50GbAggregation

    X770 X870-96x-8c

    100GbUplinks

    X670-G2

    100GbUplinks

    Server PODs

    770 / 870 Spine

    Data Center Private Cloud

    vC-1 vC-2

    vC-N

    The XYZ Account the VxLan forwarding plane for NSX control:

    This is where logical switches span across physical hosts and network switches. Application

    continuity is delivered with scale. Scalable Multi-tenancy across data center.

    Enabling L2 over L3 Infrastructure - Pool resources from multiple data centers with the ability to

    recover from disasters faster.

    Address Network Sprawl with an VXLAN overlay. Deeper Integration with infrastructure and

    operations partners, integrations, and frameworks for IT organizations.

    Vmware NSX (Control Plane)

    Management Plane deliver

    by the NSX Manager.

    Control Plane NSX Controller

    Manages Logical networks

    and data plane resources.

    Extreme delivers an open

    high performance data

    plane with Scale

    NSX Architecture and Components

    CORE

    CAMPUS

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    X870-32c

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    10GbAggregation

    High Density 10Gb

    Aggregation

    10Gb/40GbAggregation

    High Density 25Gb/50GbAggregation

    X770 X870-96x-8c

    100GbUplinks

    X670-G2

    100GbUplinks

    Server PODs

    770 / 870 Spine

    Data Center Private Cloud

    vC-1 vC-2

    vC-N

    This is where XYZ Account must first it must have the ability to scale with customer demand,

    delivering more than just disk space and processors.

    Scale XYZ Account must have be able the to seamlessly failover, scale up, scaled down and

    optimize management of the applications and services.

    Flexibility - The infrastructure XYZ Account must have the ability to host heterogeneous and

    interoperable technologies.

    Business - The business model costs might be optimized for operating expenses or towards

    capital investment.

    Cloud Computing (Control Plane)

    (On-Premise)

    Infrastructure(as a Service)

    Platform(as a Service)

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Data

    Applications

    Runtime

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Data

    Applications

    Runtime

    You

    man

    age M

    anaged by vendor

    Managed by vendor

    You

    man

    age

    You

    man

    age

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Applications

    Runtime

    Data

    Software(as a Service)

    Managed by vendor

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Applications

    Runtime

    Data

    Public

    Private

    MSP

    FABRIC

    This is where XYZ Account must first it must have the ability to scale with customer demand,

    delivering more than just disk space and processors.

    Scale XYZ Account must have be able the to seamlessly failover, scale up, scaled down and

    optimize management of the applications and services.

    Flexibility - The infrastructure XYZ Account must have the ability to host heterogeneous and

    interoperable technologies.

    Business - The business model costs might be optimized for operating expenses or towards

    capital investment.

    Cloud Computing (Control Plane)

    (On-Premise)

    Infrastructure(as a Service)

    Platform(as a Service)

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Data

    Applications

    Runtime

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Data

    Applications

    Runtime

    You

    man

    age M

    anaged by vendor

    Managed by vendor

    You

    man

    age

    You

    man

    age

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Applications

    Runtime

    Data

    Software(as a Service)

    Managed by vendor

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Applications

    Runtime

    Data

    Public

    Private

    MSP

    FABRIC

    This is where Azure ExpressRoute lets XYZ Account create private connections between Azure

    datacenters and XYZ Account infrastructure on or off premises.

    ExpressRoute connections don't go over the public Internet. They offer more reliability, faster

    speeds, and lower latencies, and higher security than typical Internet connections.

    XYZ Account can transfer data between on-premises systems and Azure can yield significant

    cost benefits.

    XYZ Account can establishing connections to Azure at an ExpressRoute location, such as an

    Exchange provider facility, or directly connect to Azure from your existing WAN network, such as

    a multi-protocol label switching (MPLS) VPN, provided by a network service provider

    Microsoft Assure (Control Plane)

    Cloud The key impact of this model

    for the customer is a move from

    managing physical servers to focus on

    logical management of data storage

    through policies.

    This is where Azure ExpressRoute lets XYZ Account create private connections between Azure

    datacenters and XYZ Account infrastructure on or off premises.

    ExpressRoute connections don't go over the public Internet. They offer more reliability, faster

    speeds, and lower latencies, and higher security than typical Internet connections.

    XYZ Account can transfer data between on-premises systems and Azure can yield significant

    cost benefits.

    XYZ Account can establishing connections to Azure at an ExpressRoute location, such as an

    Exchange provider facility, or directly connect to Azure from your existing WAN network, such as

    a multi-protocol label switching (MPLS) VPN, provided by a network service provider

    Microsoft Assure (Control Plane)

    Cloud The key impact of this model

    for the customer is a move from

    managing physical servers to focus on

    logical management of data storage

    through policies.

    Overlay Control

    The XYZ Account the VxLan forwarding plane for NSX control:

    This is where logical switches span across physical hosts and network switches. Application

    continuity is delivered with scale. Scalable Multi-tenancy across data center.

    Enabling L2 over L3 Infrastructure - Pool resources from multiple data centers with the ability to

    recover from disasters faster.

    Address Network Sprawl with an VXLAN overlay. Deeper Integration with infrastructure and

    operations partners, integrations, and frameworks for IT organizations.

    Vmware NSX (Control Plane)

    Management Plane deliver

    by the NSX Manager.

    Control Plane NSX Controller

    Manages Logical networks

    and data plane resources.

    Extreme delivers an open

    high performance data

    plane with Scale

    NSX Architecture and Components

    CORE

    CAMPUS

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    X870-32c

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    10GbAggregation

    High Density 10Gb

    Aggregation

    10Gb/40GbAggregation

    High Density 25Gb/50GbAggregation

    X770 X870-96x-8c

    100GbUplinks

    X670-G2

    100GbUplinks

    Server PODs

    770 / 870 Spine

    Data Center Private Cloud

    vC-1 vC-2

    vC-N

    This is where XYZ Account must first it must have the ability to scale with customer demand,

    delivering more than just disk space and processors.

    Scale XYZ Account must have be able the to seamlessly failover, scale up, scaled down and

    optimize management of the applications and services.

    Flexibility - The infrastructure XYZ Account must have the ability to host heterogeneous and

    interoperable technologies.

    Business - The business model costs might be optimized for operating expenses or towards

    capital investment.

    Cloud Computing (Control Plane)

    (On-Premise)

    Infrastructure(as a Service)

    Platform(as a Service)

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Data

    Applications

    Runtime

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Data

    Applications

    Runtime

    You

    man

    age M

    anaged by vendor

    Managed by vendor

    You

    man

    age

    You

    man

    age

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Applications

    Runtime

    Data

    Software(as a Service)

    Managed by vendor

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Applications

    Runtime

    Data

    Public

    Private

    MSP

    FABRIC

    This is where Azure ExpressRoute lets XYZ Account create private connections between Azure

    datacenters and XYZ Account infrastructure on or off premises.

    ExpressRoute connections don't go over the public Internet. They offer more reliability, faster

    speeds, and lower latencies, and higher security than typical Internet connections.

    XYZ Account can transfer data between on-premises systems and Azure can yield significant

    cost benefits.

    XYZ Account can establishing connections to Azure at an ExpressRoute location, such as an

    Exchange provider facility, or directly connect to Azure from your existing WAN network, such as

    a multi-protocol label switching (MPLS) VPN, provided by a network service provider

    Microsoft Assure (Control Plane)

    Cloud The key impact of this model

    for the customer is a move from

    managing physical servers to focus on

    logical management of data storage

    through policies.

    Overlay Control

    The XYZ Account the VxLan forwarding plane for NSX control:

    This is where logical switches span across physical hosts and network switches. Application

    continuity is delivered with scale. Scalable Multi-tenancy across data center.

    Enabling L2 over L3 Infrastructure - Pool resources from multiple data centers with the ability to

    recover from disasters faster.

    Address Network Sprawl with an VXLAN overlay. Deeper Integration with infrastructure and

    operations partners, integrations, and frameworks for IT organizations.

    Vmware NSX (Control Plane)

    Management Plane deliver

    by the NSX Manager.

    Control Plane NSX Controller

    Manages Logical networks

    and data plane resources.

    Extreme delivers an open

    high performance data

    plane with Scale

    NSX Architecture and Components

    CORE

    CAMPUS

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    X870-32c

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    10GbAggregation

    High Density 10Gb

    Aggregation

    10Gb/40GbAggregation

    High Density 25Gb/50GbAggregation

    X770 X870-96x-8c

    100GbUplinks

    X670-G2

    100GbUplinks

    Server PODs

    770 / 870 Spine

    Data Center Private Cloud

    vC-1 vC-2

    vC-N

    This is where XYZ Account must first it must have the ability to scale with customer demand,

    delivering more than just disk space and processors.

    Scale XYZ Account must have be able the to seamlessly failover, scale up, scaled down and

    optimize management of the applications and services.

    Flexibility - The infrastructure XYZ Account must have the ability to host heterogeneous and

    interoperable technologies.

    Business - The business model costs might be optimized for operating expenses or towards

    capital investment.

    Cloud Computing (Control Plane)

    (On-Premise)

    Infrastructure(as a Service)

    Platform(as a Service)

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Data

    Applications

    Runtime

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Data

    Applications

    Runtime

    You

    man

    age M

    anaged by vendor

    Managed by vendor

    You

    man

    age

    You

    man

    age

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Applications

    Runtime

    Data

    Software(as a Service)

    Managed by vendor

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Applications

    Runtime

    Data

    Public

    Private

    MSP

    FABRIC

    This is where Azure ExpressRoute lets XYZ Account create private connections between Azure

    datacenters and XYZ Account infrastructure on or off premises.

    ExpressRoute connections don't go over the public Internet. They offer more reliability, faster

    speeds, and lower latencies, and higher security than typical Internet connections.

    XYZ Account can transfer data between on-premises systems and Azure can yield significant

    cost benefits.

    XYZ Account can establishing connections to Azure at an ExpressRoute location, such as an

    Exchange provider facility, or directly connect to Azure from your existing WAN network, such as

    a multi-protocol label switching (MPLS) VPN, provided by a network service provider

    Microsoft Assure (Control Plane)

    Cloud The key impact of this model

    for the customer is a move from

    managing physical servers to focus on

    logical management of data storage

    through policies.

    Compute Storage

    Data Center Architecture

    Considerations

    ComputeCache

    Database

    Storage

    Client

    Response

    80% North-South Traffic

    Oversubscription : upto 200:1 (Client

    Request +Server Response = 20%

    traffic).

    Inter-rack latency: 150 micros.

    Lookup Storage = 80% traffic.

    Scale: Up to 20 racks (features Non-

    blocking 2 tier designs optimal).

    VM

    VM VM

    VM

    Data Center Architecture

    Considerations

    ComputeCache

    Database

    Storage

    Client

    Response

    80% North-South Traffic

    Oversubscription : upto 200:1 (Client

    Request +Server Response = 20%

    traffic).

    Inter-rack latency: 150 micros.

    Lookup Storage = 80% traffic.

    Scale: Up to 20 racks (features Non-

    blocking 2 tier designs optimal).

    VM

    VM VM

    VM

    Purchase "vanity free"

    This is where..Open Compute might allow companies to

    purchase "vanity free". Previous outdated

    data center designs support more

    monolithic computing.

    Low density X620 might help XYZ

    Account to avoid stranded ports.

    Availability - Dual X620s can be

    deployed to minimize impact to

    maintenance.

    Flexibility of the X620 can offer

    flexibility to support both 1G and 10G to

    servers and storage.

    One RACK Design

    Closely

    coupled

    Nearly

    coupled

    Loosely

    coupled

    Shared Combo Ports4x10GBASE-T & 4xSFP+

    100Mb/1Gb/10GBASE-T

    The monolithic datacenter

    is dead.

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    ServersServers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Purchase "vanity free"

    This is where..Open Compute might allow companies to

    purchase "vanity free". Previous outdated

    data center designs support more

    monolithic computing.

    Low density X620 might help XYZ

    Account to avoid stranded ports.

    Availability - Dual X620s can be

    deployed to minimize impact to

    maintenance.

    Flexibility of the X620 can offer

    flexibility to support both 1G and 10G to

    servers and storage.

    One RACK Design

    Closely

    coupled

    Nearly

    coupled

    Loosely

    coupled

    Shared Combo Ports4x10GBASE-T & 4xSFP+

    100Mb/1Gb/10GBASE-T

    The monolithic datacenter

    is dead.

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Open Compute - Two Rack Design

    This is where, XYZ Account can reduce OPEX and

    leverage a repeatable solution.

    With the spline setup, XYZ Account can put

    redundant switches in the middle and link

    each server to those switches.

    Fewer Hops between Servers - The important

    thing is that each server is precisely one hop

    from any other server.

    Avoid Stranded ports Designs often have a

    mix of fat and skinny nodes. If XYZ Account

    deploys a 48-port leaf switches many

    configurations might have anywhere from 16

    to 24 stranded ports.

    Two RACK

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Typical spline setup

    Two RACK

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Typical spline setup

    Open Compute - Two Rack Design

    This is where, XYZ Account can reduce OPEX and

    leverage a repeatable solution.

    With the spline setup, XYZ Account can put

    redundant switches in the middle and link

    each server to those switches.

    Fewer Hops between Servers - The important

    thing is that each server is precisely one hop

    from any other server.

    Avoid Stranded ports Designs often have a

    mix of fat and skinny nodes. If XYZ Account

    deploys a 48-port leaf switches many

    configurations might have anywhere from 16

    to 24 stranded ports.

    Two RACK

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Typical spline setup

    Open Compute : Eight Rack POD Design

    This is where

    Typical spline setup : Eight Rack POD

    Leaf

    Spine

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    ServersServers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    ServersServers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    ServersServers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    ServersServers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Open Compute : Eight Rack POD Design

    This is where

    Typical spline setup : Eight Rack POD

    Leaf

    Spine

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    ServersServers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Data Center Architecture

    Considerations

    ComputeCache

    Database

    Storage

    Client

    Response

    80% North-South Traffic

    Oversubscription : upto 200:1 (Client

    Request +Server Response = 20%

    traffic).

    Inter-rack latency: 150 micros.

    Lookup Storage = 80% traffic.

    Scale: Up to 20 racks (features Non-

    blocking 2 tier designs optimal).

    VM

    VM VM

    VM

    Purchase "vanity free"

    This is where..Open Compute might allow companies to

    purchase "vanity free". Previous outdated

    data center designs support more

    monolithic computing.

    Low density X620 might help XYZ

    Account to avoid stranded ports.

    Availability - Dual X620s can be

    deployed to minimize impact to

    maintenance.

    Flexibility of the X620 can offer

    flexibility to support both 1G and 10G to

    servers and storage.

    One RACK Design

    Closely

    coupled

    Nearly

    coupled

    Loosely

    coupled

    Shared Combo Ports4x10GBASE-T & 4xSFP+

    100Mb/1Gb/10GBASE-T

    The monolithic datacenter

    is dead.

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Open Compute - Two Rack Design

    This is where, XYZ Account can reduce OPEX and

    leverage a repeatable solution.

    With the spline setup, XYZ Account can put

    redundant switches in the middle and link

    each server to those switches.

    Fewer Hops between Servers - The important

    thing is that each server is precisely one hop

    from any other server.

    Avoid Stranded ports Designs often have a

    mix of fat and skinny nodes. If XYZ Account

    deploys a 48-port leaf switches many

    configurations might have anywhere from 16

    to 24 stranded ports.

    Two RACK

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Typical spline setup

    Open Compute : Eight Rack POD Design

    This is where

    Typical spline setup : Eight Rack POD

    Leaf

    Spine

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    ServersServers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Compute Storage

    Data Center Architecture

    Considerations

    ComputeCache

    Database

    Storage

    Client

    Response

    80% North-South Traffic

    Oversubscription : upto 200:1 (Client

    Request +Server Response = 20%

    traffic).

    Inter-rack latency: 150 micros.

    Lookup Storage = 80% traffic.

    Scale: Up to 20 racks (features Non-

    blocking 2 tier designs optimal).

    VM

    VM VM

    VM

    Purchase "vanity free"

    This is where..Open Compute might allow companies to

    purchase "vanity free". Previous outdated

    data center designs support more

    monolithic computing.

    Low density X620 might help XYZ

    Account to avoid stranded ports.

    Availability - Dual X620s can be

    deployed to minimize impact to

    maintenance.

    Flexibility of the X620 can offer

    flexibility to support both 1G and 10G to

    servers and storage.

    One RACK Design

    Closely

    coupled

    Nearly

    coupled

    Loosely

    coupled

    Shared Combo Ports4x10GBASE-T & 4xSFP+

    100Mb/1Gb/10GBASE-T

    The monolithic datacenter

    is dead.

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Open Compute - Two Rack Design

    This is where, XYZ Account can reduce OPEX and

    leverage a repeatable solution.

    With the spline setup, XYZ Account can put

    redundant switches in the middle and link

    each server to those switches.

    Fewer Hops between Servers - The important

    thing is that each server is precisely one hop

    from any other server.

    Avoid Stranded ports Designs often have a

    mix of fat and skinny nodes. If XYZ Account

    deploys a 48-port leaf switches many

    configurations might have anywhere from 16

    to 24 stranded ports.

    Two RACK

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Typical spline setup

    Open Compute : Eight Rack POD Design

    This is where

    Typical spline setup : Eight Rack POD

    Leaf

    Spine

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    ServersServers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Chassis V Spline

    Fabric Modules (Spine)

    I/O Modules (Leaf)

    Spine

    Leaf

    Proven value with legacy approach.

    Can not access Line cards.

    No L2/l3 recovery inside.

    No access to Fabric.

    Disaggregated value...

    Control Top-of-Rack Switches

    L2/L3 protocols inside the Spline

    Full access to Spine Switches

    No EGO, Complexity or Vendor Lock-in).

    Fat-Tree

    Clos / Cross-Bar

    Traditional 3-tier model (Less cabling).

    Link speeds must increase at every hop (Less

    predictable latency).

    Common in Chassis based architectures (Optimized

    for North/South traffic).

    Every Leaf is connected to every Spine (Efficient

    utilization/ Very predictable latency).

    Always two hops to any leaf (More resiliency,

    flexibility and performance).

    Friendlier to east/west traffic (The uplink to the

    rest of the network is just another leaf).

    The XYZ Account handshake layer:

    This is where convergence needs to happen LAN/SAN, FCoE, ETS. Stop or allow

    whatever you can (Efficient Multicasting).

    Virtualization happens with VXLAN and VMotion (Control by the overlay).

    N plus one fabric design needs to happen here (Delivers simple no vanity future proofing,

    No-forklift migrations, interop between vendors and hit-less operation).

    This is where,

    a Fabric outperforms the Big Uglies

    ONE to ONE: Spine Leaf

    The XYZ Account Ethernet Expressway Layer: deliver massive scale...

    This is where low latency is critical, switch as quickly as you can. DO NOT slow down

    the core keep it simple (Disaggregated Spline + One Big Ugly

    Elastic Capacity - Today s XYZ Account s spines are tomorrow s leafs. Dial-in the

    bandwidth to your specific needs with the number of uplinks.

    Availability - the state of the network is kept in each switch; no single point of failure.

    Seamless XYZ Account upgrades, easy to take a single switch out of service.

    (Cloud Fabric) Disaggregation

    Spine

    Leaf

    Legacy Challenges:

    Complex/Slow/Expensive

    Scale-up and Scale out

    Vendor lock-in

    Proprietary (HW, SW)Commodity

    Fabric Modules (Spine)

    I/O M

    odules (Lea

    f)

    Spline (Speed)

    Active - Active redundancy

    fn(x,y,z) The next convergence will be collapsing

    datacenter designs into smaller, elastic form

    factors for compute, storage and networking.

    This is where, you can never have enough.

    Customers want scale made easy.

    Hypervisor integration w cloud simplicity.

    L2L3

    L2L3

    L2L3 L2

    L3

    L2L3

    Start Small; Scale as You Grow

    This is where, you can simply add

    a Extreme Leaf Clusters

    Each cluster is independent

    (including servers, storage,

    database & interconnects).

    Each cluster can be used for

    a different type of service.

    Delivers repeatable design

    which can be added as a

    commodity.

    XYZ Account Spine

    Leaf

    Cluster Cluster Cluster

    Egress

    Scale

    Ingress

    Active / Active

    VM

    VMVM

    RR RR

    BGP Route-ReflectorRR

    iBGP Adjacency

    This is where

    VXLAN (Route Distribution)

    This is where Why VxLAN? It Flattens network to a single

    tier from the XYZ Account end station

    perspective.

    All IP/BGP based (Virtual eXtensible Local

    Area Network). Host Route Distribution

    decoupled from the Underlay protocol.

    VXLAN s goal is allowing dynamic large

    scale isolated virtual L2 networks to be

    created for virtualized and multi-

    tenant environments.

    Route-Reflectors deployed for scaling

    purposes - Easy setup, small configuration.

    Traffic Engineer like ATM or MPLS

    UDP

    Sta

    rt

    Stop

    UDP UDP

    Use Existing IP Network

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VMVTEP VTEP

    Dense 10GbE

    Interconnect using

    breakout cables,

    Copper or Fiber

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VM

    VMVM

    VM

    VM

    VM

    VM

    App 1

    App 2

    App 3

    Intel, Facebook, OCP

    Facebook 4-Post Architecture - Each

    leaf or rack switch has up to 48 10G

    downlinks. Segmentation or multi -tenancy

    without routers.

    Each spine has 4 uplinks one to each

    leaf (4:1 oversubscription).

    Enable insertion of services without

    sprawl (Analytics for fabric and

    application forensics).

    No routers at spine. One failure

    reduces cluster capacity to 75%.

    (5 S's) Needs to be Scalable, Secure,

    Shared, Standardized, and Simplified.

    Network (Fit) Overlay Control

    The XYZ Account the VxLan forwarding plane for NSX control:

    This is where logical switches span across physical hosts and network switches. Application

    continuity is delivered with scale. Scalable Multi-tenancy across data center.

    Enabling L2 over L3 Infrastructure - Pool resources from multiple data centers with the ability to

    recover from disasters faster.

    Address Network Sprawl with an VXLAN overlay. Deeper Integration with infrastructure and

    operations partners, integrations, and frameworks for IT organizations.

    Vmware NSX (Control Plane)

    Management Plane deliver

    by the NSX Manager.

    Control Plane NSX Controller

    Manages Logical networks

    and data plane resources.

    Extreme delivers an open

    high performance data

    plane with Scale

    NSX Architecture and Components

    CORE

    CAMPUS

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    X870-32c

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    01

    05

    10

    15

    20

    25

    30

    35

    40

    02

    03

    04

    06

    07

    08

    09

    11

    12

    13

    14

    16

    17

    18

    19

    21

    22

    23

    24

    26

    27

    28

    29

    31

    32

    33

    34

    36

    37

    38

    39

    41

    42

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    2GbGb 1 Gb 3 Gb 4

    21

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    N3K-C3064PQ-FASTAT

    ID

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 1 2 3 4

    10GbAggregation

    High Density 10Gb

    Aggregation

    10Gb/40GbAggregation

    High Density 25Gb/50GbAggregation

    X770 X870-96x-8c

    100GbUplinks

    X670-G2

    100GbUplinks

    Server PODs

    770 / 870 Spine

    Data Center Private Cloud

    vC-1 vC-2

    vC-N

    This is where XYZ Account must first it must have the ability to scale with customer demand,

    delivering more than just disk space and processors.

    Scale XYZ Account must have be able the to seamlessly failover, scale up, scaled down and

    optimize management of the applications and services.

    Flexibility - The infrastructure XYZ Account must have the ability to host heterogeneous and

    interoperable technologies.

    Business - The business model costs might be optimized for operating expenses or towards

    capital investment.

    Cloud Computing (Control Plane)

    (On-Premise)

    Infrastructure(as a Service)

    Platform(as a Service)

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Data

    Applications

    Runtime

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Data

    Applications

    Runtime

    You

    man

    age M

    anaged by vendor

    Managed by vendor

    You

    man

    age

    You

    man

    age

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Applications

    Runtime

    Data

    Software(as a Service)

    Managed by vendor

    Storage

    Servers

    Networking

    O/S

    Middleware

    Virtualization

    Applications

    Runtime

    Data

    Public

    Private

    MSP

    FABRIC

    This is where Azure ExpressRoute lets XYZ Account create private connections between Azure

    datacenters and XYZ Account infrastructure on or off premises.

    ExpressRoute connections don't go over the public Internet. They offer more reliability, faster

    speeds, and lower latencies, and higher security than typical Internet connections.

    XYZ Account can transfer data between on-premises systems and Azure can yield significant

    cost benefits.

    XYZ Account can establishing connections to Azure at an ExpressRoute location, such as an

    Exchange provider facility, or directly connect to Azure from your existing WAN network, such as

    a multi-protocol label switching (MPLS) VPN, provided by a network service provider

    Microsoft Assure (Control Plane)

    Cloud The key impact of this model

    for the customer is a move from

    managing physical servers to focus on

    logical management of data storage

    through policies.

    Compute Storage

    Data Center Architecture

    Considerations

    ComputeCache

    Database

    Storage

    Client

    Response

    80% North-South Traffic

    Oversubscription : upto 200:1 (Client

    Request +Server Response = 20%

    traffic).

    Inter-rack latency: 150 micros.

    Lookup Storage = 80% traffic.

    Scale: Up to 20 racks (features Non-

    blocking 2 tier designs optimal).

    VM

    VM VM

    VM

    Purchase "vanity free"

    This is where..Open Compute might allow companies to

    purchase "vanity free". Previous outdated

    data center designs support more

    monolithic computing.

    Low density X620 might help XYZ

    Account to avoid stranded ports.

    Availability - Dual X620s can be

    deployed to minimize impact to

    maintenance.

    Flexibility of the X620 can offer

    flexibility to support both 1G and 10G to

    servers and storage.

    One RACK Design

    Closely

    coupled

    Nearly

    coupled

    Loosely

    coupled

    Shared Combo Ports4x10GBASE-T & 4xSFP+

    100Mb/1Gb/10GBASE-T

    The monolithic datacenter

    is dead.

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Open Compute - Two Rack Design

    This is where, XYZ Account can reduce OPEX and

    leverage a repeatable solution.

    With the spline setup, XYZ Account can put

    redundant switches in the middle and link

    each server to those switches.

    Fewer Hops between Servers - The important

    thing is that each server is precisely one hop

    from any other server.

    Avoid Stranded ports Designs often have a

    mix of fat and skinny nodes. If XYZ Account

    deploys a 48-port leaf switches many

    configurations might have anywhere from 16

    to 24 stranded ports.

    Two RACK

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Management

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Typical spline setup

    Open Compute : Eight Rack POD Design

    This is where

    Typical spline setup : Eight Rack POD

    Leaf

    Spine

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Servers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    ServersServers

    Storage

    Summit

    Management

    Switch

    Summit

    Summit

    Storage

    Reference Great Networks POV October 2016 v2.vsdxData Center


Recommended