gpx data ingestion & dx presentataion at aws mumbai summit 2017

23
Bulk Data Upload using AWS Direct Connect @ Tier IV DC GPX India Private Limited, 001, Boomerang, Chandivali Farm Road, Andheri East, Mumbai – 400072 www.gpxglobal.net Manoj Paul Managing Director

Upload: cloudxchangeio

Post on 21-Jan-2018

163 views

Category:

Technology


1 download

TRANSCRIPT

Bulk Data Upload using

AWS Direct Connect

@Tier IV DC

GPX India Private Limited, 001, Boomerang, Chandivali Farm Road, Andheri East, Mumbai – 400072www.gpxglobal.net

Manoj PaulManaging Director

GPX DC Location in Mumbai

GPX Mumbai Data Center• Asia’s first and India’s only Uptime Institute design certified Tier IV DC service provider offering 99.999% uptime

SLA

• India’s most “Carrier Content Cloud” rich Data Center with an ecosystem comprised of Carriers (~ 12 TelecomService Providers), ISP’s(50+), content service providers, hosting companies, and cloud service providers

• India’s first AWS Direct Connect PoP @ GPX DC and only DX with Singapore and India region connectivity

• Awarded “State of Art Data Center” by Economic times

Revisiting : What is AWS Direct Connect• Dedicated network connection from your Data Center / office to AWS• Using MPLS or dedicated leased lines, avoiding the internet route

Customer premises or DCAWS Cloud

Virtual Private Cloud

100mbps1 Gb/s

10 Gb/sSESAmazon

GlacierElastic

BeanstalkSQS

Amazon EMR

AmazonRedshift

EC2

AWS Direct Connect @ GPX

Mumbai

Direct Connect

MPLS

Leased Line

Direct Connect Advantages:Direct Connect is a Dedicated, High-Capacity, Secure Link to

Amazon Web Services

Source: http://aws.amazon.com/directconnect/

Data Download charges

AWS DX Set up: Scenario 1Scenario – I . Customers with their racks collocated @ GPX DC

Router A Router R

Cloud Uplink (A)10GBASE-LR (1310nm)

10GBASE-LR (1310nm)

Legend:A – ActiveR - Redundant

1 G/100mbps/500mbps Fiber/Copper links to

customer racks

Cloud Uplink (R)

10GBASE-LR (1310nm)

GPX Switch Fabric

AWS Router AWS Router

1 G/100mbps/500mbps Fiber / Copper links to customer

racks

GPX Cross connect

GP

X D

ata

cen

ter

AWS Network

Customer Racks

Benefits of Co-located racks at GPX• Advantages of colocation at GPX for Direct Connect:

Zero ISP bandwidth cost

Elasticity : Connect over 1Gbps , flexibility for heavy upload/ downloads

Cost effective redundant link to AWS cloud

Higher security

• DR on cloud using Direct Connect

Ideal for a DC _ DR set up:

Host your servers in a Tier IV DC offering 99.999% uptime

Connect to AWS over LAN: zero DC/DR bandwidth cost

Pay only when AWS resources are used-ideal for DR

Scenario – II . For Customers outside GPX Data Center

Router A Router R

Cloud Uplink (A)10GBASE-LR (1310nm)

10GBASE-LR (1310nm)

Legend:A – ActiveR - Redundant

1 G/100mbps/500mbps Fiber/Copper point to point links to customer premises

Cloud Uplink (R)10GBASE-LR (1310nm)

GPX Switch Fabric

AWS Router AWS Router

1 G/100mbps/500mbps Fiber/Copper point to point links to customer premises

Customer Data center

Carrier MUX Carrier MUX

PtoP / MPLS Links from TSP

GP

X D

ata

cen

terGPX Switch Fabric

Customer IT infra NOT colocated at GPX DC

Customers not collocated at GPX• Set up steps:

1. AWS Direct Connect port - subscribe on AWS console

2. MPLS/Leased Line to GPX DC - provided by Telecom Service Provider

3. Cross Connect from TSP rack at carrier room to GPX Switch fabric or direct cross connect to AWS router for 1gbps/10Gbps - provided by GPX

• Constrains :

Varying bandwidth requirement becomes difficult to implement

Two MPLS/Leased line links needed for redundancy

Bulk Data Ingestion

• With launch of India region of AWS, many customers would migrate from own

set up to cloud

– One time requirement of uploading huge existing data

• Enterprises are setting up DR on cloud

– One time requirement of uploading bulk existing data

• Media companies

– infrequent bulk video file uploads

Objective: Need for bulk upload

• Internet:

• Time consuming- 100 TB data can take 100+ days over 100mbps link;• Expensive

• Snowball: Generally used for very high volume ( Petabyte) data

Bulk Data Ingestion Service

@

Data Center

Pre Configured Set up for Bulk Data Ingest

AWS DX

• Pre configured set- up at GPX DC, enabling uploading of 10/100 Terra bytes of data

Cost effective

Efficient

• Elastic solution for differed requirements :

elastic bandwidth as bandwidth requirement would vary

Variable Duration

• Space for keeping storage hardware and hooking it up to the cloud

Features :

Steps:Customer:• Customer brings in storage hardware to GPX DC

GPX :• Provide temporary dedicated rack compartment with power to place customer

hardware.• Preconfigured cross connect using fibre port for 1G/10G or CAT6 FE port for

sub 1G cross connects• For sub 1 gig, GPX will provision direct connect on AWS console• Temporary workstation with power and cross connect for customer technician

Customer:• Performs the data upload

Case Study

10 TB

Case Study Details

• Leading Petrochem Engineering company head quartered in France.

• They were looking for DR and archival as a service for their DC in Mumbai

Proposed Solution :

DR as a Service for 8 servers with 15 TB data.

Archival as a Service for 15 TB data stored on tapes and NAS storages

Data ingestion service with direct connect

Proposed Solution Architecture

X86 File Servers

SAN/ LAN

Customer DC

AWS MUM Region DC

Old Data On Tapes and

Servers

GPX DC

Storage Appliance for data movement

Data Injection Steps

• Storage appliance provided with 12 TB usable capacity

• Data copied from multiple servers to storage appliance using double take and

robo copy

• Storage appliance physically moved to GPX and connected to direct connect

port provided by GPX

• Pre Configured direct connect with 500 Mbps speed was available

• Migrated data to EBS using robo copy and FTP

Type of data migrated

• Autocad engineering drawings

• 3D images & videos for engineering applications

• Videos

• Office files

More than 60% data were engineering drawings and videos

Data Injection activity report

• Total data migrated = 10 TB

• Direct connect port speed configured : 500 Mbps

• Total time took to complete transfer : 92 hours • Note : First 5 TB was transferred within 36 hours, remaining took some

time.

• Fully secured process with customer having full control over the data