+ All Categories
Home > Technology > AWS Webcast - Managing Big Data in the AWS Cloud_20140924

AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Date post: 02-Nov-2014
Category:
Upload: amazon-web-services
View: 721 times
Download: 1 times
Share this document with a friend
Description:
This presentation deck will cover specific services such as Amazon S3, Kinesis, Redshift, Elastic MapReduce, and DynamoDB, including their features and performance characteristics. It will also cover architectural designs for the optimal use of these services based on dimensions of your data source (structured or unstructured data, volume, item size and transfer rates) and application considerations - for latency, cost and durability. It will also share customer success stories and resources to help you get started.
Popular Tags:
87
Managing Big Data in the AWS Cloud Siva Raghupathy Principal Solutions Architect Amazon Web Services
Transcript
Page 1: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Managing Big Data in the AWS Cloud

Siva Raghupathy

Principal Solutions Architect

Amazon Web Services

Page 2: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Agenda

• Big data challenges • AWS big data portfolio• Architectural considerations• Customer success stories• Resources to help you get started• Q&A

Page 3: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Data Volume, Velocity, & Variety

• 4.4 zettabytes (ZB) of data exists in the digital universe today– 1 ZB = 1 billion terabytes

• 450 billion transaction per day by 2020

• More unstructured data than structured data GB

TB

PB

ZB

EB

1990 2000 2010 2020

Page 4: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Big Data• Hourly server logs: how your systems were

misbehaving an hour ago

• Weekly / Monthly Bill: What you spent this past billing cycle?

• Daily customer-preferences report from your web-site’s click stream: tells you what deal or ad to try next time

• Daily fraud reports: tells you if there was fraud yesterday

Real-time Big Data• Real-time metrics: what just went wrong

now

• Real-time spending alerts/caps: guaranteeing you can’t overspend

• Real-time analysis: tells you what to offer the current customer now

• Real-time detection: blocks fraudulent use now

Big Data : Best Served Fresh

Page 5: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Gartner: User Survey Analysis: Key Trends Shaping the Future of Data Center Infrastructure Through 2011 IDC: Worldwide Business Analytics Software 2012–2016 Forecast and 2011 Vendor Shares

Available for analysis

Generated dataData volume - Gap

1990 2000 2010 2020

Data Analysis Gap

Page 6: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Big Data

Potentially massive datasets

Iterative, experimental style of data manipulation and analysis

Frequently not a steady-state workload;

peaks and valleys

Time to results is key

Hard to configure/manage

AWS Cloud

Massive, virtually unlimited capacity

Iterative, experimental style of infrastructure deployment/usage

At its most efficient with highly variable workloads

Parallel compute clusters from singe data source

Managed services

Page 7: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

AWS Big Data Portfolio

Collect / Ingest

Kinesis

Process / Analyze

EMR EC2

Redshift Data Pipeline

Visualize / ReportStore

Glacier

S3

DynamoDB

RDS

Import Export

Direct Connect

Amazon SQS

Page 8: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Ingest: The act of collecting and storing data

Page 9: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Why Data Ingest Tools?

• Data ingest tools convert random streams of data into fewer set of sequential streams

– Sequential streams are easier to process

– Easier to scale

– Easier to persist

Page 10: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Data Ingest Tools

• Facebook Scribe Data collectors

• Apache Kafka Data collectors

• Apache Flume Data Movement and Transformation

• Amazon Kinesis Data collectors

Page 11: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Real-time processing of streaming data

High throughput

Elastic

Easy to use

Connectors for EMR, S3, Redshift, DynamoDB

Amazon Kinesis

Page 12: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Amazon Kinesis ArchitectureAmazon Kinesis Architecture

Page 13: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Kinesis Stream: Managed ability to capture and store data

• Streams are made of Shards

• Each Shard ingests data up to

1MB/sec, and up to 1000 TPS

• Each Shard emits up to 2 MB/sec

• All data is stored for 24 hours

• Scale Kinesis streams by adding or

removing Shards

• Replay data inside of 24Hr. Window

Page 14: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Simple Put interface to store data in Kinesis

• Producers use a PUT call to store data in a Stream

• PutRecord {Data, PartitionKey,

StreamName}

• A Partition Key is supplied by producer and used to

distribute the PUTs across Shards

• Kinesis MD5 hashes supplied partition key over the hash

key range of a Shard

• A unique Sequence # is returned to the Producer upon a

successful PUT call

Page 15: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Building Kinesis Processing Apps: Kinesis Client LibraryClient library for fault-tolerant, at least-once, Continuous Processing

o Java client library, source available on Github

o Build & Deploy app with KCL on your EC2 instance(s)

o KCL is intermediary b/w your application & stream

Automatically starts a Kinesis Worker for each shard

Simplifies reading by abstracting individual shards

Increase / Decrease Workers as # of shards changes

Checkpoints to keep track of a Worker’s location in

the stream, Restarts Workers if they fail

o Integrates with AutoScaling groups to redistribute workers

to new instances

Page 16: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Sending & Reading Data from Kinesis Streams

HTTP Post

AWS SDK

LOG4J

Flume

Fluentd

Get* APIs

Kinesis Client Library +Connector Library

Apache Storm

Amazon Elastic MapReduce

Sending Reading

Write Read

Page 17: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

AWS Partners for Data Load and Transformation

Hparser, Big Data Edition

Flume, Sqoop

Page 18: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Storage

Page 19: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Storage

Structured – Simple QueryNoSQL

Amazon DynamoDBCache

Amazon ElastiCache (Memcached, Redis)

Structured – Complex QuerySQL

Amazon RDS Data Warehouse

Amazon RedshiftSearch

Amazon CloudSearch

Unstructured – No QueryCloud Storage

Amazon S3Amazon Glacier

Unstructured – Custom QueryHadoop/HDFS

Amazon Elastic Map Reduce

Dat

a St

ruct

ure

Com

plex

ity

Query Structure Complexity

Page 20: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Store anything

Object storage

Scalable

Designed for 99.999999999% durability

Amazon S3

Page 21: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Why is Amazon S3 good for Big Data?

• No limit on the number of Objects• Object size up to 5TB• Central data storage for all systems• High bandwidth• 99.999999999% durability• Versioning, Lifecycle Policies• Glacier Integration

Page 22: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Amazon S3 Best Practices

• Use random hash prefix for keys

• Ensure a random access pattern

• Use Amazon CloudFront for high throughput GETs and PUTs

• Leverage the high durability, high throughput design of Amazon S3 for backup

and as a common storage sink

• Durable sink between data services

• Supports de-coupling and asynchronous delivery

• Consider RRS for lower cost, lower durability storage of derivatives or copies

• Consider parallel threads and multipart upload for faster writes

• Consider parallel threads and range get for faster reads

Page 23: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Aggregate All Data in S3 Surrounded by a collection of the right tools

EMR Kinesis

Redshift DynamoDB RDS

Data Pipeline

Spark StreamingCassandra Storm

Amazon S3

Amazon S3

Page 24: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Fully-managed NoSQL database service

Built on solid-state drives (SSDs)

Consistent low latency performance

Any throughput rate

No storage limits

Amazon DynamoDB

Page 25: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

DynamoDB Concepts

attributes

items

table

schema-lessschema is defined per attribute

Page 26: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

DynamoDB: Access and Query Model

• Two primary key options• Hash key: Key lookups: “Give me the status for user abc”• Composite key (Hash with Range): “Give me all the status updates for user ‘abc’

that occurred within the past 24 hours”

• Support for multiple data types– String, number, binary… or sets of strings, numbers, or binaries

• Supports both strong and eventual consistency– Choose your consistency level when you make the API call– Different parts of your app can make different choices

• Global Secondary Indexes

Page 27: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

DynamoDB: High Availability and Durability

• Regional service• Synchronous replication to

three Availability Zones• Writes acknowledged only

when they are on disk in at least two Availability Zones

Page 28: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

What does DynamoDB handle for me?

• Scaling without down-time• Automatic sharding• Security inspections, patches, upgrades• Automatic hardware failover• Multi-AZ replication• Hardware configuration designed specifically for DynamoDB• Performance tuning

…and a lot more

Page 29: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Amazon DynamoDB Best Practices

• Keep item size small• Store metadata in Amazon DynamoDB and blobs in Amazon S3 • Use a table with a hash key for extremely high scale • Use hash-range key to model

– 1:N relationships– Multi-tenancy

• Avoid hot keys and hot partitions• Use table per day, week, month etc. for storing time series data• Use conditional updates

Page 30: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Relational Databases

Fully managed; zero admin

MySQL, PostgreSQL, Oracle & SQL Server

Amazon RDS

Page 31: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Process and Analyze

Page 32: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Processing Frameworks

• Batch Processing– Take large amount (>100TB) of cold data and ask questions– Takes hours to get answers back

• Stream Processing (real-time)– Take small amount of hot data and ask questions – Takes short amount of time to get your answer back

Page 33: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Processing Frameworks

• Batch Processing– Amazon EMR (Hadoop)– Amazon Redshift

• Stream Processing– Spark Streaming– Storm

Page 34: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Columnar data warehouse

ANSI SQL compatible

Massively parallel

Petabyte scale

Fully-managed

Very cost-effective

Amazon

Redshift

Page 35: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Amazon Redshift architecture

• Leader Node– SQL endpoint

– Stores metadata

– Coordinates query execution

• Compute Nodes– Local, columnar storage

– Execute queries in parallel

– Load, backup, restore via Amazon S3

– Parallel load from Amazon DynamoDB

• Hardware optimized for data processing

• Two hardware platforms– DW1: HDD; scale from 2TB to 1.6PB

– DW2: SSD; scale from 160GB to 256TB

10 GigE(HPC)

IngestionBackupRestore

SQL Clients/BI Tools

128GB RAM

16TB disk

16 cores

Amazon S3 / DynamoDB

JDBC/ODBC

128GB RAM

16TB disk

16 coresCompute Node

128GB RAM

16TB disk

16 coresCompute Node

128GB RAM

16TB disk

16 coresCompute Node

LeaderNode

Page 36: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Amazon Redshift Best Practices

• Use COPY command to load large data sets from Amazon S3, Amazon

DynamoDB, Amazon EMR/EC2/Unix/Linux hosts

– Split your data into multiple files

– Use GZIP or LZOP compression

– Use manifest file

• Choose proper sort key

– Range or equality on WHERE clause

• Choose proper distribution key

– Join column, foreign key or largest dimension, group by column

Page 37: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Hadoop/HDFS clusters

Hive, Pig, Impala, HBase

Easy to use; fully managed

On-demand and spot pricing

Tight integration with S3,

DynamoDB, and Kinesis

Amazon Elastic

MapReduce

Page 38: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

EMR Cluster

S3

1. Put the data into S3

2. Choose: Hadoop distribution, # of nodes, types of nodes, Hadoop

apps like Hive/Pig/HBase

4. Get the output from S3

3. Launch the cluster using the EMR console, CLI, SDK, or

APIs

How Does EMR Work?

Page 39: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

EMR

EMR Cluster

S3

You can easily resize the cluster

And launch parallel clusters using the same

data

How Does EMR Work?

Page 40: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

EMR

EMR Cluster

S3

Use Spot nodes to save time and money

How Does EMR Work?

Page 41: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

The Hadoop Ecosystem works inside of EMR

Page 42: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Amazon EMR Best Practices

• Balance transient vs persistent clusters to get the best TCO

• Leverage Amazon S3 integration– Consistent View for EMRFS

• Use Compression (LZO is a good pick)• Avoid small files (< 100MB; s3distcp can help!)• Size cluster to suit each job• Use EC2 Spot Instances

Page 43: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Amazon EMR Nodes and Size

• Tuning cluster size can be more efficient than tuning Hadoop code• Use m1 and c1 family for functional testing• Use m3 and c3 xlarge and larger nodes for production workloads• Use cc2/c3 for memory and CPU intensive jobs• hs1, hi1, i2 instances for HDFS workloads • Prefer a smaller cluster of larger nodes

Page 44: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Partners – Analytics (Scientific, algorithmic, predictive, etc)

Page 45: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Visualize

Page 46: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Partners - BI & Data Visualization

Page 47: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Putting All The AWS Data Tools Together & Architectural Considerations

Page 48: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

One tool to rule them all

Page 49: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Data Characteristics: Hot, Warm, Cold

Hot Warm Cold

Volume MB–GB GB–TB PBItem size B–KB KB–MB KB–TBLatency ms ms, sec min, hrsDurability Low–High High Very High

Request rate Very High High LowCost/GB $$-$ $-¢¢ ¢

Page 50: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Average latency

Datavolume

Item size

Request rate

Cost ($/GB/month)

Durability

Elasti-Cache

ms

GB

B-KB

Very High

Low -Moderate

$$

AmazonDynamoDB

ms

GB-TBs(no limit)

B-KB(64 KB max)

Very High

Very High

¢¢

AmazonRDS

ms.sec

High

High

¢¢

GB-TB(3 TB max)

KB(~rowsize)

CloudSearch

ms.sec

High

High

$

GB-TB

KB(1 MB max)

AmazonRedshift

sec.min

Low

High

¢

KB(64 K max)

TB-PB(1.6 PB max)

AmazonEMR (Hive)

sec.min,hrs

Low

High

¢

KB-MB

GB-PB(~nodes)

AmazonS3

Very High

¢

KB-GB(5 TB max)

GB-PB(no limit)

ms,sec,min (~size)

Low-Very High (no limit)

AmazonGlacier

Very High

¢

GB(40 TB max)

GB-PB(no limit)

hrs

Very Low(no limit)

Page 51: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Cost Conscious Design Example: Should I use Amazon S3 or Amazon DynamoDB?

“I’m currently scoping out a project that will greatly increase my team’s use of Amazon S3. Hoping you could answer some questions. The current iteration of the design calls for many small files, perhaps up to a billion during peak. The total size would be on the order of 1.5 TB per month…”

Request rate (Writes/sec)

Object size(Bytes)

Total size(GB/month)

Objects per month

300 2048 1483 777,600,000

Page 52: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Request rate (Writes/sec)

Object size(Bytes)

Total size(GB/month)

Objects per month

300 2,048 1,483 777,600,000 DynamoDB or S3?

Page 53: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Request rate (Writes/sec)

Object size(Bytes)

Total size(GB/month)

Objects per month

Scenario 1 300 2,048 1,483 777,600,000

Scenario 2 300 32,768 23,730 777,600,000

Amazon S3

Amazon DynamoDB

use

use

Page 54: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Lambda Architecture

Page 55: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Putting it all togetherDe-coupled architecture

• Multi-tier data processing architecture• Ingest & Store de-coupled from Processing• Ingest tools write to multiple data stores • Processing frameworks (Hadoop, Spark, etc.) read from data stores• Consumers can decide which data store to read from depending on

their data processing requirement

Page 56: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

SparkStreaming /

Storm

Redshift

Impala Spark

EMR/Hadoop

Redshift

EMR/Hadoop

Spark

Kinesis/Kafka

NoSQL / DynamoDB / Hadoop HDFS S3Data

Hot ColdData Temperature

Latency

Low

HighAnswers

Page 57: AWS Webcast - Managing Big Data in the AWS Cloud_20140924
Page 58: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Customer Use Cases

Page 59: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Autocomplete Search RecommendationsAutomatic spelling corrections

Page 60: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

A look at how it works

Months of user history Common misspellings

Data Analyzed Using EMR:

WestenWistin

WestanWhestin

Automatic spelling corrections

Page 61: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Months of user search data

Search terms

Misspellings

Final click throughs

Yelp web site log data goes into Amazon S3

Amazon S3

Page 62: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Amazon Elastic MapReduce spins up a 200 node Hadoop cluster

Hadoop Cluster

Amazon EMRAmazon S3

Page 63: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Hadoop Cluster

Amazon EMRAmazon S3

All 200 nodes of the cluster simultaneously look for common misspellings

Westen

Wistin

Westan

Page 64: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Hadoop Cluster

Amazon EMRAmazon S3

A map of common misspellings and suggested corrections are loaded back into Amazon S3.

Westen

Wistin

Westan

Page 65: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Then the cluster is shut down Yelp only pays for the time they used it

Hadoop Cluster

Amazon EMRAmazon S3

Page 66: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Each of Yelp’s 80 Engineers Can Do This Whenever They Have a Big Data Problem

spins up over 250 Hadoop clusters per

week in EMR.

Amazon EMRAmazon S3

Page 67: AWS Webcast - Managing Big Data in the AWS Cloud_20140924
Page 68: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Data Innovation Meets Action at Scale at NASDAQ OMX

• NASDAQ’s technology powers more than 70 marketplaces in 50 countries

• NASDAQ’s global platform can handle more than 1 million messages/second at a

median speed of sub-55 microseconds

• NASDAQ own & operate 26 markets including 3 clearinghouse & 5 central securities repositories

• More than 5,500 structured products are tied to NASDAQ’s global indexes with the notional value of at least $1 trillion

• NASDAQ powers 1 in 10 of the world’s securities transactions

Page 69: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

NASDAQ’s Big Data Challenge

• Archiving Market Data– A classic “Big Data” problem

• Power Surveillance and Business Intelligence/Analytics

• Minimize Cost– Not only infrastructure, but development/IT labor costs too

• Empower the business for self-service

Page 70: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

1 2 3 4 5 6 7 8 90

100,000,000200,000,000300,000,000400,000,000500,000,000600,000,000

NASDAQ Exchange Daily Peak Messages

MarketDataIs BigDataCharts courtesy of the Financial Information Forum

Page 71: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

NASDAQ’s Legacy Solution

• On-premises MPP DB– Relatively expensive, finite storage– Required periodic additional expenses to add more storage– Ongoing IT (administrative) human costs

• Legacy BI tool– Requires developer involvement for new data sources, reports,

dashboards, etc.

Page 72: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

New Solution: Amazon Redshift

• Cost Effective– Redshift is 43% of the cost of legacy

• Assuming equal storage capacities

– Doesn’t include IT ongoing costs!

• Performance– Outperforms NASDAQ’s legacy BI/DB solution– Insert 550K rows/second on a 2 node 8XL cluster

• Elastic– NASDAQ can add additional capacity on demand, easy to grow their cluster

Page 73: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

• Amazon Redshift partner– http://aws.amazon.com/redshift/partn

ers/pentaho/

• Self Service– Tools empower BI users to integrate

new data sources, create their own analytics, dashboards, and reports without requiring development involvement

• Cost effective

New Solution: Pentaho BI/ETL

Page 74: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Net Result

• New solution is cheaper, faster, and offers capabilities that NASDAQ didn’t have before– Empowers NASDAQ’s business users to explore data like they never

could before– Reduces IT and development as bottlenecks– Margin improvement (expense reduction and supports business

decisions to grow revenue)

Page 75: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

NEXT STEPS

Page 76: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

AWS is here to help

Solution Architects

Professional Services

Premium Support

AWS Partner Network (APN)

Page 77: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

aws.amazon.com/partners/competencies/big-data

Partner with an AWS Big Data expert

Page 78: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

http://aws.amazon.com/marketplace

Big Data Case Studies

Learn from other AWS customers

aws.amazon.com/solutions/case-studies/big-data

Page 79: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

AWS Marketplace

AWS Online Software Store

aws.amazon.com/marketplace

Shop the big data category

Page 80: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

http://aws.amazon.com/marketplace

AWS Public Data Sets

Free access to big data sets

aws.amazon.com/publicdatasets

Page 81: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

AWS Grants Program

AWS in Education

aws.amazon.com/grants

Page 82: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

AWS Big Data Test Drives

APN Partner-provided labs

aws.amazon.com/testdrive/bigdata

Page 83: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

https://aws.amazon.com/training

AWS Training & Events

Webinars, Bootcamps, and Self-Paced Labs

aws.amazon.com/events

Page 84: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

Big Data on AWS

Course on Big Data

aws.amazon.com/training/course-descriptions/bigdata

Page 85: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

reinvent.awsevents.com

Page 86: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

aws.amazon.com/big-data

Page 87: AWS Webcast - Managing Big Data in the AWS Cloud_20140924

[email protected]

Thank You!


Recommended