cse6331: cloud computingthe three v’s of big data: volume, variety, and velocity high volume (too...

25
CSE6331: Cloud Computing Leonidas Fegaras University of Texas at Arlington c 2016 by Leonidas Fegaras Cloud Computing Fundamentals Based on: J. Freire’s class notes on Big Data http://vgc.poly.edu/ ~ juliana/courses/BigData2016/ D. Abadi: Data Management in the Cloud: Limitations and Opportunities. IEEE Data Eng. Bull. 2009 VMWare white paper: Virtualization Overview

Upload: others

Post on 08-Jul-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

CSE6331: Cloud Computing

Leonidas Fegaras

University of Texas at Arlingtonc©2016 by Leonidas Fegaras

Cloud Computing Fundamentals

Based on:J. Freire’s class notes on Big Data http://vgc.poly.edu/~juliana/courses/BigData2016/

D. Abadi: Data Management in the Cloud: Limitations and Opportunities. IEEE Data Eng. Bull. 2009VMWare white paper: Virtualization Overview

Page 2: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Computing as a Utility

Cloud computing is a model for enabling convenient, on-demandnetwork access to a shared pool of configurable computing resources:

network bandwidthdata storageapplicationsservices

these resources can be rapidly provisioned and released with minimalmanagement effort or service provider interaction

Key idea:

Move computing into large shared data centersGives the illusion of infinite resources on demandPay-as-you-go (as needed), no upfront costElasticity: scale up or down, based on current needsAutomatically manages fault tolerance, redundancy, storage, loadbalancing, geographic distribution, security, software updates, etc

CSE6331: Cloud Computing Cloud Computing Fundamentals 2

Page 3: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Why?

Sharing resources among applications utilizes these resources better

While the resource needs of an application may vary dramatically, thesum of the resource needs of many independent applications variesmuch less

Upfront setup cost is amortized over many applications

Cost is proportional to use

Improved service levels, fault tolerance, and availability

Rapid prototyping and market testing

Better security infrastructure, better protection against networkattacks

Allows self-service deployment

CSE6331: Cloud Computing Cloud Computing Fundamentals 3

Page 4: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Good for...

Web hosting

Deployment of data-intensive web services

especially those with unpredictable load... that require high availability

Storing massive amounts of data (backups, videos, logs, etc)

Off-line analysis of massive data (map-reduce, Spark)

Large computing jobs on demand

jobs that require many CPUs

CSE6331: Cloud Computing Cloud Computing Fundamentals 4

Page 5: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Alternatives

HPC (High-Performance Computing)

Distributed Computing

Distributed Databases

Grid computing

Peer-to-peer computing

BitTorrent, multi-user games, Skype

CSE6331: Cloud Computing Cloud Computing Fundamentals 5

Page 6: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Current Trends

Cloud providers: Amazon, Google, Microsoft, ...

they provide infrastructure and/or deployment platformthey offer them as web services and are billed on a utility basis

Amazon web services

Simple Storage Service (SS3)Elastic Compute Cloud (EC2)Elastic MapReduce

Google App Engine

plus many office applications (Google docs), email etc

CSE6331: Cloud Computing Cloud Computing Fundamentals 6

Page 7: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Challenges

Data security

Have to depend on the security provided by the cloud hostData protection against other applicationsProtection against the cloud hostA cloud is a high value target to hackersData ownership issuesMoving sensitive data to the cloud

Performance variability

Lack of Quality of Service (QoS) guarantees

Data transfer bottlenecks

Financial cost

CSE6331: Cloud Computing Cloud Computing Fundamentals 7

Page 8: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Service Offering Models

Software as a Service (SaaS)

applications running on a cloud infrastructureusers: any end usereg, Google mail

Platform as a Service (PaaS)

deploy onto the cloud infrastructure user-defined applications createdusing a set of provided libraries, services, and toolsusers: application developerseg, Google App Engine

Infrastructure as a Service (IaaS)

deploy and run arbitrary software, including operating systems andapplications; share computing resources, storage, networkusers: IT architects, data analystseg, Amazon Web Services (AWS)

CSE6331: Cloud Computing Cloud Computing Fundamentals 8

Page 9: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Virtualization

The underlying core technology of cloud computing

The physical infrastructure owned by the service provider is sharedamong many users, increasing the resource utilization

Partitioning: may run multiple operating systems on a single physicalsystem and share the underlying hardware resources

Benefits: isolation of virtual machines and hardware-independence

Virtual machines are highly portable

Hypervisor: software that provides virtual partitioning capabilities

bare metal approach: runs directly on hardware

CSE6331: Cloud Computing Cloud Computing Fundamentals 9

Page 10: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Virtualization OverviewSlide credit: VMWare white paper: Virtualization Overview

CSE6331: Cloud Computing Cloud Computing Fundamentals 10

Page 11: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Virtualization (cont.)

Hosted Virtualization

The virtualization software relies on the host OS to provide the servicesto talk directly to the underlying hardwareThe guest OS cannot communicate directly with the underlyingphysical infrastructureThe hypervisor interacts directly with the host’s hardware resourcesand acts as a platform for the guest OSVirtualBox

Paravirtualization

The guest OS is aware that it is running within a virtualizedenvironment, and has been modified to exploit thisVirtualBox guest editions

CSE6331: Cloud Computing Cloud Computing Fundamentals 11

Page 12: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Virtualization (cont.)

Bare-Metal (Hypervisor) Virtualization

No host OSThe guest OS access the underlying hardware directly through an APIMore efficient than hosted architectures: greater scalability, robustness,and performanceVMWare, Xen

Hardware-Assisted Virtualization

Current CPUs support virtualization through hardware extensionsIntel VT-x, and AMD AMD-v

CSE6331: Cloud Computing Cloud Computing Fundamentals 12

Page 13: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Virtualization ArchitectureSlide credit: VMWare white paper: Virtualization Overview

CSE6331: Cloud Computing Cloud Computing Fundamentals 13

Page 14: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Virtualization in a Data CenterSlide credit: VMWare white paper: Virtualization Overview

Transforms farms of individual servers, storage, and networking into a poolof computing resources

CSE6331: Cloud Computing Cloud Computing Fundamentals 14

Page 15: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Large Scale Computing

Distributed systemsParallel or distributed computations on multiple disks and nodes

often using message passing interface (MPI)and parallel virtual machines (PVM)

Data are stored in a central location and copied to processing nodeswhen neededBrings the data to the computationGood for computationally intense tasks

Cloud Computing

Run on clusters of share-nothing commodity computersUse distributed storageThe data are already distributed when they are storedDistributed computation: run computations where the data areBrings the computation to the dataGood for data intense tasks

CSE6331: Cloud Computing Cloud Computing Fundamentals 15

Page 16: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

What is Big Data?

The three V’s of big data: Volume, Variety, and Velocity

High Volume (too big): petabyte-scale collections

High Variety (too hard): data comes from a variety of sources indifferent formats

irregular, unstructured, or incomplete dataData need to be cleaned, processed, and integrated

High Velocity (too fast): data need to be processed quickly(throughput) with low latency

Batch streaming, Real-time analytics

Also veracity: correctness and accuracy of data

CSE6331: Cloud Computing Cloud Computing Fundamentals 16

Page 17: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

High Demand for Data Scientists

CSE6331: Cloud Computing Cloud Computing Fundamentals 17

Page 18: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

How Big?

From Lin and Dyer, 2010:

Google:

grew from processing 100 TB of data a day with MapReduce in 2004to processing 20 PB a day with MapReduce in 2008

Facebook:

1.1 billion users the site each month2.5 PB of user data, growing at about 15 TB per day

Twitter:

517 million accounts, 250 million tweets/day

eBay’s two enormous data warehouses:

one with 2 PB of user data, and the other with 6.5 PB of user dataspanning 170 trillion records and growing by 150 billion new recordsper day

CSE6331: Cloud Computing Cloud Computing Fundamentals 18

Page 19: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Data-intensive Scientific Computing

Current scientific applications must analyze enormous amounts ofdata using complex mathematical data processing methods

Necessary for scientific breakthroughs

Capable of generating several petabytes of data per

Large Hadron Collider at CERNAstronomy’s Pan-STARRS5 array of celestial telescopesSloan Sky Survey, genome, climate data, oceanography

CSE6331: Cloud Computing Cloud Computing Fundamentals 19

Page 20: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Big Data is not New, but has Become More Important

More data are now widely available

Web, scientific data, social network data, etc

Increased variety

logs, XML, JSON, ...

Many more companies try to collect data and turn them into value

targeted advertising

Computing is cheaper and easier to access

server with 64 cores, 512GB RAM $11kcluster with 1000 cores $150kcloud computing: Amazon EC2, Microsoft Azure

Very large data centers

Better virtualization technology

Broadband Internet

CSE6331: Cloud Computing Cloud Computing Fundamentals 20

Page 21: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Data Management in the Cloud

Not all Big Data management applications are well-suited for deploymentin the cloud:

Traditional transactional databases (OLTP): NO

Good for both read- and write-intensive applicationsDo not normally use shared-nothing architecturesHard to maintain ACID guarantees in the face of data replicationThere are enormous risks in storing transactional data on an untrustedhost

Data analysis systems (OLAP): YES

Tend to be read-only with occasional batch insertsShared-nothing architecture is a good matchData analysis workloads consist of many large scans and aggregationsACID guarantees are typically not neededEasy to parallelizeSensitive data can be left out of the analysis

CSE6331: Cloud Computing Cloud Computing Fundamentals 21

Page 22: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Wish List

ScalabilityScan 100 TB on 1 node @ 50 MB/sec = 23 daysScan 100 TB on 1000-node cluster = 33 minutesUse divide-and-conquer: data partitioning

Yahoo! runs a 4000 node Hadoop clusterOverall, there are 38,000 nodes running Hadoop at Yahoo! (2009)

Data analysis is elastic, but only if workload is parallelizable

Fault toleranceFault tolerant read-only query: does not have to restart it if one of theprocessing nodes failsComplex queries can involve thousands of nodes and can take hours tocompleteClouds are typically built on top of cheap, commodity hardware, forwhich failure is not uncommonThe probability of a failure occurring during a long-running dataanalysis task is relatively highTo contrast: parallel database systems restart a query upon a failure

Google reports an average of 1.2 failures per analysis job

CSE6331: Cloud Computing Cloud Computing Fundamentals 22

Page 23: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Wish List (cont.)

Performance: optimize resources

Since you pay for resourcesEspecially, minimize the number of nodes

Ability to run in a heterogeneous environment

Performance of cloud compute nodes is often not consistentThe time to complete a query will be equal to time for the slowestcompute node to complete its assigned taskHow do we tell if a node is slow or faulty?

Ability to operate on encrypted data

Ability to work with other data analysis software

CSE6331: Cloud Computing Cloud Computing Fundamentals 23

Page 24: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

Big Data: New Technologies

Infrastructure

New computing paradigms: Hadoop MapReduce, SparkNew storage solutions: NoSQL, column stores, Big TableNew languages: Hive, JAQL, Pig LatinWe will study these and how they relate to previous technologies

Algorithms and techniques

New infrastructure demands new approaches to explore dataWe will study algorithms to process and explore data in Big-Dataenvironments

CSE6331: Cloud Computing Cloud Computing Fundamentals 24

Page 25: CSE6331: Cloud ComputingThe three V’s of big data: Volume, Variety, and Velocity High Volume (too big): petabyte-scale collections High Variety (too hard): data comes from a variety

NoSQL Systems

Stands for: Not Only SQL

They don’t use SQL

No schema

Simple and scalable data management systems

No “one size fits all” philosophy

Most are open-source

For clusters/cloud

Various types:

Key-value storesDocument storesExtensible record stores

CSE6331: Cloud Computing Cloud Computing Fundamentals 25