vorlesung 12 - p2p streaming.ppt [kompatibilitätsmodus]

24
L3S Research Center, University of Hannover Peer Peer to to Peer Streaming Peer Streaming Peer Peer- -to to- -Peer Streaming Peer Streaming Wolf Wolf- -Tilo Tilo Balke Balke and Wolf Siberski and Wolf Siberski 23.1.2008 23.1.2008 1 Peer-to-Peer Systems and Applications, Springer LNCS 3485 *with slides from Kan-Leung Cheng (U Maryland), Darshan Purandare (UCF), Hui Zhang (CMU), Long Vu (UIUC) Overview 1. IP-Level vs Application-Level Streaming 2. P2P Streaming Basics 3. PPLive 4. Summary 2 P2P Streaming L3S Research Center

Upload: others

Post on 12-Jan-2022

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

L3S Research Center, University of Hannover

PeerPeer toto Peer StreamingPeer StreamingPeerPeer--toto--Peer StreamingPeer Streaming

WolfWolf--TiloTilo BalkeBalke and Wolf Siberskiand Wolf Siberski

23.1.200823.1.2008

1Peer-to-Peer Systems and Applications, Springer LNCS 3485Peer-to-Peer Systems and Applications, Springer LNCS 3485

*with slides from Kan-Leung Cheng (U Maryland), Darshan Purandare (UCF), Hui Zhang (CMU), Long Vu (UIUC)

Overview

1. IP-Level vs Application-Level Streaming

2. P2P Streaming Basics

3. PPLive

4. Summary

2P2P StreamingL3S Research Center

Page 2: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Multicast Internet Applications

Multi-party applicationsVideo on Demand and IPTV

Audio/video conferencing

Multi-party games

Distributed simulation

Broadcast of web cams

Consider a world with ...Tens of millions of simultaneously running multi-point applications

3P2P StreamingL3S Research Center

applications

Each application with tens to several thousand of end points

Multi-unicast vs. IP Multicast

IP MulticastUnicast

4P2P StreamingL3S Research Center

Page 3: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

IP Multicast Overview

Seminal work by Steve Deering in 1989Huge amount of follow-on workResearch

1000s papers on multicast routing reliable multicast multicastcongestion1000s papers on multicast routing, reliable multicast, multicastcongestioncontrol, layered multicastSIGCOMM, ACM Multimedia award papers, ACM Dissertation Award

Standard: IPv4 and IPv6, DVMRP/CBT/PIMDevelopment: in both routers (Cisco etc) and end systems (Microsoft, all versions of Unix)Deployment: Mbone, major ISP’s

5P2P StreamingL3S Research Center

p y jApplications: vic/vat/rat/wb…

Situation todayStill not used across the Internet Reasons?

Router state issue

How to tell a packet is a multicast packet? Each group needs a group address

How to decide where and how to branch?

6P2P StreamingL3S Research Center

routing protocol needs to set up per group state at routers

Violates stateless packet forwarding principleCurrently IP layer only maintains routing state

Highly aggregated

140K routing entries today for hundreds of millions hosts

Page 4: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Ack Explosion

Large number of acknowledgements required

End-to-end ack Router-based ack

7P2P StreamingL3S Research Center

End-to-end acknowledgments inefficient

Router-based acknowledgments overloads routersrequires even larger large state maintenance

IP Multicast Issues – Summary

Poor routing scalabilityRouters need to keep per group/connection state

Violation of fundamental Internet architecture principle

Difficult to support higher functionalitiesError control, flow control, congestion control

Security concernsaccess control, both senders and receivers

Denial of Service attacks

8P2P StreamingL3S Research Center

Page 5: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

IP Architecture

“Dumb” IP layer minimal functionalities for connectivity

Unicastaddressing, forwarding, routing

Smart end systemtransport layer or application performs more sophisticated functionalities

flow control, error control, congestion control

Advantages

9P2P StreamingL3S Research Center

Advantagesaccommodate heterogeneous technologies

support diverse applications and decentralized network administration “Hourglass” model

Multicast revisited

Can we achieve efficient multi-point delivery

without support from the IP layer?

10P2P StreamingL3S Research Center

Page 6: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Application Layer Multicast (ALM)

Do stream distribution on application levelMulticasting implemented at end hosts instead of network routers

Nodes form unicast channels or tunnels between them

Use default IP infrastructure

Receivers form self-organized network (peer-to-peer principle)

S

R1 R2

E1

Unicast

Unicast

Unicast

11P2P StreamingL3S Research Center

R1 R2

E2 E3

Unicast

P2P Media Streaming

Media streaming extremely expensive1 hour of video encoded at 300Kbps = 128.7 MB

Serving 1000 users would require 125.68 GB

Approach: same idea as in P2P file sharing:Peers form overlay network

Nodes offer their uplink bandwidth while downloading and viewing the media content

Takes load off the server

12P2P StreamingL3S Research Center

Takes load off the server

Scalable

Page 7: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Peer-to-Peer Streaming Benefits

Easy to deployNo change to network infrastructure

Programmable end-hostsOverlay construction algorithms at end hosts can be easily applied

Application-specific customizationsNetwork structure

Packet forwarding strategies

13P2P StreamingL3S Research Center

Challenges

Need to playback the media in real timeQuality of Service

Procure future media stream packetsNeeds reliable neighbors and effective management

High “churn” rate – Users join and leave in betweenNeeds robust network topology to overcome churn

Internet dynamics and congestion in the interior of the network

14P2P StreamingL3S Research Center

Degrades QoS

Fairness policies extremely difficult to applyHigh bandwidth users have no incentive to contribute

Tit-for-tat doesn’t work due to asymmetry

Page 8: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Peer-to-Peer Streaming Models

Media content is broken down in small pieces and disseminated through the network

Push model: content forwarded as soon as it arrives

Pull model: nodes request missing pieces

Neighboring nodes use Gossip protocol to exchange buffer information

Nodes trade unavailable pieces

R b t a d calable b t e dela

15P2P StreamingL3S Research Center

Robust and scalable, but more delay

Network efficiency

Optimization GoalsDelay between source and receivers should be small

Relative Delay Penalty (RDP)

Number of redundant packets on any physical link should be lowPhysical Link Stress (PLS)

CMU

Stan1

Stan2Stan2CMU

Stan1

Stan2 CMU

Stan1

16P2P StreamingL3S Research Center

Gatech

“Efficient” overlay

Berk2

Berk1Berk1

High degree (unicast)

Berk2

Gatech

High latency

Berk2

GatechBerk1

Page 9: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Physical Link Stress (PLS)

PLS is given by the number of identical copies of a packet that traverse a physical link

Indicates the bandwidth inefficiency

S

R1 R2

E1Example:

PLS for link S-R1 is 2.

Average PLS is 7/5.

17P2P StreamingL3S Research Center

E2 E3

Relative Delay Penalty (RDP)

RDP is given by the ratio of the delay in the overlay and the delay in the direct unicast path.

Indicates the delay inefficiency

S

R1 R2

E1

20 ms

10 ms 10 ms

Example:Overlay delay for the path from S to E3 is 60 ms.

Unicast delay is 40 ms.

Therefore the RDP for E3

18P2P StreamingL3S Research Center

E2 E3

10 ms 10 msTherefore, the RDP for E3is 1.5 ( = 60 ms / 40 ms).

Page 10: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Network topologies

Tree BasedContent flows from server to nodes in a tree like fashion

O i t f f il f l t btOne point of failure for a complete subtree

High recovery time

Multiple trees for increased robustness

Mesh BasedOvercomes tree based flaws

19P2P StreamingL3S Research Center

Overcomes tree based flaws

Nodes maintain state information of many neighbors

High control overhead

Streaming Topologies: Tree

Tree construction based upon minimal delays

Permanent peer monitoring for tree maintenance and repair

In case individual links/path fail or become congested

In case receivers join and leave (churn)In case receivers join and leave (churn)

Push-based delivery from the source(s) to the receivers

Data forwarded along the tree with minimal delay

Multiple trees (=multiple root nodes)

Provide redundant delivery paths against failures (churn)

Provide complementary data flows to each receiver

20P2P StreamingL3S Research Center

Issues

Maintaining an “optimal” tree structure incurs a lot of overhead

Particularly in conjunction with churn

Churn may also cause disruptions to downstream receivers

Page 11: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Tree-based System: End System Multicast (ESM)

First application level streaming systemDeveloped at CMU by Hui Zhang et al. (2002)

ObjectivesObjectivesSelf-organizing: adapts to dynamic membership changes

Self-improving: automatically evolves into efficient overlays

Two versions of protocolMulti-source, smaller scale conferencing apps

21P2P StreamingL3S Research Center

Single source, larger scale broadcasting apps

Tree-based, Push model

ESM Node Join

Bootstrapping process (for node X)Connect to source (S)

Get a subset of group membership

P t l ti l ithParent selection algorithmSend probe message to known nodes

Decision criteria for parent node (P)Filter out P if it is a descendant of X

Performance of P

Delay of path from S to P

Saturation level of P

22P2P StreamingL3S Research Center

Saturation level of P

Performance of link P-X

Delay of link P-X

TCP bandwidth of link P-X

?

Page 12: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

ESM Tree Maintenance

Build separate control structure decoupled from treeEach member knows small random subset of group members

Information maintained using gossip-like algorithm

Members also maintain path from source

Continouosly apply parent selection strategyReact on nodes becoming unavailable

Repeat probing regularly

I b d idth

23P2P StreamingL3S Research Center

Improve bandwidth usage

Improve clustering

Multiple Tree Construction

… …Video stream

24P2P StreamingL3S Research Center

Page 13: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Join Procedure

Initial join

Contact video source

Receives peer list, number of trees

… …Video stream

Probe peers

Connect to multicast trees

25P2P StreamingL3S Research Center

Disconnect / Rejoin Procedure

3 treesParent of yellow tree is down

Parent leave is detected

Retransmissions requested

Yellow tree is recovered

Yellow tree is down?

26P2P StreamingL3S Research Center

Page 14: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Streaming Topologies: Mesh

Mesh construction and maintenance similar to TreeFind set of peers with minimal delays

Choose subset of the peers initially provided via bootstrapping

Gossiping protocols to learn about further peersGossiping protocols to learn about further peers

Continuous optimization of neighbor set

Active pulling of media segments from peersExchange of buffer maps (who has which data)

Explicit requests for missing chunks (receiver-controlled)

Kept locally available for forwarding to other peers

Similar to BitTorrent, but needs to consider time-constraints

27P2P StreamingL3S Research Center

Similar to BitTorrent, but needs to consider time constraints

IssuesLots of control traffic (explicit pull)

Higher end-to-end delay (due to buffering and pull-based forwarding)

CoolStreaming

X. Zhang, J. Liu, B. Li, and T.-S. Peter Yum: “CoolStreaming/DONet:

A data-driven overlay network for efficient live media streaming”

(IEEE INFOCOM, 2005)

Joining Node obtains a list of 40Joining Node obtains a list of 40 nodes from the source

Each node contacts these nodes for media content

Extend list of known nodes via gossiping

In steady state, every node has typically 4-8 active neighbors

28P2P StreamingL3S Research Center

Real world deployed and highly successful system

Stopped in 2005, due to copyright issues

Successor: PPLive

Page 15: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Cooltreaming distribution algorithm

Stream is chopped by server and disseminated

Each node periodically shares its buffer content map

with neighbors

Request segments as part of this exchange message

Reply strategySend scarce packages first (like BitTorrent)

If no package is scarce: send to peer with highest bandwidth first

29P2P StreamingL3S Research Center

L3S Research Center, University of Hannover

Quality Criteria

30Peer-to-Peer Systems and Applications, Springer LNCS 3485

Page 16: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Quality Criteria

Quality of ServiceJitter less transmissionLow end to end latency

Network efficiencyNetwork efficiencyUplink utilization

High uplink throughput leads to scalable P2P systems

Robustness and ReliabilityChurn, node failure or departure should not affect QoS

ScalabilityFairness

31P2P StreamingL3S Research Center

FairnessDetermined in terms of content served (Share Ratio)No user should be forced to upload much more than what it has downloaded

Quality of Service

QoS is the most important metric

Jitter: Unavailability of stream content at play time causes jitter

Jitterless transmission ensures good media playback

Continuous supply of stream content ensures no jitters

Latency: Difference in time between playback at server and user

Lower latency keeps users interested: a live event (e.g. soccer match) can lose importance in crucial moments if the transmission is delayed

32P2P StreamingL3S Research Center

can lose importance in crucial moments, if the transmission is delayed

Reducing hop count reduces latency

Page 17: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Uplink Utilization

Uplink is the most sparse and important resource in the network

Sum of uplinks of all nodes is the load taken off the server

Utilization = (Uplink used / Uplink Available)Needs effective node organization and topology to maximize uplink utilization

High uplink throughput means more bandwidth in the network and hence leads to scalable P2P systems

33P2P StreamingL3S Research Center

Scalability

Serve as many users as possible with an acceptable level

of QoS

Increasing number of nodes should not degrade QoS

An effective overlay node topology and high uplink

throughput ensures scalable systems

34P2P StreamingL3S Research Center

•34

Page 18: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Fairness

Measured in terms of content served to the networkShare Ratio = (Uploaded Volume / Downloaded Volume)

Randomness in network causes high disparityMany nodes upload huge volume of content

Many nodes get a free ride with no or very little contribution

Must have an incentive for an end user to contributeP2P file sharing system like BitTorrent use tit-for-tat policy to combat free riding

35P2P StreamingL3S Research Center

free riding

Not easy to use it in streaming as nodes procure pieces in real time and applying tit-for-tat can cause delays

L3S Research Center, University of Hannover

Existing Systems

36Peer-to-Peer Systems and Applications, Springer LNCS 3485

Page 19: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Overview of Existing Systems

Most noted approach in recent years: CoolStreaming

PPLive and SOPCast are derivates of CoolStreaming

37P2P StreamingL3S Research Center

PPLive and SOPCast are derivates of CoolStreaming

Proprietary and working philosophy not published

Reverse engineered and measurement studies released

PPLive Overview

One of the largest deployed P2P multimedia streaming systems

D l d i ChiDeveloped in China

Hundred thousands of simultaneous viewers

38P2P StreamingL3S Research Center

Page 20: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

PPLive Membership Protocol

Client

39P2P StreamingL3S Research Center

An overlay

PPLive analysis

Long Vu et al.: “Measurement and Modeling of a Large-scale Overlay for Multimedia Streaming”, QShine 2007

Analyzed:Analyzed:Channel size variation

Node degree

Overlay randomness

Node availability

Session length

40P2P StreamingL3S Research Center

Page 21: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Channel Size Varies over a day

41P2P StreamingL3S Research Center

Huge popularity variationPeaks at noon and night

Higher dynamics than P2P file sharing

Node Degree

Average noded l fdegree scale-free

42P2P StreamingL3S Research Center

Degree independent of channel size

Similar to P2P file sharing

Page 22: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

PPLive Peers are Impatient

50% sessions are less than 10

i tminutes

43P2P StreamingL3S Research Center

Short sessions (probably channel hopping)

Different from P2P file sharing

Conclusions

Characteristics of PPLive and P2P file sharing are different

Higher variance in item popularity over time

Shorter average session durationMuch higher network churn

44P2P StreamingL3S Research Center

Page 23: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

L3S Research Center, University of Hannover

Summary

45Peer-to-Peer Systems and Applications, Springer LNCS 3485

Current Issues

High buffering time for P2P streamingHalf a minute for popular streaming channels and around 2 minutes for less popular

Some nodes lag with their peers by more than 2 minutes Some nodes lag with their peers by more than 2 minutes in playback time

Better peering strategy needed

Uneven distribution of uplink bandwidths (unfairness)No consideration of duplicate packets

Huge volumes of cross ISP traffic

46P2P StreamingL3S Research Center

ISPs use bandwidth throttling to limit bandwidth usageDegrade QoS perceived at used end

Sub-optimal uplink utilization

Page 24: Vorlesung 12 - P2P Streaming.ppt [Kompatibilitätsmodus]

Comparison to P2P File Sharing

SimilaritiesDistribution costs move from stream provider to network provider

Need incentives for end-users to contribute resources

Scalability needs uniform usage of link capacities (content replication proportional to popularity)

DifferencesQoS constraints essential for streaming

No time consuming strategies against free riding possible

47P2P StreamingL3S Research Center

No time-consuming strategies against free-riding possible

Much higher churn due to huge fraction of short sessions

Much smaller number of shared items

Conclusion

P2P streaming efficient way to realize application level multicast

Considers heterogeneous nodes

Conforms to IP network hourglass model

Self-optimizing

Widely used P2P applicationPPLive: 75 million global installed base and 20 million monthly active users (in 2007)

48P2P StreamingL3S Research Center