(powerpoint slides)

31
1 Measurement Study of Measurement Study of Low-bitrate Internet Low-bitrate Internet Video Streaming Video Streaming Dmitri Loguinov Dmitri Loguinov Hayder Radha Hayder Radha City University of New York and Philips Research City University of New York and Philips Research

Upload: ronny72

Post on 27-May-2015

155 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: (Powerpoint slides)

1

Measurement Study of Measurement Study of Low-bitrate Internet Video Low-bitrate Internet Video StreamingStreaming

Dmitri LoguinovDmitri LoguinovHayder RadhaHayder Radha

City University of New York and Philips ResearchCity University of New York and Philips Research

Page 2: (Powerpoint slides)

2

MotivationMotivation

• Market research by Telecommunications Reports International (TRI) from August 2001

• The report estimates that out of 70.6 million households in the US with Internet access, an overwhelming majority use dialup modems (Q2 of 2001)

52.2

1.2 0.1

9.13.14.9

0

15

30

45

60

Paid dialupISPs

Free ISPs Cable DSL Internet TV Satellite

ho

use

ho

lds

(mil

lio

n)

Page 3: (Powerpoint slides)

3

Motivation (cont’d)Motivation (cont’d)

• Consider broadband (cable, DSL, and satellite) Internet access vs. narrowband (dialup and webTV)

• 89% of households use modems and 11% broadband

11.55%

88.45%

0%

20%

40%

60%

80%

100%

dialup broadband

su

bs

cri

be

rs (

pe

rce

nt)

Page 4: (Powerpoint slides)

4

Motivation (cont’d)Motivation (cont’d)

• Customer growth in Q2 of 2001:– +5.2% dialup

– +0.5% cable modem

– +29.7% DSL

– Even though DSL grew fast in early 2001, the situation is now different with many local companies going bankrupt and TELCOs increasing the prices

• A report from Cahners In-Stat Group found that despite strong forecasted growth in broadband access, most households will still be using dial-up access in the year 2005 (similar forecast in Forbes ASAP, Fall 2001, by IDC market research)

• Furthermore, 30% of US households state that they still have no need or desire for Internet access

Page 5: (Powerpoint slides)

5

Motivation (cont’d)Motivation (cont’d)

• Our study was conducted in late 1999 – early 2000

• At that time, even more Internet users connected through dialup modems, which sparked our interest in using modems as the primary access technology for out study

• However, even today, our study could be considered up-to-date given the numbers from recent market research reports

• In fact, the situation is unlikely to change until year 2005

• Our study examines the performance of the Internet from the angle of an average Internet user

Page 6: (Powerpoint slides)

6

Motivation (cont’d)Motivation (cont’d)

• Note that previous end-to-end studies positioned end-points at universities or campus networks with typically high-speed (T1 or faster) Internet access

• Hence, the performance of the Internet perceived by a regular home user remains largely undocumented

• Rather than using TCP or ICMP probes, we employed real-time video traffic in our work, because performance studies of real-time streaming in the Internet are quite scarce

• The choice of NACK-based flow control was suggested by RealPlayer and Windows MediaPlayer, which dominate the current Internet streaming market

Page 7: (Powerpoint slides)

7

Overview of the TalkOverview of the Talk• Experimental Setup

• Overview of the Experiment

• Packet Loss

• Underflow Events

• Round-trip Delay

• Delay Jitter

• Packet Reordering

• Asymmetric Paths

• Conclusion

Page 8: (Powerpoint slides)

8

Experimental SetupExperimental Setup• MPEG-4 client-server real-time streaming architecture

• NACK-based retransmission and fixed streaming bitrate (i.e., no congestion control)

• Stream S1 at 14 kb/s (16.0 kb/s IP rate), Nov-Dec 1999

• Stream S2 at 25 kb/s (27.4 kb/s IP rate), Jan-May 2000

National dial-up ISP

UUNET (MCI)

Internet

server

T1 link

client

campus network

modem link

Page 9: (Powerpoint slides)

9

OverviewOverview• Three ISPs (Earthlink, AT&T WorldNet, IBM Global Net)

• Phone database included 1,813 dialup points in 1,188 cities

• The experiment covered 1,003 points in 653 US cities

• Over 34,000 long-distance phone calls

• 85 million video packets, 27.1 GBytes of data

• End-to-end paths with 5,266 Internet router interfaces

• 51% of routers from dialup ISPs and 45% from UUnet

• Sets D1p and D2p contain successful sessions with streams S1 and S2, respectively

Page 10: (Powerpoint slides)

10

Overview (cont’d)

24

14

27

25

4

5

3

8

2425

2229 22

9

138 23

20

12

21

1311713

7

13

14

248

16

125

4

3

45

5

167

3

19

46

66

22

Legend(total 653)

20 to 29 (14)14 to 20 (6)

9 to 14 (10)6 to 9 (9)1 to 6 (12)

• Cities per state that participated in the experiment

Page 11: (Powerpoint slides)

11

Overview (cont’d)Overview (cont’d)• Streaming success rate during the day shown below

• Varied between 80% during the night (midnight – 6 am) to 40% during the day (9 am – noon)

0%

15%

30%

45%

60%

75%

90%

0:00 - 3:00 3:00 - 6:00 6:00 - 9:00 9:00 - 12:00 12:00 - 15:00 15:00 - 18:00 18:00 - 21:00 21:00 - 0:00

time (E DT)

Page 12: (Powerpoint slides)

12

Overview (cont’d)Overview (cont’d)• Average end-to-end hop count 11.3 in D1p and 11.9 in D2p

• The majority of paths contained between 11 and 13 hops

• Shortest path – 6 hops, longest path – 22 hops

0%

5%

10%

15%

20%

25%

30%

6 7 8 9 10 11 12 13 14 15 16 17 18 19 2 21 2

end-to-end hops

D1p D2p

Page 13: (Powerpoint slides)

13

Packet LossPacket Loss• Average packet loss was 0.5% in both datasets

• 38% of 10-min sessions with no packet loss

• 75% with loss below 0.3%

• 91% with loss below 2%

• During the day, average packet loss varied between 0.2% (3 am - 6 am) and 0.8% (9am - 6pm EDT)

• Per-state packet loss varied between 0.2% (Idaho) to 1.4% (Oklahoma), but did not depend on the average RTT or the average number of end-to-end hops in the state (correlation –0.16 and –0.04, respectively)

Page 14: (Powerpoint slides)

14

Packet Loss (cont’d)Packet Loss (cont’d)

0.0%

0.2%

0.4%

0.6%

0.8%

1.0%

1.2%

1.4%

1.6%

state

0

2

4

6

8

10

12

14

16

Average loss Average hops

Page 15: (Powerpoint slides)

15

Packet Loss (cont’d)Packet Loss (cont’d)• 207,384 loss bursts and 431,501 lost packets

• Loss burst lengths (PDF and CDF):

0%

10%

20%

30%

40%

50%

60%

70%

80%

0 5 10 15 20

loss burst length (packets)

0%

10%

20%

30%

40%

50%

60%

70%

80%

90%

100%

PDF CDF

Page 16: (Powerpoint slides)

16

Packet Loss (cont’d)Packet Loss (cont’d)• Most bursts contained no more than 5 packets (however,

the tail reached to over 100 packets)

• RED was disabled in the backbone; still 74% of loss bursts contained only 1 packet apparently dropped in FIFO queues

• Average burst length was 2.04 packets in D1p and 2.10 packets in D2p

• Conditional probability of packet loss was 51% and 53%, respectively

• Over 90% of loss burst durations were under 1 second (maximum 36 seconds)

• The average distance between lost packets was 21 and 27 seconds, respectively

Page 17: (Powerpoint slides)

17

Packet Loss (cont’d)Packet Loss (cont’d)• Apparently heavy-tailed distributions of loss burst lengths

• Pareto with = 1.34; however, note that data was non-stationary (time of day or access point non-stationarity)

1.E-04

1.E-03

1.E-02

1.E-01

1.E+00

1 10 100

loss burst length (packets)

Loss Burst Lengths Power (Loss Burst Lengths)Expon. (Loss Burst Lengths)

Page 18: (Powerpoint slides)

18

Underflow eventsUnderflow events• Missing packets (frames) at their decoding deadlines cause

buffer underflows at the receiver

• Startup delay used in the experiment was 2.7 seconds

• 63% (271,788) of all lost packets were discovered to be missing before their deadlines

• Out of these 63% of lost packets:

– 94% were recovered in time

– 3.3% were recovered late

– 2.1% were never recovered

• Retransmission appears quite effective in dealing with packet loss, even in the presence of large end-to-end delays

Page 19: (Powerpoint slides)

19

Underflow events (cont’d)Underflow events (cont’d)• 37% (159,713) of lost packets were discovered to be

missing after their deadlines had passed

• This effect was caused by large one-way delay jitter

• Additionally, one-way delay jitter caused 1,167,979 data packets to be late for decoding

• Overall, 1,342,415 packets were late (1.7% of all sent packets), out of which 98.9% were late due to large one-way delay jitter rather than due to packet loss combined with large RTT

• All late packets caused the “freeze-frame” effect for 10.5 seconds on average in D1p and 8.5 seconds in D2p (recall

that each session was 10 minutes long)

Page 20: (Powerpoint slides)

20

Underflow events (cont’d)Underflow events (cont’d)• 90% of late retransmissions missed the deadline by no more

than 5 seconds, 99% by no more than 10 seconds

• 90% of late data packets missed the deadline by no more than 13 seconds, 99% by no more than 27 seconds

0%

10%

20%

30%

40%

50%

60%

70%

80%

90%

100%

0.01 0.1 1 10 100

seconds

data retransmissions

Page 21: (Powerpoint slides)

21

Round-trip DelayRound-trip Delay• 660,439 RTT samples

• 75% of samples below 600 ms and 90% below 1 second

• Average RTT was 698 ms in D1p and 839 ms in D2p

• Maximum RTT was over 120 seconds

• Data-link retransmission combined with low-bitrate connection were responsible for pathologically high RTTs

• However, we found access points with 6-7 second IP-level buffering delays

• Generally, RTTs over 10 seconds were considered pathological in our study

Page 22: (Powerpoint slides)

22

Round-trip Delay (cont’d)Round-trip Delay (cont’d)• Distributions of the RTT in both datasets (PDF) were similar

and contained a very long tail

0%

10%

20%

30%

40%

50%

60%

0.1 1 10

RTT (seconds)

D1p D2p

Page 23: (Powerpoint slides)

23

Round-trip Delay (cont’d)Round-trip Delay (cont’d)• Distribution tails closely matched hyperbolic distributions

(Pareto with between 1.16 and 1.58)

1.E-05

1.E-04

1.E-03

1.E-02

1.E-01

1.E+00

0.1 1 10 100

RTT (seconds)

D1p D2p Power (D2p)

Page 24: (Powerpoint slides)

24

Round-trip Delay (cont’d)Round-trip Delay (cont’d)• The average RTT varied during the day between 574 ms (3

am – 6 am) and 847 ms (3 pm – 6 pm) in D1p

• Between 723 ms and 951 ms in D2p

• Relatively small increase in the RTT during the day (by only 30-45%) compared to that in packet loss (by up to 300%)

• Per-state RTT varied between 539 ms (Maine) and 1,053 ms (Alaska); Hawaii and New Mexico also had average RTTs above 1 second

• Little correlation between the RTT and geographical distance of the state from NY

• However, much stronger positive correlation between the number of hops and the average state RTT: = 0.52

Page 25: (Powerpoint slides)

25

Delay JitterDelay Jitter• Delay jitter is one-way delay variation

• Positive values of delay jitter are packet expansion events

• 97.5% of positive samples below 140 ms

• 99.9% under 1 second

• Highest sample 45 seconds

• Even though large delay jitter was not frequent, once a single packet was delayed by the network, the following packets were also delayed creating a “snowball” of late packets

• Large delay jitter was very harmful during the experiment

Page 26: (Powerpoint slides)

26

Packet ReorderingPacket Reordering• Average reordering rates were low, but noticeable

• 6.5% of missing packets (or 0.04% of sent) were reordered

• Out of 16,852 sessions, 1,599 (9.5%) experienced at least one reordering event

• The highest reordering rate per ISP occurred in AT&T WorldNet, where 35% of missing packets (0.2% of sent packets) were reordered

• In the same set, almost half of the sessions (47%) experienced at least one reordering event

• Earthlink had a session where 7.5% of sent packets were reordered

Page 27: (Powerpoint slides)

27

Packet Reordering (cont’d)Packet Reordering (cont’d)• Reordering delay Dr is time between detecting a missing

packet and receiving the reordered packet

• 90% of samples Dr below 150 ms, 97% below 300 ms, 99% below 500 ms, and the maximum sample was 20 seconds

0%

10%

20%

30%

40%

50%

0 200 400 600 800 1000

Reordering delay Dr (ms)

Page 28: (Powerpoint slides)

28

Packet Reordering (cont’d)Packet Reordering (cont’d)• Reordering distance is the number of packets received

during the reordering delay (84.6% of the time a single packet, 6.5% exactly 2 packets, 4.5% exactly 3 packets)

• TCP’s triple-ACK avoids 91.1% of redundant retransmits and quadruple-ACK avoids 95.7%

6.5% 4.5% 2.2% 1.4% 0.5% 0.2% 0.1% 0.0% 0.0%

84.6%

0%

20%

40%

60%

80%

100%

1 2 3 4 5 6 7 8 9 10

reordering distance dr (packets)

Page 29: (Powerpoint slides)

29

Path AsymmetryPath Asymmetry• Asymmetry detected by analyzing the TTL of the returned

packets during the initial traceroute

• Each router reset the TTL to a default value (such as 255) when sending a “TTL expired” ICMP message

• If the number of forward and reverse hops was different, the path was “definitely asymmetric”

• Otherwise, the path was “possibly (or probably) symmetric”

• No fail-proof way of establishing path symmetry using end-to-end measurements (even using two traceroutes in reverse directions)

Page 30: (Powerpoint slides)

30

Path Asymmetry (cont’d)Path Asymmetry (cont’d)• 72% of sessions operated over definitely asymmetric paths

• Almost all paths with 14 or more end-to-end hops were asymmetric

• Even the shortest paths (with as low as 6 hops) were prone to asymmetry

• “Hot potato” routing is more likely to cause asymmetry in longer paths, because they are more likely to cross AS borders than shorter paths

• Longer paths also exhibited a higher reordering probability than shorter paths

Page 31: (Powerpoint slides)

31

ConclusionConclusion• Dialing success rates were quite low during the day (as low

as 40%)

• Retransmission worked very well even for delay-sensitive traffic and high-latency end-to-end paths

• Both RTT and packet-loss bursts appeared to be heavy-tailed

• Our clients experienced huge end-to-end delays both due to large IP buffers as well as persistent data-link retransmission

• Reordering was frequent even given our low bitrates

• Most paths were in fact asymmetric, where longer paths were more likely to be identified as asymmetric