North American Network Operators Group

Date Prev | Date Next | Date Index | Thread Index | Author Index | Historical

Re: cost of doing business

  • From: Thomas Kernen
  • Date: Tue Apr 19 09:50:15 2005



Hi Marshall,

Thanks for the links and papers, actually I didn't want to expand into Zipf's law on NANOG, I felt it would be a little OT for most people.
In parallel to the "multicast patching" I mentioned before I'm also indirectly involved in the design of a P2P distribution model using the customers STB hard drive as part of the content source, ie: if you request a VOD content that other users connected to the same POP (or whatever your metric is) those STBs will be potential sources for providing the content to the end user who just requested it.

I personally don't feel this is the way to go based on the current discussions with content owners who seem to prefer models where they don't have to trust the end user's STB since in the short/medium/long run this can't be considered a secure device (too many students with too much time) and therefore would like everything to be sourced from the SP's network or even better from their infrastructure, in the same way this is done with inter-domain multicast for live TV content from those same content providers.

Thomas

----- Original Message ----- From: "Marshall Eubanks" <[email protected]>
To: "Thomas Kernen" <[email protected]>; "Mike Macdonald" <[email protected]>; <[email protected]>; "Andrew Odlyzko" <[email protected]>
Sent: Tuesday, April 19, 2005 3:28 PM
Subject: Re: cost of doing business


Hello;

May be off topic, especially if you do not care about the statistics of content distribution.

On Tue, 19 Apr 2005 07:59:00 +0200
"Thomas Kernen" <[email protected]> wrote:
RE: cost of doing business
That will not be the biggest issue since most of it is live TV hence multicast so that stream
will be used by another customer anyway connected to the same customer termination device in the
POP. Of course that will also be depended on the port density of the customer termination device.

To a point (see below)

The real issue is VOD since each stream is a unicast stream per customer and that currently there
is no valid implementation that allows one to actually reuse some of that unicast stream content
to deliver to other users. Since most VOD is currently on a pay per view model customers should
not be actually requesting those streams if not watching them, so for some time that will allow a
certain "limitation" in the usage.

Related to VOD and bandwidth usage there is some research carried out around the world to solve
this issue and one of these solutions might be the "multicast patching", from what I've seen
there is still a lot of work to be done before it can comply with an SP's network setup. The
other unknown factor is that until now the customer has never had the choice of viewing what he
wants when he wants so AFAIK there is no statistical model that exists in the video media field
that allows to derive a mathematical model related to VOD consumption.

There has been a lot of work on this - as far back as 1994 Ann Chervenak
analyzed results from a video on demand system for her PhD thesis

http://www.isi.edu/~annc/papers/ChervenakThesisNov94.ps.Z

Among the recent work, Johan Pouwese has a interesting paper
analyzing a long stretch of BitTorrent traffic

http://www.isa.its.tudelft.nl/~pouwelse/Bittorrent_Measurements_6pages.pdf

Basically, all of this research finds Pareto (Zipf's law) or modified
Pareto distributions in rank for streaming, Video on Demand, Video rentals, books, blogs

http://homepage.mac.com/kevinmarks/powerlaws.html

even video conferencing. (I found a Mandelbrot-Pareto distribution to be
an excellent match to the usage of VideoConferencing units, for example :
http://www.americafree.tv/rankings/vc_usage.total_hours.post_fit.rank_shift_8.png )

In many recent papers the Pareto distribution is taken for granted,
and may not even be mentioned, but it's there (Jan Pouwelse told me, for example, that
his large BitTorrent logs fit a Pareto distribution beautifully, but it's not even mentioned
in the paper IIRC).

What does this mean ? Well, in a study I did for a satellite startup (that never actually, um,
started), I looked at
video rentals from a large chain; with 13,000 titles, 10 titles get almost 9% of the rentals,
and the top 100 get ~ 22%, but there is a "long tail," and 50% of the traffic comes
from the least popular 11,700 titles.

It seems to me that this has a profound implication concerning both multicast bandwidth
savings and the ability to cache (dynamically or statically) video on demand - while
a lot of bandwidth can be saved by caching or multicast, there will be (if you are truly
offering a large number of titles / shows / channels) a lot of aggregate demand for
content that is not watched by many people, so you
will still need a lot of bandwidth / servers to
service the "long tail." In the above case study, for
example, content was at 5 Mbits/sec, and typically 2 hours long (i.e., movies).
While caching, say, 200 titles in a Terabyte of local storage
might service 28% of your "hits", and reduce your bandwidth load accordingly,
you still need the ability to provide 102,000 downloads of the other
12,800 titles daily if you have 1 million customers (assuming, BTW, that each one only
watches one movie per week), which requires 42 Gbps. As you expand the amount of content
and the size of the community, these numbers only get worse.

I expect to see BitTorrent type systems expand to include distributed storage. The
same 13,000 titles in the above study only require 60 Megabyte per user. Suppose
that this was a BitTorrent community instead, with distributed storage.
If every title was striped, duplicated 5 times, and stored across the community,
for only 300 Megabytes of storage every member of this community could have instant
access to every title. As these communities expand to include basically everyone
with suitable connections, storing and distributing
every piece of video content ever made for public release becomes technically
feasible.

To get back to Randy Bush's comment on Japanese bandwidth to the home, if everyone
has 100 Mbps symmetric, I expect ideas such as this to use it up.


Thomas

Regards
Marshall Eubanks

----- Original Message ----- From: Mike Macdonald
To: Thomas Kernen ; [email protected] ; Andrew Odlyzko
Sent: Tuesday, April 19, 2005 4:17 AM
Subject: RE: cost of doing business


The problem I see coming is simple bandwidth wastage driven by NA TV habits. Many homes have
TVs on during the day full time even when not watched. Now were talking about as much as 10-20
Mbps (depending in HDTV adoption) going into a void.



Mike



-----Original Message----- From: [email protected] [mailto:[email protected]] On Behalf Of Thomas Kernen
Sent: Tuesday, April 19, 2005 1:48 AM
To: [email protected]; Andrew Odlyzko
Subject: Re: cost of doing business



>> fwiw, 100mb to the home costs about that in japan
>
>
>
> We are talking of two different things here, traffic versus access
> bandwidth.
> It will be a while before the average household generates 5 megabit/s
> traffic.
> Even in Korea and Hong Kong, where the average broadband link is in
> the 5-10 Mbps range, average traffic is about 0.1 Mbps. The main
> purpose of high speed links is to get low transaction latency (as in
> "I want that Web page on my screen NOW," or "I want that song for
> transfer to my portable device NOW"), so utilizations are low.
>

For those of us that are already running triple play architectures and working on the data
analysis related to the bandwidth usage growth (in my case over the last 18 months and adding
services one after the other) I see this with a different light:

I fully agree with the transaction latency syndrome, people are compulsive customers that want
to buy right now and you (as a service provider) want to see to them purchase the service before
they change their mind, just need to look at the ringtones market to see how much people are
willing to spend within seconds for a piece of music they will replace in a few days/weeks with
their next favorite tune from the charts that marketing is feeding them with.

Where I don't agree is on the bandwidth usage analysis, once you add the IP based TV/VOD*
services you will be carrying close to 5Mbps on average on your network in the near future.
Either for the one of the TV channels (currently the market is talking about 2 concurrent TV
channels down the same pipe to an end user's home in the North American model or 1 for the

European) or the VOD. So agreed this is not Internet traffic but you will need to carry it
beyond your access termination device (DSLAM/CMTS/ Ethernet

switch) since the economics of the IPTV/VOD market and (current?) technical scalability will
prevent you from being able to have a the full IPTV/VOD streaming (= unicast and/or multicast in
this case) in each POP to keep the traffic as local as possible. So anyhow within your metro area
network accessing and aggregating the customers the amount of bandwith required to service all
customers will grow quite a bit with IPTV/VOD services.

IMHO (of course)
Thomas

*Triple play IPTV/VOD = IP packets carrying a video signal using (name your favorite format)
either as unicast or multicast stream. This excludes the current hybrid HFC networks that still
provide digital TV via an HF stream using (insert your favorite standard here) and the Internet
access and voice service over IP. Anyhow they will migrate once DOCSIS 3.0 and the wideband
benefits have been marketed to all the cable operators as the "next big thing" they need to have
and hence run an IP only service for all the triple play services.