Keynote/Boardwatch Internet Backbone Index A better test!!!

The following should be attributed to Gene Shklar:

Please set for 80 columns.

Thanks for the suggestions. We're considering doing a follow-up and perhaps making
this a regular feature every 2 or 4 months thereafter. We've received a few suggestions
for methodology changes/enhancements, and also several emails so far denouncing our
methodology but not explaining why (which is typical of people in many areas -- politics,
the environment, economics, whatever -- who disagree emotionally but not intellectually
with the conclusions of a study).

The current methodology generally shows how a web site connected to a particular
backbone appears to the general internet population of users. The results are intended
to be a guide (but not the only one) for helping web sites select or evaluate a collocation,
hosting, or access provider.

Fine, I will point out a few specific problems with your methodology:

1. Although a variety of backbones is used, the study does not say
which ones. Also, even though the study does point out the assymetrical
routing of a web transavtion (hot-potato), it doesn't point out that
the traffic being measured is a brief web request (which is dumped to
the web server's backbone ASAP) answered by a long response (10KB in
this case, dumped to the querier's backbone).

2. The test used measures the responsiveness of a company's web servers,
which is not necessarily reflective of the response their customers get.
This test specifically measures traffic going "outbound", but suggests
that this information is useful in determining a carrier for "inbound"
traffic. This could be misleading; a web farm will have a lot more
outbound traffic than inbound, and a dial-up only provider will have more
inbound traffic than out.

3. The results show an average taken over 24 hours, ignoring the important
"time of day" factor. Compuserve, for example, may be dog slow
for three hours per day, and greatly over-provisioned for the rest of the
day. In other words, if over a three-hour period samples took 15 seconds,
but only took 2 seconds for the rest of the day, it would score better
than a provider with a consistent 6-7 second rate.

4. The transfer rate of 10KB x 5 is not the same as the transfer rate
of a 50KB file. If one backbone is significantly "burstier" than
another, this could dramatically affect throughput. For instance, a
10KB file might easily go through a bursty or bouncy backbone in just
a few seconds, while larger files require greater consistency.

5. Some companies have more popular web pages than others. Few major
providers hang servers directly off their backbone (whatever that might
mean in this context), but rather have a lobe or two attaching their farm.
Just because a provider's web farm is saturated or busy or slow, does not
indicate that the rest of their backbone is.

Looking back, some of these errors reflect a misinterpretation of mine
that the results were intended for consumers looking for a connection, as
opposed to a place to host their web server. Most of these points still
apply, however. And people will use the study results this way; if I can
make that mistake, J. Random LUser could, too.

Also, the links from the page to the graphs were diappointing. What
units are measured in the "Best Value" chart? What does the price of a
T1 have to do with web hosting?

All of the previous notwithstanding, I would be interested in a better
version of this study. Even more interesting would be to track how
providers do over the course of several studies--who responds well to
backbone congestion?

Gene Shklar GeneShklar@keynote.com
Keynote Systems, Inc. voice (415) 524-3011

Lee

Lee Howard Internet Systems Engineer
(703)208-5231 UUNET High-speed Install
lhoward@uu.net Do I speak for UUNET? [NOT IN ANY WAY]

lhoward@UU.NET writes:

1. Although a variety of backbones is used, the study does not say
which ones. Also, even though the study does point out the assymetrical
routing of a web transavtion (hot-potato), it doesn't point out that the
traffic being measured is a brief web request (which is dumped to the
web server's backbone ASAP) answered by a long response (10KB in this
case, dumped to the querier's backbone).

Ironically, the only person I can remember actively (and
publically) questioning the wisdom of hot potato routing
for several applications is Curtis Villamizar. While Tony
Bates and I had what seemed to be good reasons for doing
closest-exit routing (among other routing policies) at the
time, and there are strong economic incentives for
maintaining closest-exit routing among peers, there are
certainly drawbacks.

Considering that the Keynote/Boardwatch study purports to
help content-providers decide where to locate their
content, poor performance attributable to closest-exit
routing should have engineers considering the wisdom of
that particular routing policy with respect to that
particular market segment, depending on the focus of their
businesses.

The nice thing about the Keynote test is that it might get
some of the cleverer people around here thinking of the
Internet in a more abstract way than they might be used to.

When I think in a more end-to-end fashion about comment #1
above, it seems that a provider who is rated poorly
deserves the rating, even though the performance problem
may be attributable to a peer's network's congested lines,
broken equipment, or whatever. In effect, by taking an
end-to-end approach, this test also rates networks'
routing policies. If the routing policy drives traffic
through a competitor's suboptimal path, and this degrades
end-to-end performance, then the provider with that
routing policy will receive a poorer rating than otherwise.

One of the things that should be thought about is that
there may be ways to improve end-to-end goodput by
changing one's own network based on what one can infer
from tests similar to the Keynote ones (or indeed similar
to Vern Paxson's tests), as much as what one can infer
from what one's own equipment is doing or seeing.

All of the previous notwithstanding, I would be interested in a better
version of this study. Even more interesting would be to track how
providers do over the course of several studies--who responds well to
backbone congestion?

Personally, I would take this comment as a sign that the
study and its publication is astoundingly successful.

Finally, I would rather see people respond to congestion
in a cleverer fashion, i.e. by deploying or pressing for
the deployment of better WWW transport technology. There
are some people with very clever ideas about caching and
lazy proxies, and some clever ideas about changing HTTP in
dramatic ways, that deserve some attention, as these ideas
are more likely to help people and information meet more
quickly in the long run than any amount of tweaking the
underlying transport fabric ever will.

  Sean.

I rerecommended the study Ben Black first recommended because it was
ingeniously end to end. And real end to end, not end to web.

randy