Every internet user has had the frustrating experience of staring at streaming video while it buffers or waiting for a webpage to load. These problems are the result of internet congestion where packets of data cannot be rooted quickly enough to meet demand.
Just as frustrating is the inability to work out what is causing this congestion. Internet users are generally left scratching their heads when it comes to working out which part of the network is responsible for the snarl up.
An interesting question for most users is where the congestion actually occurs. There is a common belief that most congestion occurs in the so-called “last mile”, the final part of the network that links the home or place of work to the network. But is that true?
Other possibilities are that the congestion occurs in the so-called “middle mile”, the network run by the local Internet service provider (ISP) or further afield such as where the ISP connects to the public Internet or even somewhere else entirely.
And there are other questions too. For example, does the pattern of congestion change over time?
These puzzles are not easy to answer largely because the private companies that own various parts of the network keep the details about their traffic, capacity and network topology secret.
And yet, this strategy of secrecy does not serve internet users well and is almost certainly not in the best interests of the ISPs either. “Even the ISPs would beneﬁt from [a better understanding of congestion] as it would allow them to target infrastructure improvements at the key points in the network where return on investment, in terms of enhanced user experience, would be greatest,” say Daniel Genin and Jolene Splett at the National Institute of Standards and Technology in Gaithersburg, Maryland.
That’s partly why in 2010 the Federal Communications Commission began a long term project to gather data about internet congestion and to find out where it really occurs. This study, carried out by a private company called SamKnows, has involved over 10,000 measurement units deployed at the homes and work places of customers of 16 ISPs in the US.
These units–Netgear routers with special firmware–send data to certain servers which measure the time the data takes to get there; the units also download pages from major websites to see how quickly they load. By analysing the results of these tests over long periods of time and from many units, it’s possible to build up a picture of where the congestion occurs and how it changes.
Although the project is ongoing, the data gathered so far is publicly available and today, Genin and Splett reveal its first results. They say that congestion on ISPs offering DSL broadband follows an entirely different pattern to the congestion associated with ISPs offering cable broadband.
It’s worth pointing out that cable broadband is generally much faster than DSL. But do you get the advertised speed?
Not according to Genin and Splett. They say that DSL broadband connections general give download speeds that are at least above 80% of the advertised speed more than 80% of the time. “By contrast, a signiﬁcant fraction of cable connections received less than 80% of their assigned speed tier more than 20% of the time,” they say.
What’s more, the problems don’t appear to be in the last mile and instead occur deeper into the network. “This is somewhat contrary to the popular belief that the edge is more congested than the core,” say Genin and Splett.
There’s still significantly more work to be done to determine exactly what causes this kind of congestion and where exactly it happens and we should see the results of this analysis in the coming months and year.
In the meantime, the results from Genin and Splett should give ISPs something to think about. At the very least, they might want to consider making their traffic data public so that everyone, including the ISP’s themselves, can benefit from the greater insight this should produce.
Ref: arxiv.org/abs/1307.3696 : Where In The Internet Is Congestion?
How SpaceX’s massive Starship rocket might unlock the solar system—and beyond
With the first orbital test launch of Starship on the horizon, scientists are dreaming about what it might make possible— from trips to Neptune to planetary defense.
DeepMind says its new language model can beat others 25 times its size
RETRO uses an external memory to look up passages of text on the fly, avoiding some of the costs of training a vast neural network
The therapists using AI to make therapy better
Researchers are learning more about how therapy works by examining the language therapists use with clients. It could lead to more people getting better, and staying better.
What it will take to unleash the potential of geothermal power
Four new pilot plants funded by the US infrastructure bill could help expand the range of the “forgotten renewable.”
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.