Making Internet faster

December 09, 2010

There is no single entity behind the Internet. It is made up of different networks that are managed by service providers. These service providers - or operators - manage what data is being sent and monitor the amount of traffic being used in terms of simple web browsing, multimedia streaming or peer to peer file sharing. When the data traffic on a network is too dense what experts call "bottlenecks" can occur, slowing the delivery of information to your computer, which can result in a slower Internet experience. A EUREKA-backed project entitled TRAMMS, for Traffic Measurements and Models in Multi-Service networks, incorporating teams from Sweden, Hungary and Spain, aimed to solve this issue by gaining access to Internet networks run by operators in both Sweden and Spain and monitoring traffic over a period of three years. This gave them an excellent insight into user behaviour, enabling them to accurately measure network traffic so that in the future, service providers know how much capacity is needed and can avoid bottlenecks.

Taming the Internet beast

The particularity of this research project is that the team of experts taking part in it was given access to very sensitive data on Internet traffic measurements. Operators normally tend to guard this information jealously as it constitutes their core business. "Internet traffic measurements are very difficult to find if you are not an operator," says Mr. Andreas Aurelius, coordinator of the project and senior scientist at Acreo AB, one of the project partners. Previous research in this field has normally been limited to campus networks, and limited to a geographical area. "That is one of the unique things about this project," he says. "We were using data in access networks, not campus networks as most researchers do."

The type of information the project monitored were designed to get an overall view of traffic passing through the networks. This included IP traffic (the flow of data on the Internet), routing decisions (the selection of which path to send network traffic), quality of service (giving priority to certain applications, such as multimedia) and available bandwidth. This was innovative as the partners developed new tools that measured traffic which gave a complete picture of a network. These tools, already targeted for use by many operators will make web browsing considerably faster.

"For everyday users, this means better quality for multimedia services over the Internet, like streaming for example" says Mr. Aurelius.

Setting new standards to measure Internet traffic

Another question that springs to mind is how the team was able to acquire all of this information without flouting any privacy laws. The answer is that through agreements with the operators, the partners had access to certain information, but not all of it. "The information was post-processed, so it only contained data. It wasn't linked to any customers or IP addresses. We could see what type of application was being used, for example if it was peer to peer, but we couldn't see what file was downloaded," explains Mr. Aurelius. Getting access to such delicate information was a great coup for the project, and as a result the privacy concerns were taken very seriously.

The team managed to collect an astonishing 3 000 terabytes of data over the three years of the project. This was important as it allowed them to study trends and changes over an extended period of time amid a continuous influx of information.

The project was also notable for the fact that a number of the processes that were carried out are under consideration by the International Telecommunication Union to become standardised forms of measurement. An example of these standards is BART (Bandwidth Available in Real Time), which monitors available bandwidth between a sender and a receiver in a network such as the Internet.

Getting big companies' attention

Transnational partnerships and multicultural issues didn't seem to affect the TRAMMS project, as some of the partners had already worked together on previous projects, so they gelled together well.

"There was a tighter bond between the national partners," says Mr. Aurelius, but there were regular international meetings with the partners which helped foster cooperation. In the end, 11 partners successfully completed the project and the three countries involved complemented each others areas of specialisation making the project a perfect example of international cooperation.

Mr. Aurelius has even come back for more, with a follow-up project entitled IPNQSIS already under way. This project deals with quality of experience in network services, such as voice over IP (VoIP), video on demand (VoD), IPTV, and so on. These are sectors where network service providers are expecting huge revenue opportunities and need to improve the quality of the offered services as perceived by the users with a goal to minimize the customer churn yet maintaining their competitive edge. This makes the topic an ideal follow-up project for the team that already worked on TRAMMS.

The TRAMMS project was started under the Celtic Initiative which is a EUREKA cluster specialising in communication technology. Mr. Aurelius praised the way Celtic was involved in the set up phase of the project, finding funding and partners.

The project came to a conclusion at the end of 2009 and the first results are impressive. No less than five companies have taken up the methodology used for traffic measurements: Ericsson, Procera, Telnet-RI, NAUDIT, and GCM Communications Technology.

It is a welcome outcome for Mr. Aurelius who also had to deal with the ramifications of the global economic recession. "The economic recession hit us hard, and many partners dropped out or slowed down their cooperation," he said. But that did obviously not diminish the work done in the project. "Ultimately it was very successful - better than we could have imagined!".


Related Data Articles from Brightsurf:

Keep the data coming
A continuous data supply ensures data-intensive simulations can run at maximum speed.

Astronomers are bulging with data
For the first time, over 250 million stars in our galaxy's bulge have been surveyed in near-ultraviolet, optical, and near-infrared light, opening the door for astronomers to reexamine key questions about the Milky Way's formation and history.

Novel method for measuring spatial dependencies turns less data into more data
Researcher makes 'little data' act big through, the application of mathematical techniques normally used for time-series, to spatial processes.

Ups and downs in COVID-19 data may be caused by data reporting practices
As data accumulates on COVID-19 cases and deaths, researchers have observed patterns of peaks and valleys that repeat on a near-weekly basis.

Data centers use less energy than you think
Using the most detailed model to date of global data center energy use, researchers found that massive efficiency gains by data centers have kept energy use roughly flat over the past decade.

Storing data in music
Researchers at ETH Zurich have developed a technique for embedding data in music and transmitting it to a smartphone.

Life data economics: calling for new models to assess the value of human data
After the collapse of the blockchain bubble a number of research organisations are developing platforms to enable individual ownership of life data and establish the data valuation and pricing models.

Geoscience data group urges all scientific disciplines to make data open and accessible
Institutions, science funders, data repositories, publishers, researchers and scientific societies from all scientific disciplines must work together to ensure all scientific data are easy to find, access and use, according to a new commentary in Nature by members of the Enabling FAIR Data Steering Committee.

Democratizing data science
MIT researchers are hoping to advance the democratization of data science with a new tool for nonstatisticians that automatically generates models for analyzing raw data.

Getting the most out of atmospheric data analysis
An international team including researchers from Kanazawa University used a new approach to analyze an atmospheric data set spanning 18 years for the investigation of new-particle formation.

Read More: Data News and Data Current Events is a participant in the Amazon Services LLC Associates Program, an affiliate advertising program designed to provide a means for sites to earn advertising fees by advertising and linking to