Skip to Content

The Dizzying Data Rate Conundrum

The suffocating rate at which data is being produced in many experiments raises the question of how to store it for future generations.

When it is switched on later this year, the Large Hadron Collider (LHC) at CERN will smash particles together at the rate of 40 million collisions per second. It’s a process that will generate several petabytes of data per year, and one that the LHC has been set up specifically to handle. The data will be kicked, prodded, and crunched before being analyzed and eventually released into the community as a scientific paper for publication.

That’s its short-term fate. The question is what to do with the data in the long term. Should it be archived somewhere and kept for eternity, and if so, how (and why)?

The LHC is emblematic of a broader problem in science, say André Holzner and his buddies at CERN. That problem is a rapidly growing body of data from increasingly sophisticated experiments. Holzner and co say that an increasingly pressing problem is to understand how this data is being kept in disparate facilities around the world, so that future repositories can be designed to do the job in future.

With that in mind, and with generous funding from the European Union, they’ve questioned over 1,000 high-energy physicists linked to CERN about these questions and published the results on the arXiv.

There seems to be general agreement that data preservation is hugely important. But strangely, there is less agreement over what sort of data should be stored–for example, whether to preserve the raw data itself or some higher-level analysis of it. Stranger still is the broad range of opinion over why the data should be kept at all. Only 60 percent of respondents think that the data should be kept so that conclusions can be checked in future.

Clearly, the broad concern over the issue is matched only by the widespread befuddlement over what to do about it.

Which spells bad news for CERN and other data producers. CERN is about to switch on one of the greatest data fire hoses the world has ever seen. If there is to be any multilateral agreement over what to do in the long term with the data it and other projects produce, the discussions need to be settled sooner rather than later.

Ref: First results from the PARSE.Insight project: HEP survey on data preservation, re-use and (open) access

Keep Reading

Most Popular

AGI is just chatter for now concept
AGI is just chatter for now concept

The hype around DeepMind’s new AI model misses what’s actually cool about it

Some worry that the chatter about these tools is doing the whole field a disservice.

Workers disinfect the street outside Shijiazhuang Railway Station
Workers disinfect the street outside Shijiazhuang Railway Station

Why China is still obsessed with disinfecting everything

Most public health bodies dealing with covid have long since moved on from the idea of surface transmission. China’s didn’t—and that helps it control the narrative about the disease’s origins and danger.

Europe's AI Act concept
Europe's AI Act concept

A quick guide to the most important AI law you’ve never heard of

The European Union is planning new legislation aimed at curbing the worst harms associated with artificial intelligence.

Stay connected

Illustration by Rose WongIllustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at with a list of newsletters you’d like to receive.