Web’s Inventor Tim Berners-Lee Wins the Nobel Prize of Computing
The man who made it possible for you to read this page just received the highest honor in computer science.
In 1989 Tim Berners-Lee, a programmer at the physics laboratory CERN, proposed a system that would allow computers to publish and access linked documents and multimedia over the Internet. Today the world runs on the Web, and Berners-Lee has been given the ACM Turing Award, considered something like the Nobel of computer science. He talked with MIT Technology Review about his invention’s past, present, and future.
Take us back to 1989. What was the problem you were setting out to solve when you started this work that led to the World Wide Web?
I was working at CERN and I was frustrated because people had brought along all kinds of wonderful computers—all different. Each would have some way of keeping track of documents and manuals and help files but they were all different as well. I felt, wouldn't it be wonderful if all of these systems could be somehow part of or considered part of one big meta-system?
The first thing we did was start a Web server at CERN for the phone book. It [previously] ran on a mainframe, which was a pain to have to log in to, and a lot of people just logged on to the mainframe to look up people’s phone numbers. The phone book got a few people to install a Web browser. It spread outside CERN in high-energy physics, and then ended up taking off exponentially.
Tim Berners-Leeu2019s career
Writes a proposal for a “distributed hypertext system.”
The first website goes online.
Founds World Wide Web Consortium for Web standards.
Calls for development of a Semantic Web readable by computers.
Founds World Wide Web Foundation to widen Web access.
The Web now feels indispensable, and is part of many people’s daily personal and professional life. What still needs working on?
Now we have to talk about it as a human right. It’s not as basic as water, but the difference in economic and social power between someone who has it and someone who doesn’t means they’re massively disadvantaged. If you're in a village in Africa and you don't have access because you can't afford it, or you do have access and you can't use it because you're not literate, then this is a problem.
When we started the World Wide Web foundation [in 2009] we started pointing out that if 20 percent of the world has it, they owe the other 80 percent to try and get them connected as quickly as possible. [The UN said last November that 47 percent of the world’s population is now online.]
Looking ahead, you have spoken about the need for the Web to be “re-decentralized.” And you’re part of a community working on technology intended to do that. What’s the thinking behind this movement?
In the late ‘90s there was a massive excitement about what an incredibly empowering system this is, and de-centralization was an important part of that. Without asking anybody else, I could get a computer, put some software on it, plug it into the Internet, and I’d have a blog and a voice. People thought those voices would mount up and provide really exciting things. We have got great things like Wikipedia and crowdfunding, but a lot of people spend all their time in social network silos. A social network is disempowering because you put a lot of energy into it, all your personal data out there, and tell it who your friends are. You can only use that information inside the silo of that particular social network.
We had a few workshops, and we've got people working in the lab on things like Solid project at MIT, which is saying we could go to a world where everybody is in charge of their own data. Imagine if all the applications that you run point at data you control. That is an exciting new mode of operating.
Geoffrey Hinton tells us why he’s now scared of the tech he helped build
“I have suddenly switched my views on whether these things are going to be more intelligent than us.”
Meet the people who use Notion to plan their whole lives
The workplace tool’s appeal extends far beyond organizing work projects. Many users find it’s just as useful for managing their free time.
Learning to code isn’t enough
Historically, learn-to-code efforts have provided opportunities for the few, but new efforts are aiming to be inclusive.
Deep learning pioneer Geoffrey Hinton has quit Google
Hinton will be speaking at EmTech Digital on Wednesday.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.