We take some 80 billion photographs each year which would require
around 400 petabytes to store if they were all saved. Finding your
cherished shot of Aunt Marjory’s 80th birthday party among that lot is
going to take some special kind of search algorithm. And of course,
various groups are working on just how to solve this problem.
But if you want to build the next generation of image search
algorithms, you need a database on which to test it, say Andrea Esuli
and pals at the Institute of Information Science and Technologies in
Pisa, Italy. And they have one: a database of 100 million high quality
digital images taken from Flickr. For each image they have extracted
five descriptive features such as colours, shape, and texture, as
defined by the MPEG-7 image standard.
That’s no mean feat. Esuli and co point out that such an image database would normally require the download and processing of up to 50 TB of data, something that would take take about 12 years on a standard PC and about 2 years using a high-end multi-core PC. Instead, they simply decided to crawl the Flickr site, where the pictures are already stories, taking what data they need as descripitors. This paper describes the trials and tribulations of building such a database.
Elusi and co also announce that the resulting collection is now open to the research community for experiments and comparisons. So if you’re testing the next generation of image search algorithm, this is the database you need to set it loose on.
Don’t settle for half the story.
Get paywall-free access to technology news for the here and now.