Looking at the photo above, you see a person on a tennis court, wielding a tennis racket and chasing a...lemon. Right?
Wrong. You don’t think it’s a lemon. You know it's a tennis ball.
A computer might not be so perceptive. A computer with the latest image labeling algorithms would have no problem making the following list of objects for the photo above: person, tennis racket, tennis court, lemon.
The only lemon I can imagine on this tennis court is in the water bottle of the line judge.
Computer scientists at
UC San Diego and UCLA are looking to give automated image labeling systems a little more common sense. And that common sense comes in the form of context. And they are squeezing some of that common sense out of a little-known widget from
Google Labs called
Google Sets.
“We think our paper is the first to bring external semantic context to the problem of object recognition,” said computer science professor
Serge Belongie from
UC San Diego's Jacobs
School of Engineering.
Belongie and his students (including Carolina Galleguillos -- the lead singer for the band
Audition Lab) are presenting their "lemon blaster" this week at
ICCV 2007 – the 11
th IEEE International Conference on Computer Vision in Rio
de Janeiro. The computer scientists show that the Google Sets can be used to provide external contextual information to automated object identifiers. The context is added in a post-processing step that comes after the image is split up into parts and labeled by a computer.
A full press release will be available
here, on the Jacobs School Web site.
A copy of the paper is
available hereCheck out the write up on the
Wired Science blog