Google Teaches AIs To Describe Key Image Features

It's no secret that Google is heavily investing in the development of artificial intelligence (AI), and the Mountain View-based tech giant just made some more progress on that front. At this week's Computer Vision and Pattern Recognition conference in Las Vegas, Nevada, the company's scientists and researchers revealed what they've been up to recently. More specifically, they've demonstrated their accomplishments with computer vision systems which they've taught to describe key features of images. Among other things, their latest AIs are capable of determining the most important person in a video, detect and track specific body extremities, and—best of all—tell us about what they've "seen"!

Speaking of its extremities-detecting system, Google's researchers demonstrated it by presenting a system that detects legs of tigers which they've developed in collaboration with the University of Edinburgh. They've explained that the same technology can also be used for not only tracking individual parts of people's bodies but also machines and basically any other object. The implications here are potentially pretty huge as this system could theoretically analyze footage with the goal of singling out - say - people with weapons in their hands. Yes, this type of technology could also allow for some pretty effective spying, but from a technological standpoint, it's incredibly impressive.

The feature which determines "events and key actors" was developed by Google in collaboration with Stanford and creates the so-called "attention mask" for each frame of the analyzed video footage. After determining what's in focus during most of the frames, it can relatively reliably rate the relevance of each person and object present in the footage. That's much more impressive than playing Atari games, right?

Last but not least, Google's scientists presented an AI system capable of accurately describing images in great detail. While that accomplishment could potentially be useful to people with seeing disabilities, it has even greater implications when you look at it in opposite terms. For example, the fact that an AI can "see" and explain that there are three hats on the round coffee table is cool but the potential of you telling your futuristic robot companion to grab you the white hat from the round coffee table is potentially revolutionary. At the moment, the system Google presented in Las Vegas is capable of the former and hopefully, it'll soon make the latter a realistic possibility.

Copyright ©2019 Android Headlines. All Rights Reserved
This post may contain affiliate links. See our privacy policy for more information.
You May Like These
More Like This:
About the Author

Dominik Bosnjak

Head Editor
Dominik started at AndroidHeadlines in 2016 and is the Head Editor of the site today. He’s approaching his first full decade in the media industry, with his background being primarily in technology, gaming, and entertainment. These days, his focus is more on the political side of the tech game, as well as data privacy issues, with him looking at both of those through the prism of Android. Contact him at [email protected]