Apple releases first scholarly research paper, discusses use of artificial intelligence to help improve recognition of images in computer vision systems
Date: Thursday, December 29th, 2016, 05:07
Category: Hardware, News, Software
Apple has entered the world of academia.
Last week, the company published its first scholarly research paper with an article covering methods of improving recognition in computer vision systems, marking a new direction for the traditionally secretive company.
The paper, titled “Learning from Simulated and Unsupervised Images through Adversarial Training,” was submitted for review in mid-November before seeing publication through the Cornell University Library on Dec. 22.
The article arrives less than a month after Apple declared that it would no longer bar employees from publishing research relating to artificial intelligence.
In the article, Apple describes techniques of training computer vision algorithms to recognize objects using synthetic, or computer generated, images.
The work describes training models based solely on real-world images, wherein models leveraging synthetic data are often more efficient given that computer generated images are usually labeled. For example, a synthetic image of an eye or hand is annotated as such, while real-world images depicting similar material are unknown to the algorithm and thus need to be described by a human operator.
The author also describes how relying completely on simulated images might yield unsatisfactory results, as computer generated content is sometimes not realistic enough to provide an accurate learning set. The paper suggests bridging the gap via a system of refining a simulator’s output through “Simulated+Unsupervised learning.”
Such a technique combines unlabeled real image data with annotated synthetic images. The technique is based in large part on Generative Adversarial Networks (GANs), which applies two competing neural networks —generator and discriminator —against each other to better discern generated data from real data. A fairly recent development, the process has seen success in the generation of photorealistic “super-resolution” images.
While not necessarily presaging upcoming consumer technology, it is interesting to note Apple elected to apply its modified GAN to the evaluation of gaze and hand pose estimation. In addition, the company says it hopes to one day move S+U learning beyond static images to video input.
The research paper was written by vision expert Ashish Shrivastava and a team of engineers including Tomas Pfister, Oncel Tuzel, Wenda Wang, Russ Webb and Apple Director of Artificial Intelligence Research Josh Susskind. Susskind also made the announcement of Apple’s newfound interest in scholarly pursuits earlier this month, a move some believe will help future recruitment efforts.
So, neat stuff, albeit it’s unclear as to how this might affect future Apple products or technologies.
Stay tuned for additional details as they become available.