Recent advances in machine learning have profoundly influenced our study of computer vision. Successes in this field have demonstrated the expressive power of learning representations directly from visual imagery — both in terms of practical utility and unexpected expressive abilities. In this talk I will discuss several contributions which have helped improve our ability to learn representations of images. First, I will describe recent advances for constructing models for extracting semantic information from images by leveraging transfer learning and meta-learning techniques. Such learned models outperform human-invented architectures and are readily scalable across a range of computational budgets. Second, I will highlight recent efforts focused on the converse problem of synthesizing images through the rich visual vocabulary of painting styles and visual textures. This work permits a unique exploration of visual space and offers a window on to the structure of the learned representation of visual imagery. My hope is that these works will highlight common threads in machine and human vision and point towards opportunities for future research.
Speaker Bio: Jon Shlens is a senior research scientist at Google since 2010. Prior to joining Google Research, he was a research fellow at the Howard Hughes Medical Institute and a Miller Fellow at UC Berkeley. His research interests include machine perception, statistical signal processing, machine learning and biological neuroscience.