Deep visual analogy-making

SE Reed, Y Zhang, Y Zhang… - Advances in neural …, 2015 - proceedings.neurips.cc
Advances in neural information processing systems, 2015proceedings.neurips.cc
In addition to identifying the content within a single image, relating images and generating
related images are critical tasks for image understanding. Recently, deep convolutional
networks have yielded breakthroughs in producing image labels, annotations and captions,
but have only just begun to be used for producing high-quality image outputs. In this paper
we develop a novel deep network trained end-to-end to perform visual analogy making,
which is the task of transforming a query image according to an example pair of related …
Abstract
In addition to identifying the content within a single image, relating images and generating related images are critical tasks for image understanding. Recently, deep convolutional networks have yielded breakthroughs in producing image labels, annotations and captions, but have only just begun to be used for producing high-quality image outputs. In this paper we develop a novel deep network trained end-to-end to perform visual analogy making, which is the task of transforming a query image according to an example pair of related images. Solving this problem requires both accurately recognizing a visual relationship and generating a transformed query image accordingly. Inspired by recent advances in language modeling, we propose to solve visual analogies by learning to map images to a neural embedding in which analogical reasoning is simple, such as by vector subtraction and addition. In experiments, our model effectively models visual analogies on several datasets: 2D shapes, animated video game sprites, and 3D car models.
proceedings.neurips.cc
Showing the best result for this search. See all results