A Neural-Symbolic Architecture for Inverse Graphics Improved by Lifelong Meta-Learning
2019-05-22Code Available0· sign in to hype
Michael Kissner, Helmut Mayer
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/Kayzaks/VividNetOfficialIn papernone★ 0
Abstract
We follow the idea of formulating vision as inverse graphics and propose a new type of element for this task, a neural-symbolic capsule. It is capable of de-rendering a scene into semantic information feed-forward, as well as rendering it feed-backward. An initial set of capsules for graphical primitives is obtained from a generative grammar and connected into a full capsule network. Lifelong meta-learning continuously improves this network's detection capabilities by adding capsules for new and more complex objects it detects in a scene using few-shot learning. Preliminary results demonstrate the potential of our novel approach.