SOTAVerified

Training Deep Neural Networks via Direct Loss Minimization

2015-11-19Code Available0· sign in to hype

Yang Song, Alexander G. Schwing, Richard S. Zemel, Raquel Urtasun

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Supervised training of deep neural nets typically relies on minimizing cross-entropy. However, in many domains, we are interested in performing well on metrics specific to the application. In this paper we propose a direct loss minimization approach to train deep neural networks, which provably minimizes the application-specific loss function. This is often non-trivial, since these functions are neither smooth nor decomposable and thus are not amenable to optimization with standard gradient-based methods. We demonstrate the effectiveness of our approach in the context of maximizing average precision for ranking problems. Towards this goal, we develop a novel dynamic programming algorithm that can efficiently compute the weight updates. Our approach proves superior to a variety of baselines in the context of action classification and object detection, especially in the presence of label noise.

Tasks

Reproductions