SOTAVerified

Hierarchical spike coding of sound

2012-12-01NeurIPS 2012Unverified0· sign in to hype

Yan Karklin, Chaitanya Ekanadham, Eero P. Simoncelli

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We develop a probabilistic generative model for representing acoustic event structure at multiple scales via a two-stage hierarchy. The first stage consists of a spiking representation which encodes a sound with a sparse set of kernels at different frequencies positioned precisely in time. The coarse time and frequency statistical structure of the first-stage spikes is encoded by a second stage spiking representation, while fine-scale statistical regularities are encoded by recurrent interactions within the first-stage. When fitted to speech data, the model encodes acoustic features such as harmonic stacks, sweeps, and frequency modulations, that can be composed to represent complex acoustic events. The model is also able to synthesize sounds from the higher-level representation and provides significant improvement over wavelet thresholding techniques on a denoising task.

Tasks

Reproductions