SOTAVerified

RADAM: Texture Recognition through Randomized Aggregated Encoding of Deep Activation Maps

2023-03-08Code Available1· sign in to hype

Leonardo Scabini, Kallil M. Zielinski, Lucas C. Ribas, Wesley N. Gonçalves, Bernard De Baets, Odemir M. Bruno

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Texture analysis is a classical yet challenging task in computer vision for which deep neural networks are actively being applied. Most approaches are based on building feature aggregation modules around a pre-trained backbone and then fine-tuning the new architecture on specific texture recognition tasks. Here we propose a new method named Random encoding of Aggregated Deep Activation Maps (RADAM) which extracts rich texture representations without ever changing the backbone. The technique consists of encoding the output at different depths of a pre-trained deep convolutional network using a Randomized Autoencoder (RAE). The RAE is trained locally to each image using a closed-form solution, and its decoder weights are used to compose a 1-dimensional texture representation that is fed into a linear SVM. This means that no fine-tuning or backpropagation is needed. We explore RADAM on several texture benchmarks and achieve state-of-the-art results with different computational budgets. Our results suggest that pre-trained backbones may not require additional fine-tuning for texture recognition if their learned representations are better encoded.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
DTDRADAM (ConvNeXt-L)Accuracy84Unverified
FMD (materials)RADAM (ConvNeXt-L)Accuracy (%)95.2Unverified
KTH-TIPS2RADAM (ConvNeXt-XL)Accuracy (%)94.4Unverified

Reproductions