SOTAVerified

ANDHRA Bandersnatch: Training Neural Networks to Predict Parallel Realities

2024-11-28Code Available0· sign in to hype

Venkata Satya Sai Ajay Daliparthi

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Inspired by the Many-Worlds Interpretation (MWI), this work introduces a novel neural network architecture that splits the same input signal into parallel branches at each layer, utilizing a Hyper Rectified Activation, referred to as ANDHRA. The branched layers do not merge and form separate network paths, leading to multiple network heads for output prediction. For a network with a branching factor of 2 at three levels, the total number of heads is 2^3 = 8 . The individual heads are jointly trained by combining their respective loss values. However, the proposed architecture requires additional parameters and memory during training due to the additional branches. During inference, the experimental results on CIFAR-10/100 demonstrate that there exists one individual head that outperforms the baseline accuracy, achieving statistically significant improvement with equal parameters and computational cost.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CIFAR-10ABNet-2G-R3-CombinedPercentage correct96.38Unverified
CIFAR-10ABNet-2G-R3Percentage correct96.09Unverified
CIFAR-10ABNet-2G-R2Percentage correct95.9Unverified
CIFAR-10ABNet-2G-R1Percentage correct95.54Unverified
CIFAR-10ABNet-2G-R0Percentage correct94.12Unverified
CIFAR-100ABNet-2G-R3-CombinedPercentage correct82.78Unverified
CIFAR-100ABNet-2G-R3Percentage correct80.83Unverified
CIFAR-100ABNet-2G-R2Percentage correct80.35Unverified
CIFAR-100ABNet-2G-R1Percentage correct78.79Unverified
CIFAR-100ABNet-2G-R0Percentage correct73.93Unverified

Reproductions