SOTAVerified

Sparse Interpretable Deep Learning with LIES Networks for Symbolic Regression

2025-06-09Code Available0· sign in to hype

Mansooreh Montazerin, Majd Al Aawar, Antonio Ortega, Ajitesh Srivastava

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Symbolic regression (SR) aims to discover closed-form mathematical expressions that accurately describe data, offering interpretability and analytical insight beyond standard black-box models. Existing SR methods often rely on population-based search or autoregressive modeling, which struggle with scalability and symbolic consistency. We introduce LIES (Logarithm, Identity, Exponential, Sine), a fixed neural network architecture with interpretable primitive activations that are optimized to model symbolic expressions. We develop a framework to extract compact formulae from LIES networks by training with an appropriate oversampling strategy and a tailored loss function to promote sparsity and to prevent gradient instability. After training, it applies additional pruning strategies to further simplify the learned expressions into compact formulae. Our experiments on SR benchmarks show that the LIES framework consistently produces sparse and accurate symbolic formulae outperforming all baselines. We also demonstrate the importance of each design component through ablation studies.

Tasks

Reproductions