AdLU: adaptive double parametric activation functions

Warning The system is temporarily closed to updates for reporting purpose.

Güney Duman, Merve and Koparal, Sibel and Ömür, Neşe and Ertürk, Alp and Aptoula, Erchan (2026) AdLU: adaptive double parametric activation functions. Digital Signal Processing, 168 (Part C). ISSN 1051-2004 (Print) 1095-4333 (Online)

Full text not available from this repository. (Request a copy)

Abstract

Activation functions are critical components of neural networks, introducing the necessary nonlinearity for learning complex data relationships. While widely used functions such as ReLU and its variants have demonstrated notable success, they still suffer from limitations such as vanishing gradients, dead neurons, and limited adaptability at various degrees. This paper proposes two novel differentiable double-parameter activation functions (AdLU1 and AdLU2) designed to address these challenges. They incorporate tunable parameters to optimize gradient flow and enhance adaptability. Evaluations on benchmark datasets, MNIST, FMNIST, USPS, and CIFAR-10, using ResNet-18 and ResNet-50 architectures, demonstrate that the proposed functions consistently achieve high classification accuracy. Notably, AdLU1 improves accuracy by up to 5.5 % compared to ReLU, particularly in deeper architectures and more complex datasets. While introducing some computational overhead, their performance gains establish them as competitive alternatives to both traditional and modern activation functions.
Item Type: Article
Uncontrolled Keywords: Activation functions; AdLU; Deep neural networks; ResNet-18; ResNet-50
Divisions: Faculty of Engineering and Natural Sciences
Depositing User: Erchan Aptoula
Date Deposited: 26 Sep 2025 15:44
Last Modified: 26 Sep 2025 15:44
URI: https://research.sabanciuniv.edu/id/eprint/52574

Actions (login required)

View Item
View Item