SigmoReLU: An Improvement Activation Function by Combining Sigmoid and ReLU<strong> </strong>

Two of the most common activation functions (AF) in deep neural networks (DNN) training are Sigmoid and ReLU. Sigmoid was tend to be more popular the previous decades, but it was suffering with the common vanishing gradient problems. ReLU has resolved these problems by using zero gradient and not tiny values for negative weights and the value “1” for all positives. Although it significant resolves the vanishing of the gradients, it poses new issues with dying neurons of the zero values. Recent approaches for improvements are in a similar direction by just proposing variations of the AF, such as Leaky ReLU (LReLU), while maintaining the solution within the same unresolved gradient problems. In this paper, the combining of the Sigmoid and ReLU in one single function is proposed, as a way to take the advantages of the two. The experimental results demonstrate that by using the ReLU’s gradient solution on positive weights, and Sigmoid’s gradient solution on negatives, has a significant improvement on performance of training Neural Networks on image classification of diseases such as COVID-19, text and tabular data classification tasks on five different datasets..

Medienart:

Preprint

Erscheinungsjahr:

2021

Erschienen:

2021

Enthalten in:

Preprints.org - (2021) vom: 10. Juni Zur Gesamtaufnahme - year:2021

Sprache:

Englisch

Beteiligte Personen:

Mastromichalakis, Stamatis [VerfasserIn]

Links:

Volltext [kostenfrei]

Themen:

620
Engineering

doi:

10.20944/preprints202106.0252.v1

funding:

Förderinstitution / Projekttitel:

PPN (Katalog-ID):

preprintsorg031970117