RMIT University
Browse

Enhancement of neural networks with an alternative activation function tanhLU

journal contribution
posted on 2024-11-02, 19:33 authored by Shui-Long Shen, Ning Zhang, Annan ZhouAnnan Zhou, Zhen-Yu Yin
A novel activation function (referred to as tanhLU) that integrates hyperbolic tangent function (tanh) with a linear unit is proposed as a promising alternative to tanh for neural networks. The tanhLU is inspired by the boundlessness of rectified linear unit (ReLU) and the symmetry of tanh. Three variable parameters in tanhLU controlling activation values and gradients could be preconfigured as constants or adaptively optimized during the training process. The capacity of tanhLU is first investigated by checking the weight gradients in error back propagation. Experiments are conducted to validate the improvement of tanhLUs on five types of neural networks, based on seven benchmark datasets in different domains. tanhLU is then applied to predict the highly nonlinear stress–strain relationship of soils by using the multiscale stress–strain (MSS) dataset. The experiment results indicate that using constant tanhLU leads to apparent improvement on FCNN and LSTM with lower loss and higher accuracy compared with tanh. Adaptive tanhLUs achieved the state-of-the-art performance for multiple deep neural networks in image classification and face recognition.

History

Related Materials

  1. 1.
    DOI - Is published in 10.1016/j.eswa.2022.117181
  2. 2.
    ISSN - Is published in 09574174

Journal

Expert Systems with Applications

Volume

199

Number

117181

Start page

1

End page

13

Total pages

13

Publisher

Elsevier

Place published

United Kingdom

Language

English

Copyright

© 2022 Elsevier Ltd. All rights reserved.

Former Identifier

2006114900

Esploro creation date

2022-06-16

Usage metrics

    Scholarly Works

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC