MS (Master of Science)
Date of Award
Committee Chair or Co-Chairs
Jeff R. Knisley
Teresa W. Haynes, Debra J. Knisley
This thesis presents the use of a new sigmoid activation function in backpropagation artificial neural networks (ANNs). ANNs using conventional activation functions may generalize poorly when trained on a set which includes quirky, mislabeled, unbalanced, or otherwise complicated data. This new activation function is an attempt to improve generalization and reduce overtraining on mislabeled or irrelevant data by restricting training when inputs to the hidden neurons are sufficiently small. This activation function includes a flattened, low-training region which grows or shrinks during back-propagation to ensure a desired proportion of inputs inside the low-training region. With a desired low-training proportion of 0, this activation function reduces to a standard sigmoidal curve. A network with the new activation function implemented in the hidden layer is trained on benchmark data sets and compared with the standard activation function in an attempt to improve area under the curve for the receiver operating characteristic in biological and other classification tasks.
Thesis - Open Access
Bonnell, Jeffrey A., "Implementation of a New Sigmoid Function in Backpropagation Neural Networks." (2011). Electronic Theses and Dissertations. Paper 1342. https://dc.etsu.edu/etd/1342
Copyright by the authors.