Optimization of ReLU Activation Function for Deep-Learning-based Breast Cancer Classification on Mammogram Images

The use of the deep Convolutional Neural Network (CNN) in breast cancer classification of mammogram images has been widely investigated to aid radiologists in better clinical diagnoses. Multiple levels of convolution and non-linearity repetitions in CNN's architecture are required to extract si...

Full description

Bibliographic Details
Published in:2024 IEEE International Conference on Automatic Control and Intelligent Systems, I2CACIS 2024 - Proceedings
Main Author: Razali N.F.; Isa I.S.; Sulaiman S.N.; Osman M.K.; Karim N.K.A.; Nordin S.A.
Format: Conference paper
Language:English
Published: Institute of Electrical and Electronics Engineers Inc. 2024
Online Access:https://www.scopus.com/inward/record.uri?eid=2-s2.0-85203830703&doi=10.1109%2fI2CACIS61270.2024.10649623&partnerID=40&md5=253a66e726ed21956914929da0764700
Description
Summary:The use of the deep Convolutional Neural Network (CNN) in breast cancer classification of mammogram images has been widely investigated to aid radiologists in better clinical diagnoses. Multiple levels of convolution and non-linearity repetitions in CNN's architecture are required to extract significant data to be represented. However, the vanishing gradient effect occurs when deeper network training as a product of the partial derivative of loss function on each weightage update can cause no meaningful network learning, even with additional epochs. Overcoming this using the activation function of rectified linear unit (ReLU) by allowing neurons to be activated to allow non-linearity when the output is more than zero could lessen the problem. However, restrictive allowance of non-linearity for <0 for final feature extraction when producing output probability on highly complex data such as mammogram images leads to dropped network performance. To overcome this, this study proposed an adaptive ReLU based on genetic algorithm (GA) profiling to determine the best threshold value for allowing neuron activation based on mutation and adaptation to improve the restrictive capability of the original ReLU. We modified the adaptive ReLU on the final learning layer of two CNN architectures and observed the performance on a public mammogram dataset of INbreast. Our experiments show improved accuracy from 95.0% to 98.5% and improved classification performance compared to other well-known activation functions. Applying evolutionary-based GAs to activation functions can represent an exciting frontier in meta-learning for neural networks. © 2024 IEEE.
ISSN:
DOI:10.1109/I2CACIS61270.2024.10649623