An efficient and effective convolutional neural network for visual pattern recognition

Convolutional neural networks (CNNs) are a variant of deep neural networks (DNNs) optimized for visual pattern recognition, which are typically trained using first order learning algorithms, particularly stochastic gradient descent (SGD). Training deeper CNNs (deep learning) using large data sets (b...

Full description

Bibliographic Details
Main Author: Liew, Shan Sung
Format: Thesis
Language:English
Published: 2016
Subjects:
Online Access:http://eprints.utm.my/60714/1/LiewShanSungPFKE2016.pdf
_version_ 1796860960820953088
author Liew, Shan Sung
author_facet Liew, Shan Sung
author_sort Liew, Shan Sung
collection ePrints
description Convolutional neural networks (CNNs) are a variant of deep neural networks (DNNs) optimized for visual pattern recognition, which are typically trained using first order learning algorithms, particularly stochastic gradient descent (SGD). Training deeper CNNs (deep learning) using large data sets (big data) has led to the concept of distributed machine learning (ML), contributing to state-of-the-art performances in solving computer vision problems. However, there are still several outstanding issues to be resolved with currently defined models and learning algorithms. Propagations through a convolutional layer require flipping of kernel weights, thus increasing the computation time of a CNN. Sigmoidal activation functions suffer from gradient diffusion problem that degrades training efficiency, while others cause numerical instability due to unbounded outputs. Common learning algorithms converge slowly and are prone to hyperparameter overfitting problem. To date, most distributed learning algorithms are still based on first order methods that are susceptible to various learning issues. This thesis presents an efficient CNN model, proposes an effective learning algorithm to train CNNs, and map it into parallel and distributed computing platforms for improved training speedup. The proposed CNN consists of convolutional layers with correlation filtering, and uses novel bounded activation functions for faster performance (up to 1.36x), improved learning performance (up to 74.99% better), and better training stability (up to 100% improvement). The bounded stochastic diagonal Levenberg-Marquardt (B-SDLM) learning algorithm is proposed to encourage fast convergence (up to 5.30% faster and 35.83% better than first order methods) while having only a single hyperparameter. B-SDLM also supports mini-batch learning mode for high parallelism. Based on known previous works, this is among the first successful attempts of mapping a stochastic second order learning algorithm to be deployed in distributed ML platforms. Running the distributed B-SDLM on a 16- core cluster achieves up to 12.08x and 8.72x faster to reach a certain convergence state and accuracy on the Mixed National Institute of Standards and Technology (MNIST) data set. All three complex case studies tested with the proposed algorithms give comparable or better classification accuracies compared to those provided in previous works, but with better efficiency. As an example, the proposed solutions achieved 99.14% classification accuracy for the MNIST case study, and 100% for face recognition using AR Purdue data set, which proves the feasibility of proposed algorithms in visual pattern recognition tasks.
first_indexed 2024-03-05T19:49:08Z
format Thesis
id utm.eprints-60714
institution Universiti Teknologi Malaysia - ePrints
language English
last_indexed 2024-03-05T19:49:08Z
publishDate 2016
record_format dspace
spelling utm.eprints-607142021-01-03T01:31:19Z http://eprints.utm.my/60714/ An efficient and effective convolutional neural network for visual pattern recognition Liew, Shan Sung TK Electrical engineering. Electronics Nuclear engineering Convolutional neural networks (CNNs) are a variant of deep neural networks (DNNs) optimized for visual pattern recognition, which are typically trained using first order learning algorithms, particularly stochastic gradient descent (SGD). Training deeper CNNs (deep learning) using large data sets (big data) has led to the concept of distributed machine learning (ML), contributing to state-of-the-art performances in solving computer vision problems. However, there are still several outstanding issues to be resolved with currently defined models and learning algorithms. Propagations through a convolutional layer require flipping of kernel weights, thus increasing the computation time of a CNN. Sigmoidal activation functions suffer from gradient diffusion problem that degrades training efficiency, while others cause numerical instability due to unbounded outputs. Common learning algorithms converge slowly and are prone to hyperparameter overfitting problem. To date, most distributed learning algorithms are still based on first order methods that are susceptible to various learning issues. This thesis presents an efficient CNN model, proposes an effective learning algorithm to train CNNs, and map it into parallel and distributed computing platforms for improved training speedup. The proposed CNN consists of convolutional layers with correlation filtering, and uses novel bounded activation functions for faster performance (up to 1.36x), improved learning performance (up to 74.99% better), and better training stability (up to 100% improvement). The bounded stochastic diagonal Levenberg-Marquardt (B-SDLM) learning algorithm is proposed to encourage fast convergence (up to 5.30% faster and 35.83% better than first order methods) while having only a single hyperparameter. B-SDLM also supports mini-batch learning mode for high parallelism. Based on known previous works, this is among the first successful attempts of mapping a stochastic second order learning algorithm to be deployed in distributed ML platforms. Running the distributed B-SDLM on a 16- core cluster achieves up to 12.08x and 8.72x faster to reach a certain convergence state and accuracy on the Mixed National Institute of Standards and Technology (MNIST) data set. All three complex case studies tested with the proposed algorithms give comparable or better classification accuracies compared to those provided in previous works, but with better efficiency. As an example, the proposed solutions achieved 99.14% classification accuracy for the MNIST case study, and 100% for face recognition using AR Purdue data set, which proves the feasibility of proposed algorithms in visual pattern recognition tasks. 2016-06 Thesis NonPeerReviewed application/pdf en http://eprints.utm.my/60714/1/LiewShanSungPFKE2016.pdf Liew, Shan Sung (2016) An efficient and effective convolutional neural network for visual pattern recognition. PhD thesis, Universiti Teknologi Malaysia, Faculty of Electrical Engineering. http://dms.library.utm.my:8080/vital/access/manager/Repository/vital:94077
spellingShingle TK Electrical engineering. Electronics Nuclear engineering
Liew, Shan Sung
An efficient and effective convolutional neural network for visual pattern recognition
title An efficient and effective convolutional neural network for visual pattern recognition
title_full An efficient and effective convolutional neural network for visual pattern recognition
title_fullStr An efficient and effective convolutional neural network for visual pattern recognition
title_full_unstemmed An efficient and effective convolutional neural network for visual pattern recognition
title_short An efficient and effective convolutional neural network for visual pattern recognition
title_sort efficient and effective convolutional neural network for visual pattern recognition
topic TK Electrical engineering. Electronics Nuclear engineering
url http://eprints.utm.my/60714/1/LiewShanSungPFKE2016.pdf
work_keys_str_mv AT liewshansung anefficientandeffectiveconvolutionalneuralnetworkforvisualpatternrecognition
AT liewshansung efficientandeffectiveconvolutionalneuralnetworkforvisualpatternrecognition