Online sequential learning with non-iterative strategy for feature extraction, classification and data augmentation

Loading...
Thumbnail Image

Date

Authors

Paul, Adhri Nandini

Journal Title

Journal ISSN

Volume Title

Publisher

Abstract

Network aims to optimize for minimizing the cost function and provide better performance. This experimental optimization procedure is widely recognized as gradient descent, which is a form of iterative learning that starts from a random point on a function and travels down its slope, in steps, until it reaches to the steepest point which is time-consuming and slow to converge. Over the last couple of decades, several variations of the non-iterative neural network training algorithms have been proposed, such as Random Forest and Quicknet. However, the non-iterative neural network training algorithms do not support online training that given a very largesized training data, one needs enormous computing resources to train neural network. In this thesis, a non-iterative learning strategy with online sequential has been exploited. In Chapter 3, a single layer Online Sequential Sub-Network node (OS-SN) classifier has been proposed that can provide competitive accuracy by pulling the residual network error and feeding it back into hidden layers. In Chapter 4, a multilayer network is proposed where the first portion built by transforming multi-layer autoencoder into an Online Sequential Auto-Encoder(OS-AE) and use OS-SN for classification. In Chapter 5, OS-AE is utilized as a generative model that can construct new data based on subspace features and perform better than conventional data augmentation techniques on real-world image and tabular datasets.

Description

Keywords

Neural networks, Network training algorithm, Machine learning, Online sequential learning, Autoencoder

Citation

Endorsement

Review

Supplemented By

Referenced By