Copy For Citation
Sekmen A., Parlaktuna M., Abdul-Malek A., Erdemir E., Koku A. B.
Discover Artificial Intelligence, vol.1, no.12, pp.1-11, 2021 (Peer-Reviewed Journal)
Abstract
This paper introduces two deep convolutional neural network training
techniques that lead to more robust feature subspace separation in
comparison to traditional training. Assume that dataset has M labels. The first method creates M deep convolutional neural networks called {DCNNi}i=1M" role="presentation">{DCNNi}Mi=1. Each of the networks DCNNi" role="presentation">DCNNi is composed of a convolutional neural network (CNNi" role="presentation">CNNi) and a fully connected neural network (FCNNi" role="presentation">FCNNi). In training, a set of projection matrices {Pi}i=1M" role="presentation">{Pi}Mi=1 are created and adaptively updated as representations for feature subspaces {Si}i=1M" role="presentation">{Si}Mi=1. A rejection value is computed for each training based on its projections on feature subspaces. Each FCNNi" role="presentation">FCNNi acts as a binary classifier with a cost function whose main parameter is rejection values. A threshold value ti" role="presentation">ti is determined for ith" role="presentation">ith network DCNNi" role="presentation">DCNNi. A testing strategy utilizing {ti}i=1M" role="presentation">{ti}Mi=1
is also introduced. The second method creates a single DCNN and it
computes a cost function whose parameters depend on subspace separations
using the geodesic distance on the Grasmannian manifold of subspaces Si" role="presentation">Si and the sum of all remaining subspaces {Sj}j=1,j≠iM" role="presentation">{Sj}Mj=1,j≠i.
The proposed methods are tested using multiple network topologies. It
is shown that while the first method works better for smaller networks,
the second method performs better for complex architectures.