Feature Extraction of Video Using Artificial Neural Network

Feature Extraction of Video Using Artificial Neural Network

Yoshihiro Hayakawa (National College of Technology, Sendai, Japan), Takanori Oonuma (National College of Technology, Sendai, Japan), Hideyuki Kobayashi (National College of Technology, Sendai, Japan), Akiko Takahashi (Sendai National College of Technology, Sendai, Japan), Shinji Chiba (National College of Technology, Sendai, Japan) and Nahomi M. Fujiki (National College of Technology, Sendai, Japan)
DOI: 10.4018/IJCINI.2017040102
OnDemand PDF Download:
$30.00
List Price: $37.50

Abstract

In deep neural networks, which have been gaining attention in recent years, the features of input images are expressed in a middle layer. Using the information on this feature layer, high performance can be demonstrated in the image recognition field. In the present study, we achieve image recognition, without using convolutional neural networks or sparse coding, through an image feature extraction function obtained when identity mapping learning is applied to sandglass-style feed-forward neural networks. In sports form analysis, for example, a state trajectory is mapped in a low-dimensional feature space based on a consecutive series of actions. Here, we discuss ideas related to image analysis by applying the above method.
Article Preview

Introduction

According to the 2015 Telecommunications White Paper from the Japanese Ministry of Internal Affairs and Communications, the number of “things” related to the Internet of Things (IoT) is expected to increase tremendously; a wide variety of data will be collected and analyzed via IoT and movements to link these are expected to enhance work efficiency, among other things. Naturally, this includes the treatment of big data, and when analyzing data collected in such large quantities, their analysis could be performed interactively, allowing interpretations from many different angles. In particular, security camera images, while considering privacy, are linked to a large volume of data collected using IoT. The reason for this is exceptionally simple. For example, to monitor the daily behavioral patterns of elderly people and children, it is convenient to construct surveillance systems using data collected from IoT because it is portable and allows us to analyze and extract data in order to evaluate whether their behaviors have changed or are in danger of doing so. There are the researches in the computer network for the realization of such a system (Shintaro, Mariko, Mingrui, Yoshikazu, & Toshimitsu, 2013). However, with this simple idea, the main point is how to construct a system in which the behavior pattern data collected from IoT can be linked to actual behavior. Although the search engine to connect sensor data with an event has been already proposed (Koji, Yutaka, Takuya, Yasue, Yasushi, & Takeshi, 2010; Takeshi, Yasue, Takuya, Koji, Yutaka, & Yasushi, 2010), we believe that if a video is captured at the same time as the IoT data are collected, specification of these data is a simple matter.

In relation to the treatment of images, deep learning using deep neural networks (DNNs) has attracted a lot of attention in recent years because it successfully demonstrated automatic organization of neurons selectively reacting to images of a cat (Quoc,2013) and high recognition performance in relation to difficult tasks, such as general object recognition. In other words, this success engendered several methods to mitigate model complexity including weight decay (Stephen, & Lorien, 1989) and weight sharing (Le, Bernhard, John, Donnie, Richard, Wayne, & Lawrence, 1990; Steven, & Geoffrey, 1992) and to roughly form DNN at an early stage through prior learning (Geoffrey, Simon, & Yee-Whye, 2006). Furthermore, by creating an ensemble of multiple neural networks, such as DropOut (Geoffrey, Nitish, Alex, Ilya, & Ruslan, 2012) or DropConnect (Li, Matthew, Sixin, Yann, & Rob, 2013), which is a model assumed to have information deficiency in it, it has greatly contributed to the development of various parameter tuning for keeping generality.

Analyzing the action parameters from videos captured at the same time as data is collected from IoT, it is clear that although human eyes can specify actions easily by directly viewing them, due to issues such as their processing ability and fatigue, they soon encounter limitations when processing large quantities for long periods. Therefore, there is a demand for the automation of action pattern analysis using such images, and DNN can be used to achieve this.

In DNN image processing, if a convolutional neural network (CNN) (Alex, Ilya, & Geoffrey, 2012) or sparse coding is used, the response neurons corresponding to image input in the feature extraction layer learn to respond and become able to recognize objects. By utilizing this process on a video frame by frame, it is possible to consider a discrete feature vector corresponding to this video (Yasuyuki, Takashi, & Kuniaki, 2015). However, we have considered utilizing a continuous trajectory of an object or person moving consecutively. These trajectories are formed from a nonlinear mapping into a low-dimensional feature space by handling feed-forward neural networks without CNN or sparse coding.

Complete Article List

Search this Journal:
Reset
Open Access Articles: Forthcoming
Volume 11: 4 Issues (2017): 3 Released, 1 Forthcoming
Volume 10: 4 Issues (2016)
Volume 9: 4 Issues (2015)
Volume 8: 4 Issues (2014)
Volume 7: 4 Issues (2013)
Volume 6: 4 Issues (2012)
Volume 5: 4 Issues (2011)
Volume 4: 4 Issues (2010)
Volume 3: 4 Issues (2009)
Volume 2: 4 Issues (2008)
Volume 1: 4 Issues (2007)
View Complete Journal Contents Listing