A Framework Model for Integrating Social Media, the Web, and Proprietary Services Into YouTube Video Classification Process

A Framework Model for Integrating Social Media, the Web, and Proprietary Services Into YouTube Video Classification Process

Mohamad Hammam Alsafrjalani (University of Miami, Miami, USA)
DOI: 10.4018/IJMDEM.2019040102
OnDemand PDF Download:
List Price: $37.50
10% Discount:-$3.75


Online video streaming has gained ubiquity in disparate educational, governmental, and corporate environments. The ubiquity of these videos elicits new challenges in video classification that is used to promote relative videos and block unwanted content. These challenges include the incorporation of contextual information, rapid development of the ad-hoc query modules, and keeping pace with contemporary contextual information. In this article, the authors present a framework model for incorporating contextual information into video classification information. To illustrate the model, the authors propose a framework which comprises video classification, web search engines, social media platforms, and third-party classification modules. The modules enable the framework's flexibility and adaptivity to different contextual environment—educational, governmental, and corporate. Additionally, the model emphasizes standardized module interface to enable the framework's extensibility and rapid development of future modules.
Article Preview

1. Introduction

The importance of video classification has increased in the last decade due to the ubiquity and exponential growth of crowd-generated videos—more than 80% of the internet traffic will be for video streaming (Cisco, 2017). Demands for classifying videos are elicited by various corporate, government, and educational policies that require identification of harmful videos (e.g., phishing and spam), hate- and crime-promoting content, pornography, and/or cyber bullying, which can be spread through online video streaming platforms. Alternatively, the classification information can be used to promote relative content, generate targeted advertisement, and retain more users (Duverger & Steffes, 2012; Xu, Zhang, et al., 2008).

To classify YouTube videos, machine learning algorithms (e.g. Convolutional Neural Networks (CNN) (Karpathy et al., 2014)) are used to extract key features from the videos’ frames (e.g., (Roach, Mason, & Pawlewski, 2001)), text (e.g., (Brezeale & Cook, 2006)), audio (e.g., (Z. Liu, Wang, & Chen, 1998)), or a combination of these data (e.g., (Qi, Gu, Jiang, Chen, & Zhang, 2000)). This information is used to train the models offline using a set of pre-downloaded videos. After the models are trained, they are used to predict the class of new, unknown videos. However, the drawbacks of these approaches include lengthy training time and high computation complexity, and a priori access to training videos.

To reduce time and computation complexity, other research targeted key frames and segments (Lu, Drew, & Au, 2001), the video’s texts (Huang, Fu, & Chen, 2010), such as the title, description, and/or comments. However, these approaches can only predict classification of new videos using training on pre-determined features that were available during training time, and they did not consider new, unprecedented classes (e.g., a new online challenge). Furthermore, the exponential growth of the number of video creators on social media—over 1 million YouTube channels (YouTube, 2017) and 8 billion Facebook videos watched per day (Tang et al., 2017), the number of new, subjective video classes will increase beyond what a single model can determine. Furthermore, the subjectivity of the information technology’s (IT) policies deployed at different entities—government agencies, corporate office, and education institution, exacerbate the relativity of the classification information. Since these policies are focused on the entity’s demands, the policies require adaptivity to the contextual environment in which the video is being streamed.

A generic video classification information (sport, news, entertainment, …) must be accompanied with contextual information. Whereas the generic classification information provides valid information, awareness of the contextual environment will impact the IT decision about these videos. For instance, a video that was classified into commercial advertisement, as a result of pre-trained classification model, could be also classified as an online challenge based on recent events in the online and social media. Since the recent events, and possibly the category itself, were unavailable during the module training period, the trained module would fail to predict the online challenge as a category of the video. If the recent events, contextual information, is obtained as part of a classification framework, the IT policies can provide more accurate decisions.

Complete Article List

Search this Journal:
Volume 14: 1 Issue (2023): Forthcoming, Available for Pre-Order
Volume 13: 4 Issues (2022): 1 Released, 3 Forthcoming
Volume 12: 4 Issues (2021)
Volume 11: 4 Issues (2020)
Volume 10: 4 Issues (2019)
Volume 9: 4 Issues (2018)
Volume 8: 4 Issues (2017)
Volume 7: 4 Issues (2016)
Volume 6: 4 Issues (2015)
Volume 5: 4 Issues (2014)
Volume 4: 4 Issues (2013)
Volume 3: 4 Issues (2012)
Volume 2: 4 Issues (2011)
Volume 1: 4 Issues (2010)
View Complete Journal Contents Listing