An Improved Image Compression Model Enabled by Adaptive Active Contour and Supervised Learning-Based ROI Classification

An Improved Image Compression Model Enabled by Adaptive Active Contour and Supervised Learning-Based ROI Classification

Santosh Kumar B. P., Venkata Ramanaiah K.
Copyright: © 2022 |Pages: 26
DOI: 10.4018/IJAMC.290536
Article PDF Download
Open access articles are freely available for download

Abstract

This paper plans to develop a novel image compression model with four major phases. (i) Segmentation (ii) Feature Extraction (iii) ROI classification (iv) Compression. The image is segmented into two regions by Adaptive ACM. The result of ACM is the production of two regions, this model enables separate ROI classification phase. For performing this, the features corresponding to GLCM are extracted from the segmented parts. Further, they are subjected to classification via NN, in which new training algorithm is adopted. As a main novelty JA and WOA are merged together to form J-WOA with the aim of tuning the ACM (weighting factor and maximum iteration), and training algorithm of NN, where the weights are optimized. This model is referred as J-WOA-NN. This classification model exactly classifies the ROI regions. During the compression process, the ROI regions are handled by JPEG-LS algorithm and the non-ROI region are handled by wavelet-based lossy compression algorithm. Finally, the decompression model is carried out by adopting the same reverse process.
Article Preview
Top

1. Introduction

Over the decades, the demand for multimedia products attaints faster growth in the field of communication. The digital images make the bandwidth insufficient as well as consume huge storage space of the memory devices (Miaou, et al., 2009; Lin & Hao, 2005). Thus, there is a necessity to diminish the data redundancy in the image with the desire of saving the hardware space and the transmission bandwidth. In fact, compressing the images plays a very important role in storage as well as transmission purpose by performing the compression into the least number of bits with no loss in the essential information content encapsulated in the original image. In each image, there is redundant data, which represents the duplication of data (Chen, et al., 2005; Shen & Rangayyan, 1997). The redundancy may be due to the frequent repetition of the pixels across the image. The redundancies in the image are distinguished into, psycho-visual redundancy spatial redundancy, and coding redundancy. The elimination of the correlation among the pixels in the natural image via transform coding or predictive coding is referred to as Inter-pixel Redundancy or spatial redundancy. The psycho-visual redundancy is done to reduce the quantity of data to make the visual information equally sensitive to all human eyes. The coding redundancy takes place with variable-length codes of the statistical model (Sanchez, et al., 2008; Srikanth & Ramakrishnan, 2005).

In general, the main approaches in image compression are categorized as Lossless technique and Lossy technique on the basis of the reconstruction of the original image from the compressed image (Taquet & Labit, 2012). With the lossless compression, the reconstruction of the original image from the compressed image is a bit easier while compared to the lossy compression. In the lossy compression, there is a minute difference between the reconstructed image and original image and with the lossy compression, there is a higher compression rate (Velisavljevic, et al., 2007; Creusere, 1997; Sanchez, et al., 2010; de Queiroz, et al., 2000). Run Length Encoding, Entropy Encoding, Huffman Encoding, Arithmetic Coding come under the lossless image compression technique. The lossy image compression includes Scalar Quantization and Vector Quantization (Kim & Cho, 2014; Lin, et al., 2018). With the lossy image compression, the speed of encoding and decoding of images as well as the Signal to Noise Ratio and compression ratio are high.

The most popular image compression technique is DCT as it is efficient in blocking the artifact effect and making the sub-images visible. The JPEG baseline coding system is the common mode employed in DCT as it fits most of the compression applications. This JPEG technique has the lowest compression ratio, and RLE is sufficient only for the files with high repetitive data. The Fractal Encoding has an excellent mathematical-encoding frame, and in contrast, its encoding scheme is slow (Aliaga & Carlbom, 2005; Lee, et al., 2015; Pang, et al., 2019). The Arithmetic encoding makes uses of the fractional values, and these values have complex computations. With the Vector Quantization, there lacks coefficient quantization and hence this scheme is simple to use (Fu, et al., 2018). But, the codebook generation in Vector Quantization is much slower and has low bit rates. The Huffman Encoding is more effective for text or program files other than images. Thus, with the desire of overcoming the issues in the existing image compression techniques, there is a necessity to have an appropriate image compression technique that will have a better compression and reconstruction rate.

The major contribution of this paper is portrayed below:

Complete Article List

Search this Journal:
Reset
Volume 15: 1 Issue (2024)
Volume 14: 1 Issue (2023)
Volume 13: 4 Issues (2022): 2 Released, 2 Forthcoming
Volume 12: 4 Issues (2021)
Volume 11: 4 Issues (2020)
Volume 10: 4 Issues (2019)
Volume 9: 4 Issues (2018)
Volume 8: 4 Issues (2017)
Volume 7: 4 Issues (2016)
Volume 6: 4 Issues (2015)
Volume 5: 4 Issues (2014)
Volume 4: 4 Issues (2013)
Volume 3: 4 Issues (2012)
Volume 2: 4 Issues (2011)
Volume 1: 4 Issues (2010)
View Complete Journal Contents Listing