Clustering-Based Approach for Clustering Journals in Computer Science

Clustering-Based Approach for Clustering Journals in Computer Science

J. K. D. B. G. Jayaneththi, Banage T. G. S. Kumara
DOI: 10.4018/IJSSOE.2019040103
OnDemand:
(Individual Articles)
Available
$37.50
No Current Special Offers
TOTAL SAVINGS: $37.50

Abstract

In the present scientific world, most of the authors of scientific literature are seeking effective ways to share their research findings with large peer groups. But finding a high-quality journal to publish paper is a huge challenge. Most of the journals present today are predatory and less-quality. The main aim of this study is to help the researchers in identifying the quality level of computer science journals by introducing a data mining model based on six journal quality metrics (Journal Impact Factor, SCImago Journal Rank, Eigenfactor, H-index, Source Normalized Impact per Paper, and Article Influence). Further, another objective is to identify the best metrics to measure the quality of journals out of the six attributes. A sample dataset of 200 journals was used and journals were clustered into five clusters using K-means clustering algorithm. When finding the best quality metrics, Pearson's and Spearman's correlation coefficients were calculated. A more accurate clustering model with an accuracy of 0.9171 was developed considering only suitable attributes.
Article Preview
Top

Introduction

Computer Science subject domain is developing rapidly, and many researchers find different types of new inventions. These researchers search for better and effective ways to share those new inventions with the research community. When considering sharing new research findings with large peer groups, most of the scientific authors focus on publishing their findings in internationally recognized journals. At present, mainly with the development of the Internet and the increase of open-access publishers, a large number of journals have come into operation. But, most of the journals available today are low-quality and predatory ones. Most of them only focus on earning money by publishing papers without maintaining quality. Predatory publishers claim that they provide good editorial services and peer reviews, but in truth, they will publish almost everything that they receive (and paid for) without proper quality control process. Because of the unavailability of a proper quality control process, poorly conducted and false research studies could be published in many journals. That causes greater consequences because these papers are cited and used as the basis for further studies. Another cause for concern is that honest researchers are sometimes duped into believing that journals are legitimate and may end up publishing their valuable efforts in them. Most of these journals have false editorial boards. People with a good reputation are named as editors without their knowledge.

Therefore, measuring the quality of the available journals and the development of suitable methods and metrics to measure the quality of journals has become a crucial need in the present academic society. Many researchers have paved their attention on finding ways to identify the legitimate journals out of the predatory ones. As a conspicuous solution, Jeffrey Beall (a library scientist at the University of Colorado, Denver, USA) has compiled a list of predatory journals and publishers, which can be used by authors to identify predatory journals out of the legitimate ones in different subject fields. But, only identifying predatory journals was not enough. Researchers examined what quality metrics can be used to measure the quality of journals. Many indicators, such as Journal Impact Factor (JIF), SCImago Journal Rank (SJR), Eigenfactor, H-index, Source Normalized Impact per Paper (SNIP), and Article Influence (AI) were developed in order to measure the quality of journals.

As well as by considering these quality metrics many reputed and prestigious databases such as Scopus, Scimago Journal and Country Rank website, eigenfactor.org website, and Web of Science (WOS) were developed to rank and categorize journals. But most of those databases have used a limited number of quality attributes for ranking journals. As examples, Scimago Journal and Country Rank website have used SJR for ranking, WOS has used JIF for ranking, and Scopus has used SNIP for ranking and categorizing journals. Mainly this study aims at developing a new categorization model for Computer Science journals according to their quality by considering the values of their quality attributes such as JIF, SJR, Eigenfactor, H-index, SNIP, and AI. Further, this focuses on finding the most suitable metrics for measuring the quality of Computer Science journals out of the selected six attributes and developing a more accurate data mining model by considering those quality metrics.

The remainder of this paper is organized as follows. Section 2 discusses related work. Section 3 describes the proposed clustering approach. Section 4 discusses our experiments and evaluation. Section 5 is the discussion section. Finally, Section 6 concludes the study.

Complete Article List

Search this Journal:
Reset
Volume 13: 1 Issue (2024): Forthcoming, Available for Pre-Order
Volume 12: 2 Issues (2022): 1 Released, 1 Forthcoming
Volume 11: 2 Issues (2021)
Volume 10: 2 Issues (2020)
Volume 9: 2 Issues (2019)
Volume 8: 4 Issues (2018)
Volume 7: 4 Issues (2017)
Volume 6: 4 Issues (2016)
Volume 5: 4 Issues (2015)
Volume 4: 4 Issues (2014)
Volume 3: 4 Issues (2012)
Volume 2: 4 Issues (2011)
Volume 1: 4 Issues (2010)
View Complete Journal Contents Listing