Quality Measures for Semantic Web Application

Quality Measures for Semantic Web Application

Adiraju Prasanth Rao (Anurag Group of Institution, India)
Copyright: © 2016 |Pages: 10
DOI: 10.4018/978-1-4666-9764-5.ch005
OnDemand PDF Download:
$30.00
List Price: $37.50

Abstract

The Semantic Web is a standard of Common Data Formats on WWW with aim to convert the current web data of unstructured and semi-structured documents into a common framework that allows data to be shared and reused across applications, enterprises. The main purpose of the Semantic Web is driving the evolution of the current Web by enabling users to find, share, and combine information more easily. Humans are capable of using the Web to carry out tasks such as searching for the lowest price for a LAPTOP. However, machines cannot accomplish all of these tasks without human direction, because web pages are designed to be read by people, not machines. The semantic web is a vision of information that can be readily interpreted by machines, so machines can perform more of the tedious work involved in finding, combining, and acting upon information on the web. The chapter presents the architecture of semantic web, its challenging issues and also data quality principles. These principles provide a better decision making within organization and will maximize long term data integration and interoperability.
Chapter Preview
Top

Introduction

This is the age of information which can be gathered from different sources. Information-oriented communication networks, services and applications in various domains have recently undergone rapid changes due to sudden increases in number of network enabled devices and sensors deployed in physical environments. Within the next ten years, it is expected that billions of devices will generate large volumes of real world data for many applications and services in various domains such as smart phones, health care, transport environmental monitoring system, mobile applications etc. Data generated by these devices is expected to be mostly multimodal in nature (like temperature, light, sound and video) and manifold in character. This implies that the quality of data can change with devices, location and time.

The amount of data available on the world-wide-web is already huge and is increasing at a rapid pace. About 2.5 billion bytes of data is generated each day, which includes sensory data and data from various other sources. The generated data from different sources can be analyzed to extract actionable information which enables better understanding about the physical world and value added products and services provided by the manufacturers.

Currently, every business transaction or decision is data based. Data has also become more and more important for social activities. Large amount of business related data is published on the web with the development of internet from “Web of Documents” to “Web of Data”. Today’s web consists of large libraries of web documents and interconnected documents that are transmitted by systems and are available to the public. Applications let people view, search and combine data like calendars, address books, playlist and spreadsheet. This technology has been developed from hypertext systems to which anyone can contribute. They also reveal that the quality of information or the consistency of documents cannot be constantly assured. The present web or traditional web can be categorized into the 2nd generation. The traditional web was an association among internet users, content providers and enterprises. Originally, data was posted on web sites and users simply read or downloaded the content. This has led to the concept of the internet as a huge distributed database with users performing three major operations - search, integration and web data mining. These operations are expanded in the following paragraphs.

  • Search: The main goal of search is to identify and access the information or resources on the web.

  • Integration: Integration is the process of combining and aggregating different resources to accomplish a specialized task. For example, searching for Indian food in the United States of America requires two resources - a restaurant and Indian food items. By integrating these two resources, we can look forward to a nice dinner.

  • Web Data Mining: Web data mining is the nontrivial extraction of useful information from large data sets or databases. The author’s D. Artz & Y. Gil. (2007 [5]) discussed Special data mining techniques are used to automatically discover and extract information dynamically from web documents and services. This involves three main tasks viz. resource finding which retrieves the intended web documents, information selection and preprocessing which automatically selects and preprocesses specific information retrieved from web resources and lastly a generalization which dynamically discovers the patterns in individual web sites as well as across multiple sites.

For all three categories of internet users, the internet is entirely meant for reading and is purely display oriented and is therefore based on keyword matching only. The main reason for this is that the current internet contains mostly unstructured data. The current web involves too much manual work and needs to involve more automation. In its current version, the web data mining search mechanism could be very expensive because applications are highly specialized and application oriented.

We thus identify three major problems faced by all three internet user categories and would look for solutions to make these activities more efficient. The expected web has following features:

  • 1.

    The output data from search results should be relevant

  • 2.

    More automation is needed in the case of integration searching

  • 3.

    Web data mining needs to be less expensive.

Complete Chapter List

Search this Book:
Reset