Big Data and Service Science

Big Data and Service Science

Tu-Bao Ho, Siriwon Taewijit, Quang-Bach Ho, Hieu-Chi Dam
DOI: 10.4018/978-1-4666-9562-7.ch009
OnDemand:
(Individual Chapters)
Available
$37.50
No Current Special Offers
TOTAL SAVINGS: $37.50

Abstract

Big data is about handling huge and/or complex datasets that conventional technologies cannot handle or handle well. Big data is currently receiving tremendous attention from both industry and academia as there is much more data around us than ever before. This chapter addresses the relationship between big data and service science, especially how big data can contribute to the process of co-creation of service value. In particular, the value co-creation in terms of customer relationship management is mentioned. The chapter starts with brief descriptions of big data, machine learning and data mining methods, service science and its model of value co-creation, and then addresses the key idea of how big data can contribute to co-create service value.
Chapter Preview
Top

Big Data

What Is Big Data?

Big Data is an emerging issue in information technology (IT) in the last two years. In his paper “Big Data: Defining its definition” on ZDNet (2012), Andrew Brust gave a widely accepted definition: “Big Data is about the technologies and practice of handling huge data sets that conventional database management systems cannot handle them efficiently, and sometimes cannot handle them at all. Often these data sets are fast-streaming too, meaning practitioners don’t have lots of time to analyze them in a slow, deliberate manner, because the data just keeps coming.”

In other words, Big Data are datasets that are very big and/or very complex that the current IT methods and tools cannot handle them well. It is worth noting that the name “Big Data” can mislead people to think of only the big size of the data, but ignore or do not know about the complexity of the data. However, the big and complex aspects of such data always go together, in which the complexity of the data is even more typical and challenging for “Big Data” than the big size. We can also see this feature in the IBM definition of Big Data by four dimensions of Variety, Velocity, Volume and Veracity. The first dimension (variety) is about the cross-links between data sources which can be any type of data structured and unstructured data such as text, sensor data, audio, video, click streams, log files and more. The second dimension (velocity) is about “time-sensitive processes such as catching fraud, Big Data must be used as it streams into your enterprise in order to maximize its value to your business.” The third dimension (volume) is about the ever-growing data of all types, at levels of terabytes (1012), petabytes (1015 bytes), and even zetabytes (1018 bytes). The four dimension (veracity) is about truthfulness, accuracy or precision, correctness of the data. It is about uncertainty due to data inconsistency and incompleteness, ambiguities, latency, deception, or model approximations.

It has been said that Cloud Computing, Smart-Devices, and Big Data are the three emerging IT technologies that can strongly influence other disciplines. We all know that data is the source of all information that people can have. However, information usually is not ready to use but requires to be analyzed (processed) for being usable. Roughly speaking, the bigger and the more complex data the more difficult to analyze it, and therefore in many cases we still not be able to do it well (Franks, 2012).

Where Does Big Data Come From?

The phenomenon that we have much data around us than before is objective. There are Big Data in many organizations, in social, business, scientific activities that potentially contains big values. But where does Big Data come from?

The Big Data come from various sources, typically the three followings:

  • 1.

    Social Media Data: For example each day there are 230 millions tweets on twitters in the world, or 2.7 billion comments to Facebook, and 86,400 hours of video to YouTube.

  • 2.

    Machine Data: Such as industrial equipment, sensors and monitor machinery, web logs tracks user behavior online. For example, the Large Hadron collider at CERN (the European Organization for Nuclear Research) generates 40 terabytes of data each second.

  • 3.

    Transactional Data: Such as Product IDs, prices, payment, manufacturer and distributor data, and much more. For example, the products sales of Amazon.com in the third Quarter of 2011 are about 10 billion USD, or the pizza chain Domino reaches 1 million customers per day.

Complete Chapter List

Search this Book:
Reset