Improving Energy-Efficiency of Computational Grids via Scheduling

Improving Energy-Efficiency of Computational Grids via Scheduling

Ziliang Zong (South Dakota School of Mines and Technology, USA), Xiaojun Ruan (Auburn University, USA), Adam Manzanares (Auburn University, USA), Kiranmai Bellam (Auburn University, USA) and Xiao Qin (Auburn University, USA)
DOI: 10.4018/978-1-60960-472-1.ch801

Abstract

High performance Grid platforms and parallel computing technologies are experiencing their golden age because of the convergence of four critical momentums: high performance microprocessors, high-speed networks, free middleware tools, and highly increased needs of computing capability. We are witnessing the rapid development of computational Grid technologies. Dozens of exciting Grid infrastructures and projects like Grid-tech, Grid Portals, Grid Fora, and Commercial Grid Initiatives are being built all over the world. However, the fast growing power consumption of data centers has caused serious concerns for building more large-scale supercomputers, clusters, and Grids. Therefore, designing energy-efficient computational Grids to make them economically attractive and environmentally friendly for parallel applications becomes highly desirable. Unfortunately, most previous studies in Grid computing primarily focus on the improvement of performance, security, and reliability, while completely ignoring the energy conservation issue. To address this problem, we propose a general architecture for building energy-efficient computational Grids and discuss the potential possibilities for incorporating power-aware techniques to different layers of the proposed Grid architecture. In this chapter, we first provide necessary background on computational Grids, Grid computing, and parallel scheduling. Next, we illustrate the general Grid architecture and explain the functionality of different layers. Followed by that, we discuss the design and implementation details of applying the energy-efficient job-scheduling technique, which is called Communication Energy Conservation Scheduling (or CECS for short), to computational Grids. Finally, we present extensive simulation results to prove the improvement of energy-efficiency of computational Grids.
Chapter Preview
Top

1. Introduction

We are now in an era of information explosion. Billions of data is generated in the time it takes to blink your eyes. In order to process these massive data sets, large-scale high-performance computing platforms, like clusters and computational Grids, have been widely deployed. This accommodates the rapid growth of cluster and Grid applications in both academic and the commercial area. A large fraction of applications running in these high-performance computing platforms are computing-intensive and storage-intensive, since these applications deal with a large amount of data transferred either between memory and storage systems or among hundreds of computing nodes via interconnection networks. There is no doubt that high performance computing has significantly changed our lives today. We will definitely benefit more once computation power provided by Grids become standard and pervasive similar to electric and water services.

However, these large computational power increases come at a cost. Increasing evidence has shown that the powerful computing capability is actually at the cost of huge energy consumption demands. For example, Energy User News stated that the power requirements of today’s data centers range from 75 W/ft2 to 150-200 W/ft2 and will increase to 200-300 W/ft2 in the near future (Moore, 2002). The new data center capacity projected for 2005 in the U.S. would require approximately 40 TWh ($4B at $100 per MWh) per year to run 24x7 unless they become more efficient (Kappiah et al., 2005). The supercomputing center in Seattle is forecast to increase the city's power demands by 25% (Bryce, 2000). The Environment Protection Agency reported that the total energy consumption of servers and data centers of the United States was 61.4 billion KWh in 2006, which is more than double the energy usage for the same purpose in 2000 (Environment Protection Agency, 2006). Even worse, the EPA has predicted that the power usage of servers and data centers will be doubled again within five years if historical trends continue (Environment Protection Agency, 2006). Even worse, huge-energy consumption will simultaneously cause serious environmental pollution. Based on the data from EPA, generating 1 kWh of electricity in the United States results in an average 1.55 pounds (lb) of carbon dioxide (CO2) emissions. Therefore, it is highly desirable to design economically attractive and environmentally friendly supercomputing platforms like computational Grids.

Complete Chapter List

Search this Book:
Reset