Cloud computing has demonstrated that processing very large datasets over commodity can be done simply, given the right programming model and infrastructure. In paper, we describe the design and implementation of the Sector storage cloud and Sphere compute cloud. By contrast with the existing storage and compute clouds, can manage data not only within a data centre, but also across geographically data centres. Similarly, the Sphere compute cloud supports user-defined (UDFs) over data both within and across data centres. As a special case, style programming can be implemented in Sphere by using a Map UDF by a Reduce UDF. We describe some experimental studies comparing /Sphere and Hadoop using the Terasort benchmark. In these studies, Sector is twice as fast as Hadoop. Sector/Sphere is open source. © 2009 The Royal Society.
CITATION STYLE
Gu, Y., & Grossman, R. L. (2009). Sector and sphere: The design and of a high-performance data cloud. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 367(1897), 2429–2445. https://doi.org/10.1098/rsta.2009.0053
Mendeley helps you to discover research relevant for your work.