The explosive growth of enterprise and scientific data poses a huge challenge to the storage and management of massive data. In recent years, as a data intensive computing platform, Hadoop has attracted much attention from both industry and academy. In the distributed file system, in order to availability and persistent data, the data are usually divided into files of the same size block, through the form of distributed memory multi copies are stored in different geographic locations. Large cloud computing service providers have their own data center, different parts of these data centers are often distributed in the world, the data in different geographic locations can achieve fault tolerance mechanism better as a copy distribution, thus providing the persistence of high availability of data storage. Based on the Hadoop platform, this paper proposes a framework of storage and management of massive data, and analyzes key techniques that we need to implement the proposed framework.