Please wait a minute...
Frontiers of Computer Science

ISSN 2095-2228

ISSN 2095-2236(Online)

CN 10-1014/TP

Postal Subscription Code 80-970

2018 Impact Factor: 1.129

Front. Comput. Sci.    2017, Vol. 11 Issue (1) : 75-87    https://doi.org/10.1007/s11704-016-6036-y
RESEARCH ARTICLE
HWM: a hybrid workload migration mechanism of metadata server cluster in data center
Jian LIU1,2,Huanqing DONG1,Junwei ZHANG1,Zhenjun LIU1,Lu XU1()
1. Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
2. University of Chinese Academy of Sciences, Beijing 100049, China
 Download: PDF(710 KB)  
 Export: BibTeX | EndNote | Reference Manager | ProCite | RefWorks
Abstract

In data center, applications of big data analytics pose a big challenge to massive storage systems. It is significant to achieve high availability, high performance and high scalability for PB-scale or EB-scale storage systems. Metadata server (MDS) cluster architecture is one of the most effective solutions to meet the requirements of applications in data center. Workload migration can achieve load balance and energy saving of cluster systems. In this paper, a hybrid workload migration mechanism of MDS cluster is proposed and named as HWM. In HWM, workload of MDS is classified into two categories: metadata service and state service, and they can be migrated rapidly from a source MDS to a target MDS in different ways. Firstly, in metadata service migration, all the dirty metadata of one sub file system is flushed to a shared storage pool by the source MDS, and then is loaded by the target MDS. Secondly, in state service migration, all the states of that sub file system are migrated from source MDS to target MDS through network at file granularity, and then all of the related structures of these states are reconstructed in targetMDS. Thirdly, in the process of workload migration, instead of blocking client requests, the source MDS can decide which MDS will respond to each request according to the operation type and the migration stage. The proposed mechanismis implemented in the BlueWhaleMDS cluster. The performance measurements show that the HWM mechanism is efficient to migrate the workload of a MDS cluster system and provides low-latency access to metadata and states.

Keywords data center      metadata server cluster      hybrid workload migration      metadata service      state service      lowlatency access     
Corresponding Author(s): Lu XU   
Just Accepted Date: 28 September 2016   Issue Date: 11 January 2017
 Cite this article:   
Jian LIU,Huanqing DONG,Junwei ZHANG, et al. HWM: a hybrid workload migration mechanism of metadata server cluster in data center[J]. Front. Comput. Sci., 2017, 11(1): 75-87.
 URL:  
https://academic.hep.com.cn/fcs/EN/10.1007/s11704-016-6036-y
https://academic.hep.com.cn/fcs/EN/Y2017/V11/I1/75
1 Turner V, Gantz J F, Reinsel D, Minton S. The digital universe of opportunities: rich data and the increasing value of the Internet of things. IDC Analyze the Future, 2014
2 Ghemawat S, Gobioff H, Leung S T. The Google file system. ACM SIGOPS Operating Systems Review, 2003, 37(5): 29–43
https://doi.org/10.1145/1165389.945450
3 McKusick K, Quinlan S. GFS: evolution on fast-forward. Communications of the ACM, 2010, 53(3): 42–49
https://doi.org/10.1145/1666420.1666439
4 Makoto S, Hiroki K, Shoji K. Performance evaluation of scaleout NAS for HDFS. In: Proceedings of the 3rd International Conference on Advances in Information Mining and Management. 2013, 32–35
5 Xia M, Saxena M, Blaum M, Pease D A. A tale of two erasure codes in HDFS. In: Proceedings of the 13th USENIX Conference on File and Storage Technologies. 2015, 213–226
6 Jain R, Sarkar P, Subhraveti D. GPFS-SNC: an enterprise cluster file system for big data. IBM Journal of Research and Development, 2013, 57(3/4): 5:1–5:10
7 Davies A, Orsaria A. Scale out with GlusterFS. Linux Journal, 2013, (235): 1
8 Chasapis K, Dolz M F, Kuhn M, Ludwig T. Evaluating Lustre’s metadata server on a multi-socket platform. In: Proceedings of the 9th Parallel Data Storage Workshop. 2014, 13–18
https://doi.org/10.1109/pdsw.2014.5
9 Kim T, Noh S H. PNFS for everyone: an empirical study of a low-cost, highly scalable networked storage. International Journal of Computer Science and Network Security, 2014, 14(3): 52–59
10 Weil S A, Brandt S A, Miller E L, Long D D, Maltzahn C. Ceph: a scalable, high-performance distributed file system. In: Proceedings of the 7th Symposium on Operating Systems Design and Implementation. 2006, 307–320
11 Wang F, Nelson M, Oral S, Atchley S, Weil S, Settlemyer B W, Caldwell B, Hill J. Performance and scalability evaluation of the Ceph parallel file system. In: Proceedings of the 8th Parallel Data StorageWorkshop. 2013, 14–19
https://doi.org/10.1145/2538542.2538562
12 Sevilla M A, Watkins N, Maltzahn C, Nassi I, Brandt S A, Weil S A, Farnum G, Fineberg S. Mantle: a programmable metadata load balancer for the Ceph file system. In: Proceedings of the 27th International Conference for High Performance Computing, Networking, Storage and Analysis. 2015, 1–12
https://doi.org/10.1145/2807591.2807607
13 Sinnamohideen S, Sambasivan R R, Hendricks J, Liu L, Ganger G R. A transparently-scalable metadata service for the UrsaMinor storage system. In: Proceedings of USENIX Annual Technical Conference. 2010, 13–26
14 Abd-El-Malek M, Courtright IIWV, Cranor C, Ganger G R, Hendricks J, Klosterman A J, Mesnier M P, Prasad M, Salmon B, Sambasivan R R, S Sinnamohideen, Strunk J D, Thereska E, Wachs M, Wylie J J. Ursa Minor: versatile cluster-based storage. In: Proceedings of the 4th USENIX Conference on File and Storage Technologies. 2005, 59–72
15 Menon J, Pease D A, Rees R, Duyanovich L, Hillsberg B. IBM Storage Tank—a heterogeneous scalable SAN file system. IBM Systems Journal, 2003, 42(2): 250–267
https://doi.org/10.1147/sj.422.0250
16 Thomasian A. Storage research in industry and universities. ACM SIGARCH Computer Architecture News, 2010, 38(2): 1–48
https://doi.org/10.1145/1823838.1823840
17 An overview of NFSv4: NFSv4.0, NFSv4.1, pNFS, and proposed NFSv4.2 features. SNIA Ethernet Storage Forum, 2012
18 Mohr R, Peltz Jr P. Benchmarking SSD-based Lustre file system configurations. In: Proceedings of the 2014 Annual Conference on Extreme Science and Engineering Discovery Environment. 2014
https://doi.org/10.1145/2616498.2616544
19 Aishwarya K, Sreevatson M, Babu C, Prabavathy B. Efficient prefetching technique for storage of heterogeneous small files in hadoop dis tributed file system federation. In: Proceedings of the 15th International Conference on Advanced Computing. 2013, 523–530
20 Chen G, Jagadish H, Jiang D, Maier D, Ooi B C, Tan K L, Tan W C. Federation in cloud data management: challenges and opportunities. IEEE Transactions on Knowledge and Data Engineering, 2014, 26(7): 1670–1678
https://doi.org/10.1109/TKDE.2014.2326659
21 Patil S, Gibson G A. Scale and concurrency of GIGA+: file system directories with millions of files. In: Proceedings of the 9th USENIX Conference on File and Storage Technologies. 2011, 1–14
22 Patil S V, Gibson G A, Lang S, Polte M. GIGA+: scalable directories for shared file systems. In: Proceedings of the 2nd International Workshop on Petascale Data Storage. 2007, 26–29
https://doi.org/10.1145/1374596.1374604
23 Douceur J R, Howell J. Distributed directory service in the Farsite file system. In: Proceedings of the 7th Symposium on Operating Systems Design and Implementation. 2006, 321–334
24 Ma H, Liu Z, Zhang H, Feng S, Han X, Xu L. Experiences with hierarchical storage management support in blue whale file system. In: Proceedings of the 11th International Conference on Parallel and Distributed Computing, Applications and Technologies. 2010, 369–374
https://doi.org/10.1109/pdcat.2010.35
25 Solar R, Gil-Costa V, Marin M. Dynamic load balance for approximate parallel simulations with consistent hashing. In: Proceedings of the 47th Summer Computer Simulation Conference. 2015, 1–10
26 Xu Z Y, Wang X X. A predictive modified round robin scheduling algorithm for Web server clusters. In: Proceedings of the 34th Chinese Control Conference. 2015, 5804–5808
27 Xia Y, Dobra A, Han S C. Multiple-choice random network for server load balancing. In: Proceedings of the 26th IEEE International Conference on Computer Communications. 2007, 1982–1990
https://doi.org/10.1109/infcom.2007.230
28 Wu Y, Luo S, Li Q. An adaptive weighted least-load balancing algorithm based on server cluster. In: Proceedings of the 5th International Conference on Intelligent Human-Machine Systems and Cybernetics. 2013, 224–227
https://doi.org/10.1109/ihmsc.2013.60
29 Allayear S M, Salahuddin M, Ahmed F, Park S S. Introducing iSCSI protocol on online based mapreduce mechanism. In: Proceedings of the 14th International Conference on Computational Science and Its Applications. 2014, 691–706
https://doi.org/10.1007/978-3-319-09156-3_48
30 Guo T, Shen Y L, Liu Z J, Xu L. BW-FILERAID: a kind of file based distributed RAID system and optimization. Applied Mechanics and Materials, 2011, 80: 1208–1216
https://doi.org/10.4028/www.scientific.net/AMM.80-81.1208
31 Chen Y, Wu F, Yu K, Zhang L, Chen Y, Yang Y, Mao J. Instant bug testing service for linux kernel. In: Proceedings of the 10th IEEE International Conference on High Performance Computing and Communications. 2013, 1860–1865
https://doi.org/10.1109/hpcc.and.euc.2013.347
[1] Yudong QIN, Deke GUO, Lailong LUO, Geyao CHENG, Zeliu DING. Design and optimization of VLC based small-world data centers[J]. Front. Comput. Sci., 2019, 13(5): 1034-1047.
[2] Yihong GAO, Huadong MA. StreamTune: dynamic resource scheduling approach for workload skew in video data center[J]. Front. Comput. Sci., 2018, 12(4): 669-681.
[3] Hui DOU, Yong QI. An online electricity cost budgeting algorithm for maximizing green energy usage across data centers[J]. Front. Comput. Sci., 2017, 11(4): 661-674.
[4] Ruihong LIN,Yuhui DENG. Allocating workload to minimize the power consumption of data centers[J]. Front. Comput. Sci., 2017, 11(1): 105-118.
[5] Chuang LIN,Min YAO,Yin LI. Joint study on VMs deployment, assignment and migration in geographically distributed data centers[J]. Front. Comput. Sci., 2016, 10(3): 559-573.
[6] Bing YU,Yanni HAN,Hanning YUAN,Xu ZHOU,Zhen XU. A cost-effective scheme supporting adaptive service migration in cloud data center[J]. Front. Comput. Sci., 2015, 9(6): 875-886.
[7] Lailong LUO,Deke GUO,Wenxin LI,Tian ZHANG,Junjie XIE,Xiaolei ZHOU. Compound graph based hybrid data center topologies[J]. Front. Comput. Sci., 2015, 9(6): 860-874.
[8] Jian LIN, Li ZHA, Zhiwei XU. Consolidated cluster systems for data centers in the cloud age: a survey and analysis[J]. Front. Comput. Sci., 2013, 7(1): 1-19.
[9] Kaishun WU, Jiang XIAO, Lionel M. NI. Rethinking the architecture design of data center networks[J]. Front Comput Sci, 2012, 6(5): 596-603.
[10] Hui CHEN, Ping LU, Pengcheng XIONG, Cheng-Zhong XU, Zhiping WANG. Energy-aware application performance management in virtualized data centers[J]. Front Comput Sci, 2012, 6(4): 373-387.
[11] Chunjie LUO, Jianfeng ZHAN, Zhen JIA, Lei WANG, Gang LU, Lixin ZHANG, Cheng-Zhong XU, Ninghui SUN. CloudRank-D: benchmarking and ranking cloud computing systems for data processing applications[J]. Front Comput Sci, 2012, 6(4): 347-362.
[12] Yuzhong SUN, Yiqiang ZHAO, Ying SONG, Yajun YANG, Haifeng FANG, Hongyong ZANG, Yaqiong LI, Yunwei GAO. Green challenges to system software in data centers[J]. Front Comput Sci Chin, 2011, 5(3): 353-368.
[13] Zhenyong CHEN, Wei FAN, Zhang XIONG, Pingan ZHANG, Lixin LUO, . Visual data security and management for smart cities[J]. Front. Comput. Sci., 2010, 4(3): 386-393.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed