1
FusionFS: a distributed file system for large scale data-intensive computing Acknowledgement This work is supported by NSF grant OCI-1054974 Performance on Blue Gene/P Performance on Amazon cloud and Linux cluster Develop both theoretical and practical aspects of building distributed files systems scalable to exascale supporting millions of nodes and billions of concurrent IO requests Goal Performance on Applications References [1] FusionFS project website: http://datasys.cs.iit.edu/projects/FusionFS/index.html [2] Ioan Raicu, Ian Foster and Pete Beckman. Making a Case for Distributed File Systems at Exascale, ACM Workshop on Large-scale System and Application Performance (LSAP), 2011 FusionFS delivers significantly higher throughput of data and metadata than other leading file systems FusionFS provides a completely distributed and loosely-coupled metadata management FusionFS maximizes the data locality for file writes, making it a perfect fit for checkpointings FusionFS supports different replication semantics, as well as efficient data provenance We are scaling FusionFS to O(10K) nodes, and developing a library to bypass FUSE Conclusion and Future Work Dongfang Zhao 1 , Chen Shou 1 , Zhao Zhang 2 Iman Sadooghi 1 , Xiaobing Zhou 1 , Tonglin Li 1 , Ioan Raicu 1,3 1 Department of Computer Science, Illinois Institute of Technology 2 Department of Computer Science, University of Chicago 3 Mathematics and Computer Science Division, Argonne National Laboratory Motivation Current architecture (i.e. compute nodes are remotely connected to storage nodes) would unlikely scale well at exascale Proposed Architecture Distributed file system coexists on compute nodes Software Stack Data Throughput Metadata Ops/sec Provenance Efficiency Replication Efficiency (128 nodes) Amazon m1.medium instance FusionFS vs. HDFS on Linux cluster Bioinformatics application: BLAST No. 1 write-intensive application (Dec. 2011) No. 2 write-intensive application (Dec. 2011) No. 3 write-intensive application (Dec. 2011)

Performance on Blue Gene/Pdatasys.cs.iit.edu/reports/2013_GCASR13_poster104.pdf · Performance on Blue Gene/P . Performance on Amazon cloud and Linux cluster . Develop both theoretical

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Performance on Blue Gene/Pdatasys.cs.iit.edu/reports/2013_GCASR13_poster104.pdf · Performance on Blue Gene/P . Performance on Amazon cloud and Linux cluster . Develop both theoretical

FusionFS: a distributed file system for large scale data-intensive computing

Acknowledgement This work is supported by NSF grant OCI-1054974

Performance on Blue Gene/P

Performance on Amazon cloud and Linux cluster

Develop both theoretical and practical aspects of building distributed files systems scalable to exascale supporting millions of nodes and billions of concurrent IO requests

Goal

Performance on Applications

References [1] FusionFS project website: http://datasys.cs.iit.edu/projects/FusionFS/index.html [2] Ioan Raicu, Ian Foster and Pete Beckman. Making a Case for Distributed File Systems at Exascale, ACM Workshop on Large-scale System and Application Performance (LSAP), 2011

• FusionFS delivers significantly higher throughput of data and metadata than other leading file systems • FusionFS provides a completely distributed and loosely-coupled metadata management • FusionFS maximizes the data locality for file writes, making it a perfect fit for checkpointings • FusionFS supports different replication semantics, as well as efficient data provenance • We are scaling FusionFS to O(10K) nodes, and developing a library to bypass FUSE

Conclusion and Future Work

Dongfang Zhao1, Chen Shou1, Zhao Zhang2 Iman Sadooghi1, Xiaobing Zhou1, Tonglin Li1, Ioan Raicu1,3

1Department of Computer Science, Illinois Institute of Technology 2Department of Computer Science, University of Chicago

3Mathematics and Computer Science Division, Argonne National Laboratory

Motivation Current architecture (i.e. compute nodes are remotely connected to storage nodes) would unlikely scale well at exascale

Proposed Architecture Distributed file system coexists on compute nodes

Software Stack

Data Throughput Metadata Ops/sec

Provenance Efficiency Replication Efficiency (128 nodes)

Amazon m1.medium instance FusionFS vs. HDFS on Linux cluster

Bioinformatics application: BLAST No. 1 write-intensive application (Dec. 2011)

No. 2 write-intensive application (Dec. 2011) No. 3 write-intensive application (Dec. 2011)