Understanding system characteristics of online erasure coding on scalable, distributed and large-scale SSD array systems

Cited 0 time in webofscience Cited 2 time in scopus
  • Hit : 153
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorKoh, Sungjoonko
dc.contributor.authorZhang, Jieko
dc.contributor.authorKwon, Miryeongko
dc.contributor.authorYoon, Jungyeonko
dc.contributor.authorDonofrio, Davidko
dc.contributor.authorKim, Nam Sungko
dc.contributor.authorJung, Myoungsooko
dc.date.accessioned2019-12-13T12:30:58Z-
dc.date.available2019-12-13T12:30:58Z-
dc.date.created2019-11-28-
dc.date.issued2017-10-01-
dc.identifier.citation2017 IEEE International Symposium on Workload Characterization, IISWC 2017, pp.76 - 86-
dc.identifier.urihttp://hdl.handle.net/10203/269606-
dc.description.abstractLarge-scale systems with arrays of solid state disks (SSDs) have become increasingly common in many computing segments. To make such systems resilient, we can adopt erasure coding such as Reed-Solomon (RS) code as an alternative to replication because erasure coding can offer a significantly lower storage cost than replication. To understand the impact of using erasure coding on system performance and other system aspects such as CPU utilization and network traffic, we build a storage cluster consisting of approximately one hundred processor cores with more than fifty high-performance SSDs, and evaluate the cluster with a popular open-source distributed parallel file system, Ceph. Then we analyze behaviors of systems adopting erasure coding from the following five viewpoints, compared with those of systems using replication: (1) storage system I/O performance; (2) computing and software overheads; (3) I/O amplification; (4) network traffic among storage nodes; (5) the impact of physical data layout on performance of RS-coded SSD arrays. For all these analyses, we examine two representative RS configurations, which are used by Google and Facebook file systems, and compare them with triple replication that a typical parallel file system employs as a default fault tolerance mechanism. Lastly, we collect 54 block-level traces from the cluster and make them available for other researchers.-
dc.languageEnglish-
dc.publisherIEEE Computer Society-
dc.titleUnderstanding system characteristics of online erasure coding on scalable, distributed and large-scale SSD array systems-
dc.typeConference-
dc.type.rimsCONF-
dc.citation.beginningpage76-
dc.citation.endingpage86-
dc.citation.publicationname2017 IEEE International Symposium on Workload Characterization, IISWC 2017-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationSeattle, Washington-
dc.identifier.doi10.1109/IISWC.2017.8167758-
dc.contributor.localauthorJung, Myoungsoo-
dc.contributor.nonIdAuthorKoh, Sungjoon-
dc.contributor.nonIdAuthorZhang, Jie-
dc.contributor.nonIdAuthorKwon, Miryeong-
dc.contributor.nonIdAuthorYoon, Jungyeon-
dc.contributor.nonIdAuthorDonofrio, David-
dc.contributor.nonIdAuthorKim, Nam Sung-
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0