skip to main content
10.1145/2063384.2063408acmconferencesArticle/Chapter ViewAbstractPublication PagesscConference Proceedingsconference-collections
research-article

QoS support for end users of I/O-intensive applications using shared storage systems

Published: 12 November 2011 Publication History

Abstract

While the performance of compute-bound applications can be effectively guaranteed with techniques such as space sharing or QoS-aware process scheduling, it remains a challenge to meet QoS requirements for end users of I/O-intensive applications using shared storage systems because of the difficulty of differentiating I/O services for different applications with individual quality requirements. Furthermore, it is difficult for end users to accurately specify performance goals to the storage system using I/O-related metrics such as request latency or throughput. As access patterns, request rates, and the system workload change in time, a fixed I/O performance goal, such as bounds on throughput or latency, can be expensive to achieve and may not provide performance guarantees such as bounded program execution time.
We propose a scheme supporting end-users' QoS goals, specified in terms of program execution time, in shared storage environments. We automatically translate the users' performance goals into instantaneous I/O throughput bounds using a machine learning technique, and use dynamically determined service time windows to efficiently meet the throughput bounds. We have implemented this scheme in the PVFS2 parallel file system and have conducted an extensive evaluation. Our results show that this scheme can satisfy realistic end-user QoS requirements by making highly efficient use of the I/O resources. The scheme seeks to balance programs' attainment of QoS requirements, and saves as much of the remaining I/O capacity as possible for best-effort programs.

References

[1]
E. Anderson. Simple table-based modeling of storage devices. Technical Report HPL-2001-04, HP Laboratories, 2001.
[2]
J. Bruno, J. Brustoloni, E. Gabber, B. Ozden, and A. Silberschatz. Disk scheduling with quality of service guarantees. In Proc. of the IEEE International Conference on Multimedia Computing and Systems, 1999.
[3]
J. Bent, G. Gibson, G. Grider, B. McClelland, P. Nowoczynski, J. Nunez, M. Polte, and M. Wingate. PLFS: a checkpoint file system for parallel applications. In Proc. of the Supercomputing Conference, 2009.
[4]
S. Barua, R. Thulasiram, and P. Thulasiraman. High performance computing for a financial application using Fast Fourier Transform. In Proc. of EURO-PAR, 2005.
[5]
D. Chambliss, G. Alvarez, P. Pandey, D. Jadav, J. Xu, R. Menon, and T. Lee. Performance virtualization for large-scale storage systems. In Proc. of the Symposium on Reliable Distributed systems, 2003.
[6]
M. Calzarossa and G. Serazzi. Workload characterization. In Proc. of the IEEE, 81(8), 1993.
[7]
The DiskSim simulation environment (version 3.0). http://www.pdl.cmu.edu/DiskSim/, online document, 2009.
[8]
A. Gulati, A. Merchant, and P. Varman. pClock: an arrival curve based approach for QoS guarantees in shared systems. In Proc. of the ACM SIGMETRICS Conference, 2007.
[9]
A. Gulati, I. Ahmad, C. A. Waldspurger. PARDA: Proportional allocation of resources for distributed storage access. In Proc. of the 8th USENIX Conference on File and Storage Technologies (FAST), 2009.
[10]
A. Gulati, A. Merchant, and P. Varman. mClock: Handling throughput variability for hypervisor IO scheduling. In Proc. of the 9th USENIX Symposium on Operating System Design and Implementation, 2010.
[11]
L. Huang, G. Peng, and T. Chiueh. Multi-dimensional storage virtualization. In Proc. of the ACM SIGMETRICS Conference, 2004.
[12]
Geerd-R. Hoffmann. High performance computing and networking for numerical weather prediction. In Proc. of the International Conference and Exhibition on High-Performance Computing and Networking, 1994.
[13]
Interleaved or Random (IOR) benchmarks. http://www.cs.dartmouth.edu/pario/examples.html, online document, 2010.
[14]
IBM Deep Blue, http://www.research.ibm.com/deepblue.
[15]
S. Iyer and P. Druschel. Anticipatory scheduling: A disk scheduling framework to overcome deceptive idleness in synchronous I/O. In Proc. 18th ACM Symposium on Operating Systems Principles, 2001.
[16]
W. Jin, J. Chase, and J. Kaur. Interposed proportional sharing for a storage service utility. In Proc. of the ACM SIGMETRICS Conference, 2007.
[17]
T. Kelly, I. Cohen, M. Goldszmidt, and K. Keeton. Inducing models of black-box storage arrays. Technical Report HPL-2004-108, HP Laboratories, 2004.
[18]
C. Lumb, A. Merchant, and G. Alvarezg. Facade: virtual storage devices with performance guaratees. In Proc. of the USENIX Annual Technical Conference, 2003.
[19]
M. Mesnier, M. Wachs, R. Sambasivan, A. Zheng, and G. Ganger. Modeling the relative fitness of storage. In Proc. of the ACM SIGMETRICS Conference, 2007.
[20]
NAS Parallel Benchmarks, NASA AMES Research Center, http://www.nas.nasa.gov/Software/NPB, online document, 2009.
[21]
PVFS, http://www.pvfs.org, online document, 2010.
[22]
F. Popovici, A. Arpaci-Dusseau, and R. Arpaci-Dusseau. Robust, portable i/o scheduling with the disk mimic. In Proc. of the 2003 USENIX Annual Technical Conference, 2003.
[23]
G. Peng and T. Chiueh. Availability and fairness support for storage QoS guarantee. In Proc. of the IEEE ICDCS Conference, 2008.
[24]
A. Povzner, T. Kaldewey, S. Brandt, R. Golding, T. Wong, C. Maltzahn. Efficient guaranteed disk request scheduling with fahrrad. In Proc. of the The European Conference on Computer Systems, 2008.
[25]
C. Ruemmler and J. Wilkes. An introduction to disk drive modeling. IEEE Computer, 27(3), 1994.
[26]
S3aSim I/O Benchmark, http://www-unix.mcs.anl.gov/thakur/s3asim.html, online document, 2009.
[27]
E. Smirni and D. Reed. Workload characterization of input/output intensive parallel applications. In Proc. of the 9th International Conference on Computer Performance Evaluation: Modeling and Techniques and Tools, 1997.
[28]
The Human Genome Project, http://www.ornl.gov/hgmis/.
[29]
N. Tran and D. Reed. Arima time series modeling and forecasting for adaptive I/O prefetching. In Proc. of the 15th international conference on Supercomputing, 2001.
[30]
S. Uttamchandani, L. Yin, G. Alvarez, J. Palmer, and G. Agha. Chamelon: a self-evolving, fully-adaptive resource arbitrator for storage systems. In Proc. of the USENIX Annual Technical Conference, 2005.
[31]
M. Wang, K. Au, A. Ailamaki, A. Brockwell, C. Faloutsos, and G. Ganger. Storage device performance prediction with cart models. In Proc. of the 12th MASCOTS Conference, 2004.
[32]
M. Wachs, M. Abd-El-Malek, E. Thereska, and G. Ganger. Argon: performance insulation for shared storage servers. In Proc. of the 6th USENIX Conference on File and Storage Technologies, 2007.
[33]
B. Worthington, G. Ganger, and Y. Patt. Scheduling algorithm for modern disk drives. In Proc. of the ACM SIGMETRICS Conference, 1994.
[34]
X. Zhang, K. Davis, and S. Jiang. IOrchestrator: improving the performance of multi-node I/O Systems via inter-server coordination. In Proc. of the ACM/IEEE Supercomputing Conference, 2010.
[35]
X. Zhang, Y. Xu, and S. Jiang. YouChoose: A performance interface enabling convenient and efficient QoS support for consolidated storage systems. In Proc. of 27th IEEE Symposium on Massive Storage Systems and Technologies, 2011.
[36]
J. Zhang, A. Sivasubramaniam, Q. Wang, A. Riska, and E. Riedel. Storage performance virtualization via throughput and latency control. ACM Transanction on Storage, Vol. 2, Issue 3, 2006.
[37]
Y. Zhang, E. Kung, and D. Haworth. A PDF method for multidimensional modeling of HCCI engine combustion: Effects of turbulence/chemistry interactions on ignition timing and emissions. In Proc. of the 30th International Symposium on Combustion, Vol. 30, Issue 2, 2004.

Cited By

View all
  • (2023)Taming Metadata-intensive HPC Jobs Through Dynamic, Application-agnostic QoS Control2023 IEEE/ACM 23rd International Symposium on Cluster, Cloud and Internet Computing (CCGrid)10.1109/CCGrid57682.2023.00015(47-61)Online publication date: May-2023
  • (2022)Performance modeling for I/O‐intensive applications on virtual machinesConcurrency and Computation: Practice and Experience10.1002/cpe.682334:10Online publication date: 18-Jan-2022
  • (2020)Compiler aided checkpointing using crash-consistent data structures in NVMM systemsProceedings of the 34th ACM International Conference on Supercomputing10.1145/3392717.3392755(1-13)Online publication date: 29-Jun-2020
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
SC '11: Proceedings of 2011 International Conference for High Performance Computing, Networking, Storage and Analysis
November 2011
866 pages
ISBN:9781450307710
DOI:10.1145/2063384
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 12 November 2011

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. PVFS2
  2. QoS
  3. quality of service
  4. shared storage system

Qualifiers

  • Research-article

Funding Sources

Conference

SC '11
Sponsor:

Acceptance Rates

SC '11 Paper Acceptance Rate 74 of 352 submissions, 21%;
Overall Acceptance Rate 1,516 of 6,373 submissions, 24%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)9
  • Downloads (Last 6 weeks)0
Reflects downloads up to 05 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2023)Taming Metadata-intensive HPC Jobs Through Dynamic, Application-agnostic QoS Control2023 IEEE/ACM 23rd International Symposium on Cluster, Cloud and Internet Computing (CCGrid)10.1109/CCGrid57682.2023.00015(47-61)Online publication date: May-2023
  • (2022)Performance modeling for I/O‐intensive applications on virtual machinesConcurrency and Computation: Practice and Experience10.1002/cpe.682334:10Online publication date: 18-Jan-2022
  • (2020)Compiler aided checkpointing using crash-consistent data structures in NVMM systemsProceedings of the 34th ACM International Conference on Supercomputing10.1145/3392717.3392755(1-13)Online publication date: 29-Jun-2020
  • (2020)Debugging Large-scale DatalogACM Transactions on Programming Languages and Systems10.1145/337944642:2(1-35)Online publication date: 17-Apr-2020
  • (2020)Symbolic Disintegration with a Variety of Base MeasuresACM Transactions on Programming Languages and Systems10.1145/337420842:2(1-60)Online publication date: 19-May-2020
  • (2020)A Theory of Slicing for Imperative Probabilistic ProgramsACM Transactions on Programming Languages and Systems10.1145/337289542:2(1-71)Online publication date: 17-Apr-2020
  • (2019)Persistent Octrees for Parallel Mesh Refinement through Non-Volatile Byte-Addressable MemoryIEEE Transactions on Parallel and Distributed Systems10.1109/TPDS.2018.286786730:3(677-691)Online publication date: 1-Mar-2019
  • (2019)Enforcing Crash Consistency of Evolving Network Analytics in Non-Volatile Main Memory SystemsProceedings of the International Conference on Parallel Architectures and Compilation Techniques10.1109/PACT.2019.00018(124-137)Online publication date: 23-Sep-2019
  • (2019)vPFS+: Managing I/O Performance for Diverse HPC Applications2019 35th Symposium on Mass Storage Systems and Technologies (MSST)10.1109/MSST.2019.00-16(51-64)Online publication date: May-2019
  • (2019)Enforcing Crash Consistency of Scientific Applications in Non-Volatile Main Memory Systems2019 IEEE/ACM 9th Workshop on Fault Tolerance for HPC at eXtreme Scale (FTXS)10.1109/FTXS49593.2019.00007(11-20)Online publication date: Nov-2019
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media