High Availability on Jetstream: Practices and Lessons Learned. Lowe, J., M., Fischer, J., Sudarshan, S., Turner, G., Stewart, C., A., & Hancock, D., Y. In Proceedings of the 9th Workshop on Scientific Cloud Computing (ScienceCloud'18), of ScienceCloud'18, pages 4:1--4:7, 2018. ACM.
High Availability on Jetstream: Practices and Lessons Learned [link]Website  doi  abstract   bibtex   
Research computing has traditionally used high performance computing (HPC) clusters and has been a service not given to high availability without a doubling of computational and storage capacity. System maintenance such as security patching, firmware updates, and other system upgrades generally meant that the system would be unavailable for the duration of the work unless one has redundant HPC systems and storage. While efforts were often made to limit downtimes, when it became necessary, maintenance windows might be one to two hours or as much as an entire day. As the National Science Foundation (NSF) began funding non-traditional research systems, looking at ways to provide higher availability for researchers became one focus for service providers. One of the design elements of Jetstream was to have geographic dispersion to maximize availability. This was the first step in a number of design elements intended to make Jetstream exceed the NSF's availability requirements. We will examine the design steps employed, the components of the system and how the availability for each was considered in deployment, how maintenance is handled, and the lessons learned from the design and implementation of the Jetstream cloud.
@inproceedings{
 title = {High Availability on Jetstream: Practices and Lessons Learned},
 type = {inproceedings},
 year = {2018},
 keywords = {Atmosphere,XSEDE,acm reference format,atmosphere,availability,cloud,george turner,hpc,jeremy fischer,john michael lowe,research,sanjana sudarshan,xsede},
 pages = {4:1--4:7},
 websites = {http://doi.acm.org/10.1145/3217880.3217884},
 publisher = {ACM},
 city = {New York, NY, USA},
 series = {ScienceCloud'18},
 id = {5d4797f1-4581-3845-971d-edfa984ebcc2},
 created = {2019-10-01T17:20:35.320Z},
 file_attached = {false},
 profile_id = {42d295c0-0737-38d6-8b43-508cab6ea85d},
 last_modified = {2019-10-01T17:20:35.320Z},
 read = {false},
 starred = {false},
 authored = {true},
 confirmed = {true},
 hidden = {false},
 citation_key = {Lowe:2018:HAJ:3217880.3217884},
 source_type = {inproceedings},
 private_publication = {false},
 abstract = {Research computing has traditionally used high performance computing (HPC) clusters and has been a service not given to high availability without a doubling of computational and storage capacity. System maintenance such as security patching, firmware updates, and other system upgrades generally meant that the system would be unavailable for the duration of the work unless one has redundant HPC systems and storage. While efforts were often made to limit downtimes, when it became necessary, maintenance windows might be one to two hours or as much as an entire day. As the National Science Foundation (NSF) began funding non-traditional research systems, looking at ways to provide higher availability for researchers became one focus for service providers. One of the design elements of Jetstream was to have geographic dispersion to maximize availability. This was the first step in a number of design elements intended to make Jetstream exceed the NSF's availability requirements. We will examine the design steps employed, the components of the system and how the availability for each was considered in deployment, how maintenance is handled, and the lessons learned from the design and implementation of the Jetstream cloud.},
 bibtype = {inproceedings},
 author = {Lowe, John Michael and Fischer, Jeremy and Sudarshan, Sanjana and Turner, George and Stewart, Craig A and Hancock, David Y},
 doi = {10.1145/3217880.3217884},
 booktitle = {Proceedings of the 9th Workshop on Scientific Cloud Computing (ScienceCloud'18)}
}

Downloads: 0