Grace Scheduled Maintenance - Aug 15 to Aug 17

 
Dear Grace Users,
 
Please be aware that scheduled maintenance will be performed on the Grace cluster starting on Tuesday, August 15, 2023, at 8:00 am.  Maintenance is expected to be completed by the end of day, Thursday, August 17, 2023. 
 
During this time, logins will be disabled, running jobs will be terminated, and connections via Globus will be unavailable.  We ask that you save your work, close interactive applications, and logoff the system prior to the start of the maintenance.  An email notification will be sent when the maintenance has been completed, and the cluster is available.
 
As the maintenance window approaches, the Slurm scheduler will not start any job if the job’s requested wallclock time extends past the start of the maintenance period (8:00 am on August 15, 2023).  You can run the command “htnm” (short for “hours_to_next_maintenance”) to determine the number of hours until the next maintenance period, which can aid in submitting jobs that will run before maintenance begins.  If you run squeue, such jobs will show as pending jobs with the reason “ReqNodeNotAvail.”  (If your job can actually be completed in less time than you requested, you may be able to avoid this by making sure that you request the appropriate time limit using “-t” or “–time”.)  Held jobs will automatically return to active status after the maintenance period, at which time they will run in normal priority order.
 
Jobs submitted prior to maintenance will be held and will run in RHEL8 after the cluster is returned to service.  If there are concerns that jobs will not run properly on RHEL8, please cancel your pending jobs before maintenance.
 
Upgrade to Red Hat 8
 
As part of this maintenance, the operating system on Grace will be upgraded to Red Hat 8.  A new unified software tree will be created that will be shared with the McCleary cluster. 
 
Please visit the status page at research.computing.yale.edu/system-status for the latest updates.  If you have questions, comments, or concerns, please contact us at hpc@yale.edu.
 
Sincerely,
 
Paul Gluhosky