Transition Hangout - Slurm Workload Manager · 2016-11-28 · Transition Hangout - Site Overview...

3
Slurm Transition Hangout – SC16 Salt Lake City, Nov 2016 1 1 Transition Hangout (aka: How we converted to Slurm) Q-n-A with: Ryan Cox - BYU – Fulton Supercomputing Lab Bruce Pfaff - GSFC - NASA Center for Climate Simulation

Transcript of Transition Hangout - Slurm Workload Manager · 2016-11-28 · Transition Hangout - Site Overview...

Page 1: Transition Hangout - Slurm Workload Manager · 2016-11-28 · Transition Hangout - Site Overview BYU Fulton Supercomputing Lab NASA Goddard NASA Center for Climate Simulation Size

Slurm Transition Hangout – SC16 Salt Lake City, Nov 2016 1 1

Transition Hangout(aka: How we converted to Slurm)

Q-n-A with:

Ryan Cox - BYU – Fulton Supercomputing Lab

Bruce Pfaff - GSFC - NASA Center for Climate Simulation

Page 2: Transition Hangout - Slurm Workload Manager · 2016-11-28 · Transition Hangout - Site Overview BYU Fulton Supercomputing Lab NASA Goddard NASA Center for Climate Simulation Size

Slurm Transition Hangout – SC16 Salt Lake City, Nov 2016 2 2

Transition Hangout - Site Overview

BYU Fulton Supercomputing Lab

NASA Goddard NASA Center for Climate Simulation

Size of Cluster(s) ~1000 nodes (~5 clusters) 3500+ Compute nodes 100+ I/O, Admin, Misc nodes

Previous Resource Manger Moab/Torque PBS Pro

Conversion Date Feb 2013 Oct 2013

Length Of Conversion 1 week 10 days

Page 3: Transition Hangout - Slurm Workload Manager · 2016-11-28 · Transition Hangout - Site Overview BYU Fulton Supercomputing Lab NASA Goddard NASA Center for Climate Simulation Size

Slurm Transition Hangout – SC16 Salt Lake City, Nov 2016 3 3

Common QuestionsWhy did you switch from your previous Resource Manager?How did your administrators deal with the transition?How did your users react?Do the PBS wrappers REALLY work?Do you have a support contract?

How is SchedMD support?