GT Home : : Campus Maps : : GT Directory

HPC Status Update – 4/23/2010

This entry was posted by on Friday, 23 April, 2010 at

Much progress has been made over the last week, and we’re on track for availability in early May.  We’ve heard back from a couple of you with names for your clusters.  If you haven’t replied with a desired name, please do so soon.  If you have comments one way or the other regarding the purchase of the Jacket or Matlab Distributed toolkit, please let me know.  We need to make purchasing decisions soon.

Base networking – 95% complete

  • 1 gig links to BCDC (for backups) – complete
  • 475 names and addresses defined in DNS & DHCP – complete
  • 360 ethernet ports configured – complete
  • dual 10GbE uplink from HPC backbone to campus backbone – complete
  • 10 GbE uplink from Inchworm to HPC backbone – complete
  • second 10GbE uplink Inchworm to HPC backbone – next week
  • 10 GbE links to BCDC (replace 1gig) – next week

home & project directories – 95% complete

  • iSCSI targets – complete
  • configure dual 10GbE interfaces on storage servers – complete
  • configure ZFS filesystems on storage servers – complete
  • provision user filesystems and set quotes – next week
  • configure backups for user filesystems – next week

scratch storage – 80% complete

  • Panasas setup and configuration – complete
  • Infiniband router setup and configuration – complete
  • basic host network configuration – complete
  • Panasas client software install & configuration – next week

server infrastructure

  • install OS and configure support servers
  • install OS and configure head nodes (we need your names!)
  • install and configuration of DNS & DHCP appliances – complete

compute nodes

  • support scripts – 90% complete
  • creation of diskless system images (16 types) – 20% complete
  • 8 Community Cluster nodes online
  • bringup of ~275 compute nodes

Moab workload scheduler

  • creation and testing of prologue & epilogue scripts
  • initial configuration of scheduler queues

Software

  • GSL, GIT, ACML – installed
  • Intel Compiler Suite – installed
  • Portland Group PGI Server Complete – installed
  • mvapich2 (w/ gcc, intel and PGI permutations) – installed
  • mpich2 (w/ gcc, intel and PGI permutations) – installed
  • mpich (w/ gcc and intel permutations) – installed
  • ATLAS – in progress
  • lammps – in progress
  • Jacket & Matlab distributed toolkit – under discussion
  • GPU software

Penguin tasks

  • 13 of 50 Penguin blades out for capacitor/diode/resistor repair
  • 50 of 50 Penguin blades in need of BIOS information fix

Comments are closed.