RCC News

Signup for our Newsletter

We publish a monthly newsletter which showcases our events and opportunities at the Research Computing Center.

Signup »

View Archives

News Archive

System Upgrades (including Slurm) to occur July 13 - 19

We have scheduled our Slurm and RHEL7 upgrade to occur on Monday, July 13 through Sunday, July 19. During this time, our HPC and Spear systems will be unavailable.

MOAB to Slurm Migration Guide Published

We have published our MOAB to Slurm migration guide! This guide provides an overview of how to migrate your MOAB scripts to use the new Slurm scheduler.

Check it out: https://rcc.fsu.edu/docs/moab-slurm-migration-guide

Planned Spear Outage - Friday, May 1: 5-8pm

We are turning the Spear system off from 5-8pm on Friday, May 1 for a planned power outage.

Lustre Issues

UPDATE Apr 21 - We are still working our Lustre issues. Spear and Lustre are still unavailable.

Yesterday (Sunday, April 19), the data center in Dirac suffered a brief power outage. Most systems are back online now, but there are lingering issues with the Lustre filesystem. This means that Spear systems are unavailable as well.

We are currently working on a solution, and will post an update as soon as we have resolved the issues.

Scheduler Changes coming this Summer

This summer, we are going to replace the scheduling and job management software on the High Performance Computing Cluster with a new package called Slurm. Slurm will replace the current MOAB/Torque workload manager that we have been using for the past eight years. This will affect all HPC users.

Journal publication by RCC staff member

A publication by Bin Chen, a member of applications group has been accepted for publication in the journal Astrophysical Journal Supplement Series. The article titled "Algorithms And Programs For Strong Gravitational Lensing In Kerr Space-time Including Polarization" used the HPC, Matlab and cython for performance analysis.

Benchmarking MATLAB and Python at RCC

HPC jobs are traditionally compiled and run in low-level languages, such as C and Fortran. The reason for this—speed. Parallel libraries like OpenMP, MPI or gpGPU accelerators enable code to run in parallel faster on more hardware. Conversely, higher level interpreted languages like Python and MATLAB are much easier for scientists to work with, but are generally magnitudes slower than the lower-level compiled languages. However, these higher-level languages are significantly improving in speed and flexibility.

HPC Scheduler Problems RESOLEVD

UPDATE (Mar 4): These issues have been resolved; thanks for your patience.

We are experiencing some problems with our job scheduler; it is running but will not/very sporadically accept new jobs and will timeout when you try to access job information (for example, with checkjob). Running jobs will not be impacted.

Condor Enhancements and Job Submission Changes

Starting today, we have enabled the ability to submit and manage Condor jobs from the HPC Login nodes. This will provide several benefits:

Welcome, Terry Ward!

We are pleased to welcome a new member to our team, Terry Ward.

Terry has 43 years of experience working in systems administration, mostly in large data center environments. His experience encompasses a broad range of IT activities, including system architecture, network management, and software development.

Terry will join our support team, and will take over a number of critical systems management tasks, including HPC scheduler management and storage operations.