Skip to main content

Website Category

GPFS briefly unmounted at 8:02am on Wednesday; check output of jobs running at the time

Jul. 24, 2019—At approximately 8:02 AM today a command was issued to remove a misconfigured compute node from the cluster in GPFS. Due to the nature of the configuration problem and the timing of the command, this caused most or all of the ACCRE gateway and compute nodes to unmount GPFS for approximately one minute. We are...

Read more


New Python guidelines: ACCRE-optimized Python 3 builds with virtual environments are strongly encouraged

Jul. 15, 2019—We have significantly revised our documentation on using Python on the ACCRE cluster to address some of the most common issues that are brought to our attention and improve the overall cluster user experience. Here is a brief overview of some changes: Although Anaconda provides an easy way to use a large selection of software,...

Read more


Will French steps down as ACCRE’s Director of Research Computing Operations

Jul. 2, 2019—Will French, ACCRE’s director of research computing operations, has left Vanderbilt after nearly five years of service to ACCRE. Will came to Vanderbilt in 2008 as a graduate student in chemical and biomolecular engineering; he joined us shortly after completing his PhD and eventually became co-director of ACCRE in 2017.

Read more


Existing ACCRE groups: Let us help you use ACCRE to your fullest potential.

Jun. 24, 2019—Following best practices in an HPC cluster to ensure job performance and reliability may not be as trivial as it seems, especially with massive workloads, complex pipelines or parallel jobs. ACCRE users and groups can take advantage of our consulting services to optimize their workflow in order to maximize the output of their computations. This...

Read more


GPFS data migration to run most weekends, starting tonight

Jun. 7, 2019—Beginning tonight we will kick off a data migration between different pools of storage in our GPFS23 filesystem (/scratch and /data). Hot data (files that have been accessed more recently) will be moved to faster, more expensive storage while cold data (files that have not been accessed recently) will be moved to more economical storage...

Read more


Scheduled maintenance on public gateways complete

Apr. 18, 2019—Original post: This Saturday morning, April 20th, we will be taking the public gateway and portal servers offline from 7 am to 9 am in order to reboot and upgrade the operating system to CentOS 7.6 from 7.4 and the GPFS filesystem to 5.0.2 from 4.2.3. Updating the entire cluster to GPFS 5 is an...

Read more


New command: slurm_groups

Mar. 11, 2019—We’ve added a new command to the ACCRE cluster, slurm_groups, that allows you to view your current SLURM group membership. This is especially useful if you have GPU access to the cluster, as it gives you the group and partition that you will need to use. Click the link to go to the documentation.  

Read more


[Resolved] /scratch and /data are back online following weekend maintenance

Jan. 24, 2019—Update, 2/12/2019: /scratch and /data are back online and we are now accepting new jobs. We were never able to get the maintenance command to run successfully, but we were able to verify (with IBM’s assistance) the integrity of /scratch and /data, which is great news and means we will not need to take another...

Read more


Final Steps for CentOS 7 Upgrade

Jan. 10, 2019—Update, Jan 25: The CentOS 6 login is now closed. Original post below… It has been a long journey, but we are almost to the end! Please see below for a schedule of the final systems to be upgraded to CentOS 7. Note this schedule does not include a handful of custom/private gateways that still...

Read more


Adding Disk Space to /scratch and /data

Jan. 2, 2019—Beginning at noon today we will be adding a few hundred TB of additional disk space to our gpfs23 filesystem (which includes /scratch and /data). It is possible that this may cause some sluggishness on the cluster for an hour or so. Tonight we will then kick off a migration to redistribute data to these...

Read more