From 09f083021181186f52047eaa5bc5ca5449a6c4e9 Mon Sep 17 00:00:00 2001 From: Pjotr Prins Date: Sat, 21 Nov 2020 09:56:40 +0000 Subject: Updated facilities --- general/help/facilities.md | 62 +++++++++++++++++++++++++--------------------- 1 file changed, 34 insertions(+), 28 deletions(-) (limited to 'general') diff --git a/general/help/facilities.md b/general/help/facilities.md index 89771f7..2647696 100644 --- a/general/help/facilities.md +++ b/general/help/facilities.md @@ -2,28 +2,33 @@ The core GeneNetwork team maintains modern Linux servers and storage systems for genetic, genomic, and phenome analyses. Machines are -located in the main UTHSC machine room of the Lamar Alexander -Building in at UTHSC (Memphis campus). The whole team has access to this space for -upgrades and hardware maintenance. Issues and work packages are +located in the main UTHSC machine room of the Lamar Alexander Building +at UTHSC (Memphis campus). The whole team has access to this space +for upgrades and hardware maintenance. Issues and work packages are tracked through a Trello board and we use git repositories for documentation (all available on request). This computing facility has four computer racks dedicated to GeneNetwork-related work. Each rack has a mix of Dell PowerEdge -servers (from a few low-end R610s, high performance Dell R6515, and two R7425 -64-core systems - tux01 and tux02 - running the GeneNetwork web -services). We also support several more experimental systems, including a 40-core system with 196 GB RAM and 2x NVIDIA V100 GPU (tux03), and one Penguin Computing Relion 2600GT systems (Penguin2) with a NVIDIA Tesla K40 -GPU used for software development and to serve outside-facing -less secure R/shiny and Python services that run in isolated -containers. Effectively, we have three outward facing servers -that are fully used by GeneNetwork teams with a -total of 64+64+40+28 = 196 real cores. In late 2020 we set up a small -HPC cluster (Octopus), consisting of 11 PowerEdge R6515 AMD EPYC -7402P 24-core CPUs (264 cores). Most of these machines -are equipped with 128 GB RAM, but two nodes have 1 TB of memory. -All Octopus nodes run Debian and GNU Guix and use Slurm for -batch submission. All racks have dedicated high-speed Cisco switches -and firewalls that are maintained by UTHSC IT staff. +servers (from a few low-end R610s, R6515, and two R7425 AMD Epyc +64-core 256GB RAM systems - tux01 and tux02 - running the GeneNetwork +web services). We also support several more experimental systems, +including a 40-core R7425 system with 196 GB RAM and 2x NVIDIA V100 +GPU (tux03), and one Penguin Computing Relion 2600GT systems +(Penguin2) with NVIDIA Tesla K80 GPU used for software development and +to serve outside-facing less secure R/shiny and Python services that +run in isolated containers. Effectively, we have three outward facing +servers that are fully used by the GeneNetwork team with a total of +64+64+40+28 = 196 real cores. Late 2020 we added a small HPC cluster +(Octopus), consisting of 11 PowerEdge R6515 AMD EPYC 7402P 24-core +CPUs (264 real cores). Nine of these machines are equipped with 128 GB +RAM and two nodes have 1 TB of memory. Octopus is designed for +Mouse/Rat pangenome work without HIPAA restrictions. All Octopus nodes +run Debian and GNU Guix and use Slurm for batch submission. We are +adding support for distributed network file storage and running the +common workflow language (CWL) and Docker containers. The racks have +dedicated high-speed Cisco switches and firewalls that are maintained +by UTHSC IT staff. We also run some 'specials' including an ARM-based NVIDIA Jetson and a RISC-V [PolarFire @@ -33,17 +38,18 @@ have also ordered two RISC-V computers. In addition to above hardware we have batch submission access to the -cluster computing resource at the ISAAX computing facility operated -by the UT Joint Institute for Computational Sciences in a secure setup -at the DOE Oak Ridge National Laboratory and on the UT Knoxville campus. We have a 10 Gbit connection -from the machine room at UTHSC to data transfer nodes at ISAAC. -ISAAC has been upgraded in the past year (see [ISAAC system +cluster computing resource at the ISAAX computing facility operated by +the UT Joint Institute for Computational Sciences in a secure setup at +the DOE Oak Ridge National Laboratory (ORNL) and on the UT Knoxville +campus. We have a 10 Gbit connection from the machine room at UTHSC to +data transfer nodes at ISAAC. ISAAC has been upgraded in the past +year (see [ISAAC system overview](http://www.nics.utk.edu/computing-resources/acf/acf-system-overview) and now has over 3 PB of high-performance Lustre DDN storage and contains over 8000 cores with some large RAM nodes and several GPU -nodes. Drs. Prins and other team members have used ISAAC systems to -analyze genomic and genetic data sets. Note that we can -not yet use ISAAC and storage facilities for public-facing web -services because of stringent security requirements. ISAAC -however, will be highly useful for "precomputing" genomics and genetics - results using standardized pipelines. +nodes. Drs. Prins, Chen, Ashbrook and other team members use ISAAC +systems to analyze genomic and genetic data sets. Note that we can not +use ISAAC and storage facilities for public-facing web services +because of stringent security requirements. ISAAC however, can be +highly useful for precomputed genomics and genetics results using +standardized pipelines. -- cgit v1.2.3