From 0dca473701c973d4f0c69fa932028d0794056fea Mon Sep 17 00:00:00 2001 From: AndreaGuarracino Date: Sun, 6 Apr 2025 12:05:29 -0500 Subject: Add maintenance documentation for Octopus/Tux systems --- topics/octopus/maintenance | 98 ------------------------------------------ topics/octopus/maintenance.gmi | 98 ++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 98 insertions(+), 98 deletions(-) delete mode 100644 topics/octopus/maintenance create mode 100644 topics/octopus/maintenance.gmi diff --git a/topics/octopus/maintenance b/topics/octopus/maintenance deleted file mode 100644 index 65ea52e..0000000 --- a/topics/octopus/maintenance +++ /dev/null @@ -1,98 +0,0 @@ -# Octopus/Tux maintenance - -## To remember - -`fdisk -l` to see disk models -`lsblk -nd` to see mounted disks - -## Status - -octopus02 -- Devices: 2 3.7T SSDs + 2 894.3G SSDs + 2 4.6T HDDs -- **Status: Slurm not OK, LizardFS not OK** -- Notes: - - `octopus02 mfsmount[31909]: can't resolve master hostname and/or portname (octopus01:9421)`, - - **I don't see 2 drives that are physically mounted** - -octopus03 -- Devices: 4 3.7T SSDs + 2 894.3G SSDs -- Status: Slurm OK, LizardFS OK -- Notes: **I don't see 2 drives that are physically mounted** - -octopus04 -- Devices: 4 7.3 T SSDs (Neil) + 1 4.6T HDD + 1 3.7T SSD + 2 894.3G SSDs -- Status: Slurm NO, LizardFS OK (we don't share the HDD) -- Notes: no - -octopus05 -- Devices: 1 7.3 T SSDs (Neil) + 5 3.7T SSDs + 2 894.3G SSDs -- Status: Slurm OK, LizardFS OK -- Notes: no - -octopus06 -- Devices: 1 7.3 T SSDs (Neil) + 1 4.6T HDD + 4 3.7T SSDs + 2 894.3G SSDs -- Status: Slurm OK, LizardFS OK (we don't share the HDD) -- Notes: no - -octopus07 -- Devices: 1 7.3 T SSDs (Neil) + 4 3.7T SSDs + 2 894.3G SSDs -- Status: Slurm OK, LizardFS OK -- Notes: **I don't see 1 device that is physically mounted** - -octopus08 -- Devices: 1 7.3 T SSDs (Neil) + 1 4.6T HDD + 4 3.7T SSDs + 2 894.3G SSDs -- Status: Slurm OK, LizardFS OK (we don't share the HDD) -- Notes: no - -octopus09 -- Devices: 1 7.3 T SSDs (Neil) + 1 4.6T HDD + 4 3.7T SSDs + 2 894.3G SSDs -- Status: Slurm OK, LizardFS OK (we don't share the HDD) -- Notes: no - -octopus10 -- Devices: 1 7.3 T SSDs (Neil) + 4 3.7T SSDs + 2 894.3G SSDs -- Status: Slurm OK, LizardFS OK (we don't share the HDD) -- Notes: **I don't see 1 device that is physically mounted** - -octopus11 -- Devices: 1 7.3 T SSDs (Neil) + 5 3.7T SSDs + 2 894.3G SSDs -- Status: Slurm OK, LizardFS OK -- Notes: on - -tux05 -- Devices: 1 3.6 NVMe + 1 1.5T NVMe + 1 894.3G NVMe -- Status: Slurm OK, LizardFS OK (we don't share anything) -- Notes: **I don't have a picture to confirm physically mounted devices** - -tux06 -- Devices: 2 3.6 T SSDs (1 from Neil) + 1 1.5T NVMe + 1 894.3G NVMe -- Status: Slurm OK, LizardFS (we don't share anything) -- Notes: - - **Last picture reports 1 7.3 T SSD (Neil) that is missing** - - **Disk /dev/sdc: 3.64 TiB (Samsung SSD 990: free and usable for lizardfs** - - **Disk /dev/sdd: 3.64 TiB (Samsung SSD 990): free and usable for lizardfs** - -tux07 -- Devices: 3 3.6 T SSDs + 1 1.5T NVMe (Neil) + 1 894.3G NVMe -- Status: Slurm OK, LizardFS -- Notes: - - **Disk /dev/sdb: 3.64 TiB (Samsung SSD 990): free and usable for lizardfs** - - **Disk /dev/sdd: 3.64 TiB (Samsung SSD 990): mounted at /mnt/sdb and shared on LIZARDFS: TO CHECK BECAUSE IT HAS NO PARTITIONS** - -tux08 -- Devices: 3 3.6 T SSDs + 1 1.5T NVMe (Neil) + 1 894.3G NVMe -- Status: Slurm OK, LizardFS -- Notes: no - -tux09 -- Devices: 1 3.6 T SSDs + 1 1.5T NVMe + 1 894.3G NVMe -- Status: Slurm OK, LizardFS -- Notes: **I don't see 1 device that is physically mounted** - -## Neil disks -- four 8TB SSDs on the right of octopus04 -- one 8TB SSD in the left slot of octopus05 -- six 8TB SSDs bottom-right slot of octopus06,07,08,09,10,11 -- one 4TB NVMe and one 8TB SSDs on tux06, NVME in the bottom-right of the group of 4 on the left, SSD on the bottom-left of the group of 4 on the right -- one 4TB NVMe on tux07, on the top-left of the group of 4 on the right -- one 4TB NVMe on tux08, on the top-left of the group of 4 on the right diff --git a/topics/octopus/maintenance.gmi b/topics/octopus/maintenance.gmi new file mode 100644 index 0000000..65ea52e --- /dev/null +++ b/topics/octopus/maintenance.gmi @@ -0,0 +1,98 @@ +# Octopus/Tux maintenance + +## To remember + +`fdisk -l` to see disk models +`lsblk -nd` to see mounted disks + +## Status + +octopus02 +- Devices: 2 3.7T SSDs + 2 894.3G SSDs + 2 4.6T HDDs +- **Status: Slurm not OK, LizardFS not OK** +- Notes: + - `octopus02 mfsmount[31909]: can't resolve master hostname and/or portname (octopus01:9421)`, + - **I don't see 2 drives that are physically mounted** + +octopus03 +- Devices: 4 3.7T SSDs + 2 894.3G SSDs +- Status: Slurm OK, LizardFS OK +- Notes: **I don't see 2 drives that are physically mounted** + +octopus04 +- Devices: 4 7.3 T SSDs (Neil) + 1 4.6T HDD + 1 3.7T SSD + 2 894.3G SSDs +- Status: Slurm NO, LizardFS OK (we don't share the HDD) +- Notes: no + +octopus05 +- Devices: 1 7.3 T SSDs (Neil) + 5 3.7T SSDs + 2 894.3G SSDs +- Status: Slurm OK, LizardFS OK +- Notes: no + +octopus06 +- Devices: 1 7.3 T SSDs (Neil) + 1 4.6T HDD + 4 3.7T SSDs + 2 894.3G SSDs +- Status: Slurm OK, LizardFS OK (we don't share the HDD) +- Notes: no + +octopus07 +- Devices: 1 7.3 T SSDs (Neil) + 4 3.7T SSDs + 2 894.3G SSDs +- Status: Slurm OK, LizardFS OK +- Notes: **I don't see 1 device that is physically mounted** + +octopus08 +- Devices: 1 7.3 T SSDs (Neil) + 1 4.6T HDD + 4 3.7T SSDs + 2 894.3G SSDs +- Status: Slurm OK, LizardFS OK (we don't share the HDD) +- Notes: no + +octopus09 +- Devices: 1 7.3 T SSDs (Neil) + 1 4.6T HDD + 4 3.7T SSDs + 2 894.3G SSDs +- Status: Slurm OK, LizardFS OK (we don't share the HDD) +- Notes: no + +octopus10 +- Devices: 1 7.3 T SSDs (Neil) + 4 3.7T SSDs + 2 894.3G SSDs +- Status: Slurm OK, LizardFS OK (we don't share the HDD) +- Notes: **I don't see 1 device that is physically mounted** + +octopus11 +- Devices: 1 7.3 T SSDs (Neil) + 5 3.7T SSDs + 2 894.3G SSDs +- Status: Slurm OK, LizardFS OK +- Notes: on + +tux05 +- Devices: 1 3.6 NVMe + 1 1.5T NVMe + 1 894.3G NVMe +- Status: Slurm OK, LizardFS OK (we don't share anything) +- Notes: **I don't have a picture to confirm physically mounted devices** + +tux06 +- Devices: 2 3.6 T SSDs (1 from Neil) + 1 1.5T NVMe + 1 894.3G NVMe +- Status: Slurm OK, LizardFS (we don't share anything) +- Notes: + - **Last picture reports 1 7.3 T SSD (Neil) that is missing** + - **Disk /dev/sdc: 3.64 TiB (Samsung SSD 990: free and usable for lizardfs** + - **Disk /dev/sdd: 3.64 TiB (Samsung SSD 990): free and usable for lizardfs** + +tux07 +- Devices: 3 3.6 T SSDs + 1 1.5T NVMe (Neil) + 1 894.3G NVMe +- Status: Slurm OK, LizardFS +- Notes: + - **Disk /dev/sdb: 3.64 TiB (Samsung SSD 990): free and usable for lizardfs** + - **Disk /dev/sdd: 3.64 TiB (Samsung SSD 990): mounted at /mnt/sdb and shared on LIZARDFS: TO CHECK BECAUSE IT HAS NO PARTITIONS** + +tux08 +- Devices: 3 3.6 T SSDs + 1 1.5T NVMe (Neil) + 1 894.3G NVMe +- Status: Slurm OK, LizardFS +- Notes: no + +tux09 +- Devices: 1 3.6 T SSDs + 1 1.5T NVMe + 1 894.3G NVMe +- Status: Slurm OK, LizardFS +- Notes: **I don't see 1 device that is physically mounted** + +## Neil disks +- four 8TB SSDs on the right of octopus04 +- one 8TB SSD in the left slot of octopus05 +- six 8TB SSDs bottom-right slot of octopus06,07,08,09,10,11 +- one 4TB NVMe and one 8TB SSDs on tux06, NVME in the bottom-right of the group of 4 on the left, SSD on the bottom-left of the group of 4 on the right +- one 4TB NVMe on tux07, on the top-left of the group of 4 on the right +- one 4TB NVMe on tux08, on the top-left of the group of 4 on the right -- cgit 1.4.1