Hardware Specification
The use of the resources of the TalTech HPC Centre requires an active Uni-ID account, a procedure for non-employees/non-students can be found here (in Estonian).
TalTech operates the following infrastructure:
base.hpc.taltech.ee is the new cluster environment all nodes from HPC1 and HPC2 will be migrated here
a live diagram of the cluster load is here: load diagram
UniID login (please ask at hpcsupport@taltech.ee to activate access)
SLURM v20 scheduler
CentOS 7
home-directories’ location:
/gpfs/mariana/home/uni-ID
HPC1 home-directories accessible under
/beegfs/home/uni-ID
home directory file system has 1.5 PB storage, with a 2 TB/user quota
green nodes (former hpc2.ttu.ee nodes)
32 nodes 2 x Intel Xeon Gold 6148 20C 2.40 GHz, 96 GB DDR4-2666 R ECC RAM (green[1-32])
25 Gbit Ethernet
18 of these nodes are connected with low latency interconnect FDR InfiniBand (green-ib partition)
gray nodes (former hpc.ttu.ee nodes, migration in progress)
48 nodes with 2 x Intel Xeon E5-2630L 6C with 64 GB RAM and 1 TB local drive connected with low latency interconnect QDR InfiniBand (gray-ib partition)
60 nodes with 2 x Intel Xeon E5-2630L 6C with 48 GB RAM and 1 TB local drive (memory upgrade to 64 GB in progress)
1 Gbit Ethernet
mem1tb large memory node
1 node with 1TB of operating RAM
4x Intel Xeon CPU E5-4640 (together 32 cores, 64 threads)
amp GPU node, specific guide for amp (amp.hpc.taltech.ee)
1 node with 8xNvidia A100
2x 64core AMD EPYC 7742 (together 128 cores, 256 threads)
1 TB RAM
OpenACC, OpenMP, OpenCL, CUDA
Ubuntu 20.04
viz.hpc.taltech.ee Visualization node (accessible within University network and FortiVPN)
this service is intended mainly for post-processing, especially of large datasets; it is not intended as a remote desktop service; job submission is not possible from here
1 node with 2 nVidia Tesla K20Xm grapic cards (on displays :0.0 and :0.1)
this node is not for computation on GPUs
Debian 10
UniID login with ssh-keys (no password login; previous login on base.hpc.taltech.ee required to create account)
same home-directories as base, HPC1-homes accessible under /beegfs/home/uni-ID
ParaView, VisIt and COVISE available for remote visualization
VirtualGL, Xpra, VNC available for remote execution of programs (can be used with e.g. Mayavi, ParaView, VisIt, COVISE, OpenDX)
There is no backup please take care of backups yourself!
partitions:
short: (default) time limit 4 hours, default time 10 min, default mem 1 GB/thread, green nodes
common: time limit 8 days, default time 10 min, default mem 1 GB/thread, green nodes
long: time limit 22 days, default time 10 min, default mem 1 GB/thread, green nodes
green-ib: time limit 8 days, default time 10 min, default mem 1 GB/thread, green InfiniBand nodes
gray-ib: time limit 8 days, default time 10 min, default mem 1 GB/thread, gray InfiniBand nodes
gpu: amp GPU node, time limit 5 days, default time 10 min, default mem 1 GB/thread
mem1tb: mem1tb node
training.hpc.taltech.ee virtual training cluster
used for testing
running on the TalTech ETAIS OpenStack cloud service
TalTech ETAIS Cloud: 4 node OpenStack cloud
5 compute (nova) nodes with 282GB or 768GB of RAM and 80 threads each
65 TB CephFS storage (net capacity)
accessible through the ETAIS website: https://etais.ee/using/