Skip to content

Hardware

Warning

This page has not been updated to reflect latest cluster changes yet

Hardware Specification#

The use of the resources of the TalTech HPC Centre requires an active Uni-ID account, a procedure for non-employees/non-students can be found here (in Estonian).

TalTech operates the following infrastructure:

  • base.hpc.taltech.ee is the new cluster environment all nodes from HPC1 and HPC2 will be migrated here
    • uptime graph for our services can be found here
    • UniID login (please ask at hpcsupport@taltech.ee to activate access)
    • SLURM v20 scheduler
    • CentOS 7
    • home-directories' location: /gpfs/mariana/home/uni-ID
    • HPC1 home-directories accessible under /beegfs/home/uni-ID
    • home directory file system has 1.5 PB storage, with a 2 TB/user quota
    • green nodes (former hpc2.ttu.ee nodes)
      • 32 nodes 2 x Intel Xeon Gold 6148 20C 2.40 GHz, 96 GB DDR4-2666 R ECC RAM (green[1-32])
      • 25 Gbit Ethernet
      • 18 of these nodes are connected with low latency interconnect FDR InfiniBand (green-ib partition)
    • gray nodes (former hpc.ttu.ee nodes, migration in progress)
      • 48 nodes with 2 x Intel Xeon E5-2630L 6C with 64 GB RAM and 1 TB local drive connected with low latency interconnect QDR InfiniBand (gray-ib partition)
      • 60 nodes with 2 x Intel Xeon E5-2630L 6C with 48 GB RAM and 1 TB local drive (memory upgrade to 64 GB in progress)
      • 1 Gbit Ethernet
    • mem1tb large memory node
      • 1 node with 1TB of operating RAM
      • 4x Intel Xeon CPU E5-4640 (together 32 cores, 64 threads)
    • amp GPU node, specific guide for amp (amp.hpc.taltech.ee)
      • 1 node with 8xNvidia A100
      • 2x 64core AMD EPYC 7742 (together 128 cores, 256 threads)
      • 1 TB RAM
      • OpenACC, OpenMP, OpenCL, CUDA
      • Ubuntu 20.04
    • viz.hpc.taltech.ee Visualization node (accessible within University network and FortiVPN)
      • this service is intended mainly for post-processing, especially of large datasets; it is not intended as a remote desktop service; job submission is not possible from here
      • 1 node with 2 nVidia Tesla K20Xm grapic cards (on displays :0.0 and :0.1)
      • this node is not for computation on GPUs
      • Debian 10
      • UniID login with ssh-keys (no password login; previous login on base.hpc.taltech.ee required to create account)
      • same home-directories as base, HPC1-homes accessible under /beegfs/home/uni-ID
      • ParaView, VisIt and COVISE available for remote visualization
      • VirtualGL, Xpra, VNC available for remote execution of programs (can be used with e.g. Mayavi, ParaView, VisIt, COVISE, OpenDX)
    • There is no backup please take care of backups yourself!
    • partitions:
      • short: (default) time limit 4 hours, default time 10 min, default mem 1 GB/thread, green nodes
      • common: time limit 8 days, default time 10 min, default mem 1 GB/thread, green nodes
      • long: time limit 22 days, default time 10 min, default mem 1 GB/thread, green nodes
      • green-ib: time limit 8 days, default time 10 min, default mem 1 GB/thread, green InfiniBand nodes
      • gray-ib: time limit 8 days, default time 10 min, default mem 1 GB/thread, gray InfiniBand nodes
      • gpu: amp GPU node, time limit 5 days, default time 10 min, default mem 1 GB/thread
      • mem1tb: mem1tb node
  • training.hpc.taltech.ee virtual training cluster
    • used for testing
    • running on the TalTech ETAIS OpenStack cloud service
  • TalTech ETAIS Cloud: 4 node OpenStack cloud
    • 5 compute (nova) nodes with 282GB or 768GB of RAM and 80 threads each
    • 65 TB CephFS storage (net capacity)
    • accessible through the ETAIS website: https://etais.ee/using/