Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Altamira includes 158 main compute nodes, 4 additional GPU compute nodes and 2 login server.

Main compute nodes have two Intel Sandybridge E5-2670 processors, each one with 8 cores operating at 2.6 GHz and a cache of 20MB, 64 GB of RAM memory (i.e. 4 GB/core) and 500 GB local disk.

...

Main compute and login nodes run Centos 7.49

Network

The internal network in Altamira includes:

...

All the nodes are connected to a global storage system based on GPFS (Global Parallel File System) providing a total storage of 1.5 PB.

File Systems

Each nodes has several areas of disk space for installing system or programs and storing files. These areas may have size or time limits, please read carefully all this section to know about the policy of usage of each of these filesystems. There are 2 different types of storage available inside a node:

...

The job scheduling system running at Altamira is SLURM. Current version of Slurm in Altamira is 20.11.7.

...

Partition

Slurm has queues partitions to allocate jobs. In altamira, there are some queues configured. This table resume the main queues a group of nodes with same features are assigned to a partition.  This following table resumes the main partitions and the principal properties of each of them:. By default, compute is configured if --patition option is not defined in the job. Set it to res if you are coming from a RES project or use the testing partition for test jobs before submitting to the production partition (compute). Test jobs in the testing queue has a limit of 8 cpus and 3 hours of time as you can see below at qos.

PartitionDescription
computeMain partition in Altamira
resGroup of nodes for RES users
testingThis partition can be used to test new workflows and also to help new users to familiarise themselves with the SLURM batch system.  Both serial and parallel code can be tested on the testing partition.
         QoS

         Specify the Quality of Service (Qos) for each job submitted to slurm. Jobs request a QOS using the "--qos=" option to the sbatch,        salloc, and srun commands. By default,

  • the local user is assigned to the main and testing queues.
  • the users coming from a RES project belong to the res_a and testing queues.
GroupQueueDescriptionLimit number of nodesMax CPU cores per userMax Run TimeMax jobs per user
LocalmainMain queue for local users

158

5123 days1000
testingThis queue can be used to test new workflows and also to help new users to familiarise themselves with the SLURM batch system.  Both serial and parallel code can be tested on the testing queue.483 hours2
GroupQueueDescriptionPropertiesLocalmainMain queue for local users
  • Max Number of nodes: 158
  • Max Wall time:3 days
  • Max Number of Jobs Per User: 1000
  • Max Number of nodes: 64
  • Max Wall time:

    Res

    res_aMain queue for RES users

    128

    1024 (64 nodes)
    3 days
    Max Number of Jobs Per User:
    500
    res_cQueue assigned for res users when they reach the period hours limits requested
  • Max Number of nodes: 4
  • Max Wall time:
    464 (4 nodes)
    1 day
    Max Number of Jobs Per User:
    10


    Connecting to Altamira

    Once you have a username and its associated password you can get into Altamira system, connecting to one of the the 2 login nodes at login1.ifca.es in the case of the first login or login2.ifca.es for the second one (see Login node to know more information about the login nodes). The password provided is temporal, you must change this initial password after connecting to IPA. Also use a strong password (do not use a word or phrase from a dictionary and do not use a word that can be obviously tied to your person). 

    ...