Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

The account_no (or project) is important for batch executions. You need to indicate an account_no to be accounted for in the scheduler, using the flag "-A"

#SBATCH -A <account_no>

#PBS -A <account_no>

Please remember that different project projects are usually active of on different hosts. With the "saldo -b" command you can list all the accounaccount_no associated to your username. On systems like MARCONI, where different partitions are available on the same environment, you need to specify the host name in the "saldo" command:

...

> saldo -b --knl (reports projects defined on Marconi-KNL)

Budget Linearization policy

On MARCONI a linearization policy for the usage of project budgets has been defined and implemented. For each account, a monthly quota is defined as:

monthTotal = (total_budget / total_no_of_months)

Starting from the first day of each month, the collaborators of any account are allowed to use the quota at full priority. As long as the budget is consumed, the jobs submitted from the account will gradually lose priority, until the monthly budget (monthTotal) is fully consumed. At that moment, their jobs will still be considered for execution, but with a lower priority than the jobs from accounts that still have some monthly quota left.

This policy is similar to those already applied by other important HPC centers in Europe and worldwide. The goal is to improve the response time, giving users the opportunity of using the cpu hours assigned to their project in relation of their actual size (total amount of core-hours).

Disks and Filesystems

The storage organization conforms to the CINECA infrastructure (see Section Data Storage and Filesystems). 

...

$WORK area is defined for each active project on the system, reserved for all the collaborators of the project. This is a safe storage area to keep run time data for the whole life of the project.

 

 Total Dimension (TB)Quota (GB)Notes
$HOME20050
  • permanent/backuped, user specific, local
$CINECA_SCRATCH2.500no quota
  • temporary, user specific, local
  • automatic cleaning procedure of data older than 1 month
$WORK7.1001.000
  • permanent, project specific, local
  • extensions can be considered if motivated (mailto: superc@cineca.it)

 

$DRES points to the shared repository where Data RESources are maintained. This is a data archive area availble only on-request, shared with all CINECA HPC systems and among different projects. $DRES is not mounted on the compute nodes. This means that you cannot access it within a batch job: all data needed during the batch execution has to be moved on to $WORK or $CINECA_SCRATCH before the run starts.

...

For more information and examples of job scripts, see section Batch Scheduler PBSSLURM.

Submitting serial Batch jobs

The bdw_all_serial partition is available with one core and a maximum walltime of 4 hours. It runs on the login nodes and it is designed for pre/post-processing serial analysis, and for moving your data (via rsync, scp etc.) in case more than 10 minutes are required to complete the data transfer. In order to use this partition you have to specify the SLURM flag "-P":

#SBATCH -p bdw_all_serial

Submitting Batch jobs for A1 partition

...