<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="pt-BR">
	<id>https://wiki.if.ufrgs.br/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Leon</id>
	<title>Instituto de Física - UFRGS - Contribuições do usuário [pt-br]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.if.ufrgs.br/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Leon"/>
	<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php/Especial:Contribui%C3%A7%C3%B5es/Leon"/>
	<updated>2026-04-04T15:41:11Z</updated>
	<subtitle>Contribuições do usuário</subtitle>
	<generator>MediaWiki 1.39.4</generator>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2182</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2182"/>
		<updated>2025-01-09T17:27:19Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Criating and executing a Job */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication/discussion: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
The system of queues and scheduling of tasks is controlled by the [https://slurm.schedmd.com/ Slurm Workload Manager].&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 &lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The cluster is accessible using the  UFRGS virtual prived network ([https://www1.ufrgs.br/CatalogoServicos/servicos/servico?servico=3178 vpn]) through server lovelace.if.ufrgs.br. &lt;br /&gt;
&lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may configure winscp to enter the address lovelace.if.ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to use &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.&lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager! Just after the second # they are comments. To comment the whole line put ## at the beginning.)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2155</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2155"/>
		<updated>2024-04-04T13:52:01Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Rules for scheduling, access control, and usage of the research infrastructure */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication/discussion: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
The system of queues and scheduling of tasks is controlled by the [https://slurm.schedmd.com/ Slurm Workload Manager].&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 &lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The cluster is accessible using the  UFRGS virtual prived network ([https://www1.ufrgs.br/CatalogoServicos/servicos/servico?servico=3178 vpn]) through server lovelace.if.ufrgs.br. &lt;br /&gt;
&lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may configure winscp to enter the address lovelace.if.ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to use &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.&lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2154</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2154"/>
		<updated>2024-04-04T13:50:30Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Management Software */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication/discussion: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
The system of queues and scheduling of tasks is controlled by the [https://slurm.schedmd.com/ Slurm Workload Manager].&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 &lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The cluster is accessible using the  UFRGS virtual prived network ([https://www1.ufrgs.br/CatalogoServicos/servicos/servico?servico=3178 vpn]) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may configure winscp to enter the address lovelace.if.ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to use &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.&lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2153</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2153"/>
		<updated>2024-04-04T13:47:57Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Management Software */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication/discussion: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
[https://slurm.schedmd.com/ Slurm Workload Manager]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 &lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The cluster is accessible using the  UFRGS virtual prived network ([https://www1.ufrgs.br/CatalogoServicos/servicos/servico?servico=3178 vpn]) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may configure winscp to enter the address lovelace.if.ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to use &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.&lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2152</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2152"/>
		<updated>2024-04-04T13:46:09Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Management Software */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication/discussion: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[https://slurm.schedmd.com/ Slurm Workload Manager]&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 &lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The cluster is accessible using the  UFRGS virtual prived network ([https://www1.ufrgs.br/CatalogoServicos/servicos/servico?servico=3178 vpn]) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may configure winscp to enter the address lovelace.if.ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to use &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.&lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2151</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2151"/>
		<updated>2024-04-04T13:44:56Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Users Committee */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication/discussion: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 &lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The cluster is accessible using the  UFRGS virtual prived network ([https://www1.ufrgs.br/CatalogoServicos/servicos/servico?servico=3178 vpn]) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may configure winscp to enter the address lovelace.if.ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to use &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.&lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2150</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2150"/>
		<updated>2024-04-04T13:10:29Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Online scheduling */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 &lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The cluster is accessible using the  UFRGS virtual prived network ([https://www1.ufrgs.br/CatalogoServicos/servicos/servico?servico=3178 vpn]) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may configure winscp to enter the address lovelace.if.ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to use &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.&lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2149</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2149"/>
		<updated>2024-04-04T13:07:42Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Rules for scheduling, access control, and usage of the research infrastructure */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 &lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network ([https://www1.ufrgs.br/CatalogoServicos/servicos/servico?servico=3178 vpn]) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to use &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.&lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2148</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2148"/>
		<updated>2024-04-04T12:50:54Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Using softwares in the cluster */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 &lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to use &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.&lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2147</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2147"/>
		<updated>2024-04-04T12:49:58Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Installed Software */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 &lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2146</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2146"/>
		<updated>2024-04-04T12:49:25Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Hardware in ada nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2145</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2145"/>
		<updated>2024-04-03T15:10:21Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Rules for scheduling, access control, and usage of the research infrastructure existing within the institution */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2144</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2144"/>
		<updated>2024-04-03T15:07:12Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Conect to  cluster-slurm */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure existing within the institution ==&lt;br /&gt;
&lt;br /&gt;
=== Online scheduling ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2143</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2143"/>
		<updated>2024-04-03T15:06:01Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* How to use */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rules for scheduling, access control, and usage of the research infrastructure existing within the institution ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2142</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2142"/>
		<updated>2024-04-03T14:54:54Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Users Committee */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Users have two channels for communication: &lt;br /&gt;
&lt;br /&gt;
1) The fis-linux-if@grupos.ufrgs.br mailing list&lt;br /&gt;
&lt;br /&gt;
2) Direct messages to the IT department via the email fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2141</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2141"/>
		<updated>2024-04-03T14:48:07Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Management Committee */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2140</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2140"/>
		<updated>2024-04-03T14:46:26Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Management Committee */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the Physics Institute.&lt;br /&gt;
&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2139</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2139"/>
		<updated>2024-04-03T14:45:27Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Management Committee */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
The cluster is managed by professors representing the fields of Astronomy, Theoretical Physics, and Experimental Physics, in addition to an IT department employee from the IF.&lt;br /&gt;
Astronomy: Rogério Riffel&lt;br /&gt;
Theoretical Physics: Leonardo Brunnet&lt;br /&gt;
Experimental Physics: Pedro Grande&lt;br /&gt;
TI employee: Gustavo Feller&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2138</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2138"/>
		<updated>2024-04-03T14:39:25Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Hardware in lovelace nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: Three nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2137</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2137"/>
		<updated>2024-04-03T14:38:45Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Clusters Ada and Lovelace - Instituto de Física UFRGS */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The cluster is located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: hree nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2136</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2136"/>
		<updated>2024-04-03T14:38:12Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Installed Software */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: hree nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2135</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2135"/>
		<updated>2024-04-03T14:37:39Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Conect to  cluster-slurm */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: hree nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in ada)&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server lovelace.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2134</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2134"/>
		<updated>2024-04-03T14:37:02Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Conect to  cluster-slurm */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: hree nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in ada)&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible using the  UFRGS virtual prived network (vpn) through server cluster-slurm.if.ufrgs.br. &lt;br /&gt;
To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2133</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2133"/>
		<updated>2024-04-03T14:33:28Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Clusters Ada and Lovelace - Instituto de Física UFRGS */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Management Committee ==&lt;br /&gt;
&lt;br /&gt;
== Users Committee ==&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: hree nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in ada)&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2124</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2124"/>
		<updated>2024-01-25T12:22:51Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Management Software */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&lt;br /&gt;
Account request: mail to fisica-ti@ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: hree nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in ada)&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2123</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2123"/>
		<updated>2024-01-25T12:20:27Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Installed Software */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: hree nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in ada)&lt;br /&gt;
OS: Debian 12 (in lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2122</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2122"/>
		<updated>2024-01-25T12:19:46Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Hardware in lovelace nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores) + AMD 16 cores&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: hree nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2121</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2121"/>
		<updated>2024-01-25T12:17:39Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Hardware in the lovelace nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2120</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2120"/>
		<updated>2024-01-25T12:17:00Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Software in the nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installed Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2119</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2119"/>
		<updated>2024-01-25T12:16:30Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Hardware in the  ada nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2118</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2118"/>
		<updated>2024-01-24T20:47:46Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Management Software */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&lt;br /&gt;
Number of users on 1/24/2023: 150&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2117</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2117"/>
		<updated>2024-01-24T20:45:54Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Software in the nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
gcc&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2116</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2116"/>
		<updated>2024-01-24T20:45:02Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Conect to  cluster-slurm */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
GCC&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@lovelace.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br&lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2115</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2115"/>
		<updated>2024-01-24T20:44:10Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Software in the nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
GCC&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
compucel3d&lt;br /&gt;
espresso&lt;br /&gt;
gromacs&lt;br /&gt;
lammps&lt;br /&gt;
mesa&lt;br /&gt;
openmpi&lt;br /&gt;
povray&lt;br /&gt;
quantum-espresso&lt;br /&gt;
vasp&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br &lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2114</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2114"/>
		<updated>2024-01-24T20:40:49Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Management Software */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager(https://slurm.schedmd.com/)&lt;br /&gt;
&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
GCC&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br &lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2113</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2113"/>
		<updated>2024-01-24T20:39:55Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Management Software */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
Number of jobs per user controlled on demand.&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
GCC&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br &lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2112</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2112"/>
		<updated>2024-01-24T20:34:38Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Software in the nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 12 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
GCC&lt;br /&gt;
gfortran&lt;br /&gt;
python: torch, numba&lt;br /&gt;
julia&lt;br /&gt;
conda&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br &lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2111</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2111"/>
		<updated>2024-01-24T20:32:02Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Hardware in the  ada nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes with NVIDIA CUDA&lt;br /&gt;
Storage: storage Asustor 12TB&lt;br /&gt;
Inter-node connection: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 11 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
 GCC&lt;br /&gt;
 gfortran&lt;br /&gt;
 python2&lt;br /&gt;
 python3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br &lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2110</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2110"/>
		<updated>2024-01-24T20:30:06Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Hardware in the lovelace nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage with 50GB  quota per user &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage Dell 12TB &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 11 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
 GCC&lt;br /&gt;
 gfortran&lt;br /&gt;
 python2&lt;br /&gt;
 python3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br &lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2109</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2109"/>
		<updated>2024-01-24T20:28:46Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Hardware in the lovelace nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage with 50GB  quota per user &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: 64 GB each&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage with 50GB  quota per user &lt;br /&gt;
Conection inter-nodes: Gigabit&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 11 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
 GCC&lt;br /&gt;
 gfortran&lt;br /&gt;
 python2&lt;br /&gt;
 python3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br &lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2108</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2108"/>
		<updated>2024-01-24T20:14:35Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Hardware in the  ada nodes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: 16 nodes x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: 3 nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage with 50GB  quota per user &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: varia entre 64 GB&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage with 50GB  quota per user &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 11 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
 GCC&lt;br /&gt;
 gfortran&lt;br /&gt;
 python2&lt;br /&gt;
 python3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br &lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=P%C3%A1gina_principal&amp;diff=2086</id>
		<title>Página principal</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=P%C3%A1gina_principal&amp;diff=2086"/>
		<updated>2022-12-22T17:45:52Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* [//www.if.ufrgs.br/if/informatica/cluster/Cluster] */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''&amp;lt;center&amp;gt;&amp;lt;big&amp;gt;&amp;lt;big&amp;gt;Universidade Federal do Rio Grande do Sul&amp;lt;/big&amp;gt;&amp;lt;/big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''&amp;lt;big&amp;gt;&amp;lt;big&amp;gt;Instituto de Física&amp;lt;/big&amp;gt;&amp;lt;/big&amp;gt;'''&lt;br /&gt;
&amp;lt;/center&amp;gt;'''''&lt;br /&gt;
&lt;br /&gt;
__NOTOC__ &lt;br /&gt;
&lt;br /&gt;
== [[PeIF| Planejamento Estratégico IF]] ==&lt;br /&gt;
== [[Física Experimental|Física Experimental IF/UFRGS]] ==&lt;br /&gt;
== [[Laboratório de Implantação Iônica]] ==&lt;br /&gt;
== [[Observatório Astronômico do Campus do Vale]] ==&lt;br /&gt;
== [//www.if.ufrgs.br/if/comissao-de-extensao/ Comissão de Extensão IF/UFRGS] ==&lt;br /&gt;
== [//www.if.ufrgs.br/if/informatica/cluster/ Cluster] ==&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=P%C3%A1gina_principal&amp;diff=2085</id>
		<title>Página principal</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=P%C3%A1gina_principal&amp;diff=2085"/>
		<updated>2022-12-22T17:45:02Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Cluster */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''&amp;lt;center&amp;gt;&amp;lt;big&amp;gt;&amp;lt;big&amp;gt;Universidade Federal do Rio Grande do Sul&amp;lt;/big&amp;gt;&amp;lt;/big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''&amp;lt;big&amp;gt;&amp;lt;big&amp;gt;Instituto de Física&amp;lt;/big&amp;gt;&amp;lt;/big&amp;gt;'''&lt;br /&gt;
&amp;lt;/center&amp;gt;'''''&lt;br /&gt;
&lt;br /&gt;
__NOTOC__ &lt;br /&gt;
&lt;br /&gt;
== [[PeIF| Planejamento Estratégico IF]] ==&lt;br /&gt;
== [[Física Experimental|Física Experimental IF/UFRGS]] ==&lt;br /&gt;
== [[Laboratório de Implantação Iônica]] ==&lt;br /&gt;
== [[Observatório Astronômico do Campus do Vale]] ==&lt;br /&gt;
== [//www.if.ufrgs.br/if/comissao-de-extensao/ Comissão de Extensão IF/UFRGS] ==&lt;br /&gt;
== [//www.if.ufrgs.br/if/informatica/cluster/Cluster] ==&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2019</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2019"/>
		<updated>2022-03-14T18:51:47Z</updated>

		<summary type="html">&lt;p&gt;Leon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: some nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage with 50GB  quota per user &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: varia entre 64 GB&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage with 50GB  quota per user &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 11 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
 GCC&lt;br /&gt;
 gfortran&lt;br /&gt;
 python2&lt;br /&gt;
 python3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br &lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OR&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2018</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2018"/>
		<updated>2022-03-14T18:48:46Z</updated>

		<summary type="html">&lt;p&gt;Leon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Clusters Ada and Lovelace - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The clusters are located at Instituto de Física da UFRGS, in Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestruture ==&lt;br /&gt;
&lt;br /&gt;
=== Management Software ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the  ada nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: x86_64&lt;br /&gt;
RAM: varies between 8 GB - 16 GB&lt;br /&gt;
GPU: some nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage with 50GB  quota per user &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware in the lovelace nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: Ryzen (32 and 2*24 cores)&lt;br /&gt;
RAM: varia entre 64 GB&lt;br /&gt;
GPU: two nodes have NVIDIA CUDA&lt;br /&gt;
Storage: storage with 50GB  quota per user &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software in the nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (in cluster ada)&lt;br /&gt;
OS: Debian 11 (in cluster lovelace)&lt;br /&gt;
Basic packages installed:&lt;br /&gt;
 GCC&lt;br /&gt;
 gfortran&lt;br /&gt;
 python2&lt;br /&gt;
 python3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== How to use ==&lt;br /&gt;
&lt;br /&gt;
=== Conect to  cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
The clusters are accessible through server cluster-slurm.if.ufrgs.br (ou ada.if.ufrgr.br). To access through a unix-like system use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Under windows you may use winscp.&lt;br /&gt;
&lt;br /&gt;
If you are not registered, ask for registration sending an email to fisica-ti@ufrgs.br &lt;br /&gt;
&lt;br /&gt;
=== Using softwares in the cluster ===&lt;br /&gt;
&lt;br /&gt;
To execute a software in a cluster job this program must:&lt;br /&gt;
&lt;br /&gt;
1. Be already installed&lt;br /&gt;
 &lt;br /&gt;
OU&lt;br /&gt;
&lt;br /&gt;
2. Be copied to the user home &lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp my_programm &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you are compiling your program in the cluster, one option is to user &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh &amp;lt;user&amp;gt;@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This will generate file &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, which is the executable.&lt;br /&gt;
&lt;br /&gt;
Being accessible by methods 1 or 2, the program can be executed in the cluster through one &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: If you execute your executable without submitting as &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, it will be executed in the server, not in the nodes. This is not recommended since the server computational capabilities are limited and you will be slowing down the server for everyone else.  &lt;br /&gt;
&lt;br /&gt;
=== Criating and executing a Job ===&lt;br /&gt;
&lt;br /&gt;
Slurm manages jobs and each job represents a program or task being executed.&lt;br /&gt;
&lt;br /&gt;
To submit a new job, you must create a script file describing the requisites and characteristics of the Job.&lt;br /&gt;
&lt;br /&gt;
A typical example of the content of a submission script is below&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Number of cpus to be allocated (Despite the # these SBATCH lines are compiled by the slurm manager!)&lt;br /&gt;
#SBATCH -N 1 # Nummber of nodes to be allocated  (You don't have to use all requisites, comment with ##)&lt;br /&gt;
#SBATCH -t 0-00:05 # Limit execution time (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Partition to be submitted&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Your program execution commands&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In option --qos, use the partition name with &amp;quot;qos_&amp;quot; prefix:&lt;br /&gt;
&lt;br /&gt;
partition: short -&amp;gt; qos: qos_short -&amp;gt; limit  2 weeks&lt;br /&gt;
&lt;br /&gt;
partition: long -&amp;gt; qos: qos_long -&amp;gt; limit de 3 month&lt;br /&gt;
  &lt;br /&gt;
If you run on GPU, specify the &amp;quot;generic resource&amp;quot; gpu in cluster ada:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 &lt;br /&gt;
#SBATCH -N 1&lt;br /&gt;
#SBATCH -t 0-00:05 &lt;br /&gt;
#SBATCH -p long &lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To ask for a specific gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To submit the job, execute:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Usefull commands ==&lt;br /&gt;
* To list jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* To list all jobs running in the cluster now:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* To delete a running job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* To list available partitions:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* To list gpu's in the nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* To list characteristic of all nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2015</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2015"/>
		<updated>2021-03-10T21:12:23Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Como utilizar */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Ada - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
O Cluster está localizado no Instituto de Física da UFRGS, em Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestrutura ==&lt;br /&gt;
&lt;br /&gt;
=== Software de gerenciamento ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware dos nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: x86_64&lt;br /&gt;
RAM: varia entre 4 GB - 8 GB&lt;br /&gt;
GPU: alguns nodes possuem NVIDIA CUDA&lt;br /&gt;
Storage: storage em rede com quota de 50 GB por usuário, os nodes não possuem HD local &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software nos nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (Jessie) x86_64&lt;br /&gt;
Pacotes instalados:&lt;br /&gt;
 gcc&lt;br /&gt;
 python2&lt;br /&gt;
 python3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Como utilizar ==&lt;br /&gt;
&lt;br /&gt;
=== Conectar-se ao cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
O cluster é acessível através do servidor cluster-slurm (ou ada). Para acessar o server via SSH, use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh usuario@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
ou&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh usuario@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Caso você não tenha cadastro ou não é vinculado ao Instituto de Física, solicite o cadastro enviando um email para fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
=== Utilizando softwares no Cluster ===&lt;br /&gt;
&lt;br /&gt;
Para que seja possível executar um programa em um job no cluster, o programa deve:&lt;br /&gt;
&lt;br /&gt;
1. Já estar instalado&lt;br /&gt;
&lt;br /&gt;
OU&lt;br /&gt;
&lt;br /&gt;
2. Ser copiado para sua home (pasta do seu usuário)&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp meu_executavel usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Caso queira compilar o programa para uso no Cluster, uma das opções é usar o &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Isso irá gerar um arquivo &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, que é o executável.&lt;br /&gt;
&lt;br /&gt;
Estando acessível pelo método 1 ou 2, o programa pode ser executado no Cluster através de um &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: Caso você execute o programa sem submetê-lo como &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, ele não será executado nos nodes, e sim apenas no próprio server (cluster-slurm), que possui capacidades bem limitadas de processamento.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Criando e executando um Job ===&lt;br /&gt;
&lt;br /&gt;
O Slurm gerencia jobs, e cada job representa um programa ou tarefa sendo executado.&lt;br /&gt;
&lt;br /&gt;
Para submeter um novo Job, deve-se criar um arquivo de script descrevendo os requisitos e características de execução do Job.&lt;br /&gt;
&lt;br /&gt;
Formato do arquivo abaixo.&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Numero de CPU cores a serem alocados &lt;br /&gt;
#SBATCH -N 1 # Numero de nodes a serem alocados&lt;br /&gt;
#SBATCH -t 0-00:05 # Tempo limite de execucao (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Particao (fila) a ser submetido&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Na opção --qos, deve-se colocar o nome da partição com o prefixo &amp;quot;qos_&amp;quot;:&lt;br /&gt;
&lt;br /&gt;
partição: short -&amp;gt; qos: qos_short -&amp;gt; limite de 2 semanas&lt;br /&gt;
&lt;br /&gt;
partição: long -&amp;gt; qos: qos_long -&amp;gt; limite de 3 meses&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
Caso deseje rodar em GPU, é necessário especificar a fila e pedir explicitamente a ''generic resource'' gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Numero de CPU cores a serem alocados &lt;br /&gt;
#SBATCH -N 1 # Numero de nodes a serem alocados&lt;br /&gt;
#SBATCH -t 0-00:05 # Tempo limite de execucao (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Particao (fila) a ser submetido&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Para pedir alguma GPU específica, use um constraint adicionando a linha:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Para submeter o job, execute o comando &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Comandos úteis ==&lt;br /&gt;
* Para listar os seus jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* Para listar todos os jobs rodando no cluster neste momento:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* Para deletar um job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* Para listar as partições disponíveis:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* Para listar as gpus presentes nos nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* Para listar um resumo de todos os nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2014</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=2014"/>
		<updated>2021-03-10T21:11:32Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Conectar-se ao cluster-slurm */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Ada - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
O Cluster está localizado no Instituto de Física da UFRGS, em Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestrutura ==&lt;br /&gt;
&lt;br /&gt;
=== Software de gerenciamento ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware dos nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: x86_64&lt;br /&gt;
RAM: varia entre 4 GB - 8 GB&lt;br /&gt;
GPU: alguns nodes possuem NVIDIA CUDA&lt;br /&gt;
Storage: storage em rede com quota de 50 GB por usuário, os nodes não possuem HD local &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software nos nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (Jessie) x86_64&lt;br /&gt;
Pacotes instalados:&lt;br /&gt;
 gcc&lt;br /&gt;
 python2&lt;br /&gt;
 python3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Como utilizar ==&lt;br /&gt;
&lt;br /&gt;
=== Conectar-se ao cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
O cluster é acessível através do server cluster-slurm. Para acessar o server via SSH, use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh usuario@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
ou&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh usuario@ada.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Caso você não tenha cadastro ou não é vinculado ao Instituto de Física, solicite o cadastro enviando um email para fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
=== Utilizando softwares no Cluster ===&lt;br /&gt;
&lt;br /&gt;
Para que seja possível executar um programa em um job no cluster, o programa deve:&lt;br /&gt;
&lt;br /&gt;
1. Já estar instalado&lt;br /&gt;
&lt;br /&gt;
OU&lt;br /&gt;
&lt;br /&gt;
2. Ser copiado para sua home (pasta do seu usuário)&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp meu_executavel usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Caso queira compilar o programa para uso no Cluster, uma das opções é usar o &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Isso irá gerar um arquivo &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, que é o executável.&lt;br /&gt;
&lt;br /&gt;
Estando acessível pelo método 1 ou 2, o programa pode ser executado no Cluster através de um &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: Caso você execute o programa sem submetê-lo como &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, ele não será executado nos nodes, e sim apenas no próprio server (cluster-slurm), que possui capacidades bem limitadas de processamento.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Criando e executando um Job ===&lt;br /&gt;
&lt;br /&gt;
O Slurm gerencia jobs, e cada job representa um programa ou tarefa sendo executado.&lt;br /&gt;
&lt;br /&gt;
Para submeter um novo Job, deve-se criar um arquivo de script descrevendo os requisitos e características de execução do Job.&lt;br /&gt;
&lt;br /&gt;
Formato do arquivo abaixo.&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Numero de CPU cores a serem alocados &lt;br /&gt;
#SBATCH -N 1 # Numero de nodes a serem alocados&lt;br /&gt;
#SBATCH -t 0-00:05 # Tempo limite de execucao (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Particao (fila) a ser submetido&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Na opção --qos, deve-se colocar o nome da partição com o prefixo &amp;quot;qos_&amp;quot;:&lt;br /&gt;
&lt;br /&gt;
partição: short -&amp;gt; qos: qos_short -&amp;gt; limite de 2 semanas&lt;br /&gt;
&lt;br /&gt;
partição: long -&amp;gt; qos: qos_long -&amp;gt; limite de 3 meses&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
Caso deseje rodar em GPU, é necessário especificar a fila e pedir explicitamente a ''generic resource'' gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Numero de CPU cores a serem alocados &lt;br /&gt;
#SBATCH -N 1 # Numero de nodes a serem alocados&lt;br /&gt;
#SBATCH -t 0-00:05 # Tempo limite de execucao (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Particao (fila) a ser submetido&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Para pedir alguma GPU específica, use um constraint adicionando a linha:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Para submeter o job, execute o comando &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Comandos úteis ==&lt;br /&gt;
* Para listar os seus jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* Para listar todos os jobs rodando no cluster neste momento:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* Para deletar um job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* Para listar as partições disponíveis:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* Para listar as gpus presentes nos nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* Para listar um resumo de todos os nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=1955</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=1955"/>
		<updated>2019-03-19T17:03:04Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Comandos úteis */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Ada - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
O Cluster está localizado no Instituto de Física da UFRGS, em Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestrutura ==&lt;br /&gt;
&lt;br /&gt;
=== Software de gerenciamento ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware dos nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: x86_64&lt;br /&gt;
RAM: varia entre 4 GB - 8 GB&lt;br /&gt;
GPU: alguns nodes possuem NVIDIA CUDA&lt;br /&gt;
Storage: storage em rede com quota de 50 GB por usuário, os nodes não possuem HD local &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software nos nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (Jessie) x86_64&lt;br /&gt;
Pacotes instalados:&lt;br /&gt;
 gcc&lt;br /&gt;
 python2&lt;br /&gt;
 python3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Como utilizar ==&lt;br /&gt;
&lt;br /&gt;
=== Conectar-se ao cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
O cluster é acessível através do server cluster-slurm. Para acessar o server via SSH, use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh usuario@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Caso você não tenha cadastro ou não é vinculado ao Instituto de Física, solicite o cadastro enviando um email para fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
=== Utilizando softwares no Cluster ===&lt;br /&gt;
&lt;br /&gt;
Para que seja possível executar um programa em um job no cluster, o programa deve:&lt;br /&gt;
&lt;br /&gt;
1. Já estar instalado&lt;br /&gt;
&lt;br /&gt;
OU&lt;br /&gt;
&lt;br /&gt;
2. Ser copiado para sua home (pasta do seu usuário)&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp meu_executavel usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Caso queira compilar o programa para uso no Cluster, uma das opções é usar o &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Isso irá gerar um arquivo &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, que é o executável.&lt;br /&gt;
&lt;br /&gt;
Estando acessível pelo método 1 ou 2, o programa pode ser executado no Cluster através de um &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: Caso você execute o programa sem submetê-lo como &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, ele não será executado nos nodes, e sim apenas no próprio server (cluster-slurm), que possui capacidades bem limitadas de processamento.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Criando e executando um Job ===&lt;br /&gt;
&lt;br /&gt;
O Slurm gerencia jobs, e cada job representa um programa ou tarefa sendo executado.&lt;br /&gt;
&lt;br /&gt;
Para submeter um novo Job, deve-se criar um arquivo de script descrevendo os requisitos e características de execução do Job.&lt;br /&gt;
&lt;br /&gt;
Formato do arquivo abaixo.&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Numero de CPU cores a serem alocados &lt;br /&gt;
#SBATCH -N 1 # Numero de nodes a serem alocados&lt;br /&gt;
#SBATCH -t 0-00:05 # Tempo limite de execucao (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Particao (fila) a ser submetido&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Na opção --qos, deve-se colocar o nome da partição com o prefixo &amp;quot;qos_&amp;quot;:&lt;br /&gt;
&lt;br /&gt;
partição: short -&amp;gt; qos: qos_short -&amp;gt; limite de 2 semanas&lt;br /&gt;
&lt;br /&gt;
partição: long -&amp;gt; qos: qos_long -&amp;gt; limite de 3 meses&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
Caso deseje rodar em GPU, é necessário especificar a fila e pedir explicitamente a ''generic resource'' gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Numero de CPU cores a serem alocados &lt;br /&gt;
#SBATCH -N 1 # Numero de nodes a serem alocados&lt;br /&gt;
#SBATCH -t 0-00:05 # Tempo limite de execucao (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Particao (fila) a ser submetido&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Para pedir alguma GPU específica, use um constraint adicionando a linha:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Para submeter o job, execute o comando &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Comandos úteis ==&lt;br /&gt;
* Para listar os seus jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* Para listar todos os jobs rodando no cluster neste momento:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* Para deletar um job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* Para listar as partições disponíveis:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* Para listar as gpus presentes nos nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* Para listar um resumo de todos os nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=1954</id>
		<title>Cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Cluster&amp;diff=1954"/>
		<updated>2019-03-19T17:02:25Z</updated>

		<summary type="html">&lt;p&gt;Leon: /* Comandos úteis */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Cluster Ada - Instituto de Física UFRGS =&lt;br /&gt;
&lt;br /&gt;
O Cluster está localizado no Instituto de Física da UFRGS, em Porto Alegre.&lt;br /&gt;
&lt;br /&gt;
== Infraestrutura ==&lt;br /&gt;
&lt;br /&gt;
=== Software de gerenciamento ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Slurm Workload Manager&lt;br /&gt;
&lt;br /&gt;
Site :https://slurm.schedmd.com/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Hardware dos nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
CPU: x86_64&lt;br /&gt;
RAM: varia entre 4 GB - 8 GB&lt;br /&gt;
GPU: alguns nodes possuem NVIDIA CUDA&lt;br /&gt;
Storage: storage em rede com quota de 50 GB por usuário, os nodes não possuem HD local &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Software nos nodes ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
OS: Debian 8 (Jessie) x86_64&lt;br /&gt;
Pacotes instalados:&lt;br /&gt;
 gcc&lt;br /&gt;
 python2&lt;br /&gt;
 python3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Como utilizar ==&lt;br /&gt;
&lt;br /&gt;
=== Conectar-se ao cluster-slurm ===&lt;br /&gt;
&lt;br /&gt;
O cluster é acessível através do server cluster-slurm. Para acessar o server via SSH, use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ssh usuario@cluster-slurm.if.ufrgs.br&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Caso você não tenha cadastro ou não é vinculado ao Instituto de Física, solicite o cadastro enviando um email para fisica-ti@ufrgs.br.&lt;br /&gt;
&lt;br /&gt;
=== Utilizando softwares no Cluster ===&lt;br /&gt;
&lt;br /&gt;
Para que seja possível executar um programa em um job no cluster, o programa deve:&lt;br /&gt;
&lt;br /&gt;
1. Já estar instalado&lt;br /&gt;
&lt;br /&gt;
OU&lt;br /&gt;
&lt;br /&gt;
2. Ser copiado para sua home (pasta do seu usuário)&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp meu_executavel usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Caso queira compilar o programa para uso no Cluster, uma das opções é usar o &amp;lt;code&amp;gt;gcc&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Ex:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
scp -r source-code/ usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
ssh usuario@cluster-slurm.if.ufrgs.br:~/&lt;br /&gt;
cd source-code&lt;br /&gt;
gcc main.c funcoes.c&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Isso irá gerar um arquivo &amp;lt;code&amp;gt;a.out&amp;lt;/code&amp;gt;, que é o executável.&lt;br /&gt;
&lt;br /&gt;
Estando acessível pelo método 1 ou 2, o programa pode ser executado no Cluster através de um &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
OBS: Caso você execute o programa sem submetê-lo como &amp;lt;strong&amp;gt;JOB&amp;lt;/strong&amp;gt;, ele não será executado nos nodes, e sim apenas no próprio server (cluster-slurm), que possui capacidades bem limitadas de processamento.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Criando e executando um Job ===&lt;br /&gt;
&lt;br /&gt;
O Slurm gerencia jobs, e cada job representa um programa ou tarefa sendo executado.&lt;br /&gt;
&lt;br /&gt;
Para submeter um novo Job, deve-se criar um arquivo de script descrevendo os requisitos e características de execução do Job.&lt;br /&gt;
&lt;br /&gt;
Formato do arquivo abaixo.&lt;br /&gt;
&lt;br /&gt;
Ex: &amp;lt;code&amp;gt;job.sh&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Numero de CPU cores a serem alocados &lt;br /&gt;
#SBATCH -N 1 # Numero de nodes a serem alocados&lt;br /&gt;
#SBATCH -t 0-00:05 # Tempo limite de execucao (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Particao (fila) a ser submetido&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Na opção --qos, deve-se colocar o nome da partição com o prefixo &amp;quot;qos_&amp;quot;:&lt;br /&gt;
&lt;br /&gt;
partição: short -&amp;gt; qos: qos_short -&amp;gt; limite de 2 semanas&lt;br /&gt;
&lt;br /&gt;
partição: long -&amp;gt; qos: qos_long -&amp;gt; limite de 3 meses&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
Caso deseje rodar em GPU, é necessário especificar a fila e pedir explicitamente a ''generic resource'' gpu:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#!/bin/bash &lt;br /&gt;
#SBATCH -n 1 # Numero de CPU cores a serem alocados &lt;br /&gt;
#SBATCH -N 1 # Numero de nodes a serem alocados&lt;br /&gt;
#SBATCH -t 0-00:05 # Tempo limite de execucao (D-HH:MM)&lt;br /&gt;
#SBATCH -p long # Particao (fila) a ser submetido&lt;br /&gt;
#SBATCH --qos qos_long # QOS &lt;br /&gt;
#SBATCH --gres=gpu:1&lt;br /&gt;
  &lt;br /&gt;
# Comandos de execução do seu programa:&lt;br /&gt;
./a.out&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Para pedir alguma GPU específica, use um constraint adicionando a linha:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --constraint=&amp;quot;gtx970&amp;quot;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Para submeter o job, execute o comando &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
sbatch job.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Comandos úteis ==&lt;br /&gt;
* Para listar os seus jobs:&lt;br /&gt;
  squeue&lt;br /&gt;
&lt;br /&gt;
* Para listar todos os jobs:&lt;br /&gt;
  sudo squeue&lt;br /&gt;
&lt;br /&gt;
* Para deletar um job:&lt;br /&gt;
  scancel [job_id]&lt;br /&gt;
&lt;br /&gt;
* Para listar as partições disponíveis:&lt;br /&gt;
  sinfo&lt;br /&gt;
&lt;br /&gt;
* Para listar as gpus presentes nos nodes:&lt;br /&gt;
  sinfo -o &amp;quot;%N %f&amp;quot;&lt;br /&gt;
&lt;br /&gt;
* Para listar um resumo de todos os nodes:&lt;br /&gt;
  sinfo -Nel&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
	<entry>
		<id>https://wiki.if.ufrgs.br/index.php?title=Minuano&amp;diff=658</id>
		<title>Minuano</title>
		<link rel="alternate" type="text/html" href="https://wiki.if.ufrgs.br/index.php?title=Minuano&amp;diff=658"/>
		<updated>2010-06-14T12:23:18Z</updated>

		<summary type="html">&lt;p&gt;Leon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Instalação dos nós'''&lt;br /&gt;
&lt;br /&gt;
1 - Instalação atual: debian squeeze&lt;br /&gt;
&lt;br /&gt;
2 - Pacotes necessários além do básico: nfs-common, nis&lt;br /&gt;
&lt;br /&gt;
a) Configuração do NFS: editar /etc/fstab&lt;br /&gt;
       # Substituir a opção ro por defaults na montagem do /.     &lt;br /&gt;
&lt;br /&gt;
       # Acrescentar ao final do arquivo os nfs mounts:&lt;br /&gt;
         10.0.0.06:/home_agua /home_agua nfs wsize=8192,rsize=8192 0 0&lt;br /&gt;
         10.0.0.05:/home05 /home05 nfs wsize=8192,rsize=8192 0 0&lt;br /&gt;
         10.0.0.25:/home25 /home25 nfs wsize=8192,rsize=8192 0 0&lt;br /&gt;
         10.0.1.3:/home103 /home103 nfs wsize=8192,rsize=8192 0 0&lt;br /&gt;
         10.0.1.11:/work05 /work05 nfs wsize=8192,rsize=8192 0 0&lt;br /&gt;
         10.0.1.12:/work25 /work25 nfs wsize=8192,rsize=8192 0 0&lt;br /&gt;
         10.0.1.13:/work103 /work103 nfs wsize=8192,rsize=8192 0 0&lt;br /&gt;
&lt;br /&gt;
        #Teste de montagem: mount -a; df -h&lt;br /&gt;
&lt;br /&gt;
b) Na configuração do nis: &lt;br /&gt;
&lt;br /&gt;
         # Acrescentar &amp;quot;ypserver 10.0.0.254&amp;quot; na ultima linha de /etc/yp.serv e comentar o resto.&lt;br /&gt;
&lt;br /&gt;
         # Acrescentar &amp;quot;cluster.if.ufrgs.br&amp;quot; na ultima linha de /etc/defaultdomain e comentar o resto.&lt;br /&gt;
&lt;br /&gt;
         # Editar o arquivo /etc/nsswitch.conf e reescrever as linhas de passwd, group e shadow adicionando nis, i.e.&lt;br /&gt;
            passwd:        nis  compat&lt;br /&gt;
            group:         nis compat&lt;br /&gt;
            shadow:        nis compat&lt;br /&gt;
&lt;br /&gt;
         # Lançar o nis: /etc/init.d/nis restart&lt;br /&gt;
&lt;br /&gt;
         # Testes para saber se está tudo funcionando:&lt;br /&gt;
           &amp;gt; ypcat passwd&lt;br /&gt;
           &amp;gt; su leon&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
3 - Cliente do torque (pbs_mom):http://debianclusters.org/index.php/Resource_Manager:_Torque&lt;/div&gt;</summary>
		<author><name>Leon</name></author>
	</entry>
</feed>