Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
infra:howto:runjob [2025/02/27 12:54] remiinfra:howto:runjob [2026/03/11 10:25] (current) – [Quotas] marc
Line 10: Line 10:
  
  
 +===== Connection =====
 +Currently, you have to SSH directly on one of the compute node, either server [[infra:chacha|Chacha]] or [[infra:disco|Disco]] :
 +
 +  $ ssh firstname.lastname@disco.hevs.ch
 +  $ ssh firstname.lastname@chacha.hevs.ch
 +  
 +** NOTE : Please note you have to connect from the school network or the HEVS VPN to be able to connect on those servers **
 +
 +TODO : Change doc when the jump host will be ready
  
 ===== Environment ===== ===== Environment =====
  
-On either server [[infra:chacha|Chacha]] or [[infra:disco|Disco]], you have a symlink **datasets** in your home directory that is linked to the local storage of the server : its purpose is to give you a proper space to put all the data you will be working on.+ 1. On either server [[infra:chacha|Chacha]] or [[infra:disco|Disco]], you have a symlink **datasets** in your home directory that is linked to the local storage of the server : its purpose is to give you a proper space to put all the data you will be working on.
  
-You have also another symlink **shared_dataset** for jobs that needs to be run on several nodes : this filesystem is shared between nodes.+ 2. You have also another symlink **shared_dataset** for jobs that needs to be run on several nodes : this filesystem is shared between nodes.
  
 + 3. Your .bashrc / .zshrc contains by default the variable **APPTAINER_TMPDIR** set to **/home/user.name/.apptainer/** : this allows you to build containers without using the system /tmp that is restricted with a low quota, and use your larger dataset quota instead.
  
-NOTE : By default, you are the only one seeing your data : If you are working as a team on these data, **please ask for a group creation** so we can add members in it and apply suitable permissions.+ 4. By default, you are the only one seeing your data : If you are working as a team on these data, **please ask for a group creation** so we can add members in it and apply suitable permissions.
  
  
Line 27: Line 37:
 To avoid having everyone installing their libraries installed on the system or on their user directly on the physical servers, we need you to keep them cleanly packed in a container : That way you can both install what you want inside this container, and you can do it without needing any root priviledge on the server you are sharing with other researchers. To avoid having everyone installing their libraries installed on the system or on their user directly on the physical servers, we need you to keep them cleanly packed in a container : That way you can both install what you want inside this container, and you can do it without needing any root priviledge on the server you are sharing with other researchers.
  
-See [[infra:howto:apptainer_sample|How to create a simple apptainer container]]+**For examples, see : [[infra:howto:apptainer_sample|How to create a simple apptainer container]]**
  
  
Line 34: Line 44:
 To be able to run a job on the ISC Compute Center, you **MUST** run it under [[https://slurm.schedmd.com/overview.html|SLURM]]. Ressource usage is managed by Slurm on this cluster. To be able to run a job on the ISC Compute Center, you **MUST** run it under [[https://slurm.schedmd.com/overview.html|SLURM]]. Ressource usage is managed by Slurm on this cluster.
  
-For examples, see : [[infra:howto:slurm_sample|How-to create a simple SLURM job]]+**For examples, see : [[infra:howto:slurm_sample|How-to create a simple SLURM job]]**
  
  
 +
 +More information on our **[[infra:compute:slurmconfig|SLURM cluster]]**
 ===== Storage considerations ===== ===== Storage considerations =====
  
Line 64: Line 76:
 **NOTE :** Please note the filer storage is slow, it is not advised to run a job directly using those data. **NOTE :** Please note the filer storage is slow, it is not advised to run a job directly using those data.
  
 +**TODO :** set auto-cleaning for old data on each filesystem
  
 +
 +
 +==== Quotas ====
 +  - The root filesystem (/, including /home) for every researchers is 20GB for convenience. For students, this quota is lowered to 10GB to encourage proper infrastructure usage / coding with a lower threshold.
 +  - On local and shared datasets, quota will be set on a case by case basis
  
 ==== Cleaning ==== ==== Cleaning ====
Line 81: Line 99:
    * [[https://github.com/microsoft/vscode-cpptools/issues/5362]]    * [[https://github.com/microsoft/vscode-cpptools/issues/5362]]
    * [[https://learn.microsoft.com/en-us/answers/questions/1221136/visual-studio-2022-clear-local-caches]]    * [[https://learn.microsoft.com/en-us/answers/questions/1221136/visual-studio-2022-clear-local-caches]]
-  - Currently your vscode caches will be automatically removed from our servers.+  - Currently your vscode caches are automatically removed from our servers
 + 
 +==== Automation ==== 
 + 
 +If you need to use Cron to schedule something, you need to ask for your user to be added to the /etc/cron.allow whitelist.
Back to top