<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.anunna.wur.nl/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Duque004</id>
	<title>HPCwiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.anunna.wur.nl/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Duque004"/>
	<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php/Special:Contributions/Duque004"/>
	<updated>2026-04-17T22:38:17Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.1</generator>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1922</id>
		<title>Shared folders</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1922"/>
		<updated>2019-01-30T11:19:30Z</updated>

		<summary type="html">&lt;p&gt;Duque004: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Working with shared folders in the Lustre file system ==&lt;br /&gt;
&lt;br /&gt;
If you work in a group or team and use large volumes of data, it is useful to work within a shared space. User can thus share inputs to their models and make their outputs also easily available. This article explains how to do so within the Lustre file system, that presently supports the HPC.&lt;br /&gt;
&lt;br /&gt;
== Identifying a suitable user group ==&lt;br /&gt;
&lt;br /&gt;
Users access the HPC cluster with their WUR-wide account. This means that all the membership information is also available in the HPC system. To check of which groups is your user a member of, use the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups &amp;lt;username&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can result in a rather long list, reflecting permissions in the overall WUR systems. Within these groups you must then identify the one that is closer to match the team or group with which you wish to collaborate.&lt;br /&gt;
&lt;br /&gt;
For instance, if I wish to work together with colleagues at ISRIC, I can search within my groups an appropriate match:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups duque004 | grep isric&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In my case the group des-isric-users looked appropriate. Then next step is to confirm if the other users in my team are also members of the group.&lt;br /&gt;
&lt;br /&gt;
== Creating a shared folder with correct permissions ==&lt;br /&gt;
&lt;br /&gt;
The Lustre file system is accessible in the &amp;lt;code&amp;gt;/lustre&amp;lt;/code&amp;gt; folder and then divided into the &amp;lt;code&amp;gt;/backup&amp;lt;/code&amp;gt; and &amp;lt;code&amp;gt;/nobackup&amp;lt;/code&amp;gt; sections (corresponding to the different usage plans). Inside each of these folders there is a sub-folder named &amp;lt;code&amp;gt;SHARED&amp;lt;/code&amp;gt; in which users are to create their own assets.&lt;br /&gt;
&lt;br /&gt;
You start by creating a folder in this space; it is probably better if it matches the name of your group or team, e.g.:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;mkdir /lustre/nobackup/SHARED/myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Or in alternative:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cd /lustre/nobackup/SHARED&lt;br /&gt;
&lt;br /&gt;
mkdir myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
== Setting permissions ==&lt;br /&gt;
&lt;br /&gt;
Three basic steps are involved in stepping permissions correctly:&lt;br /&gt;
&lt;br /&gt;
1. Pass the ownership of the group to the team. In the example below it is applied recursively to all sub-folder and files that may exist:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chgrp -R my-team-group myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. Concede read/write permissions to the group. This allows other members of the group to read and write in the shared folder. If you wish other team members to only read from the folder then remove the &amp;lt;code&amp;gt;w&amp;lt;/code&amp;gt; character from the &amp;lt;code&amp;gt;+rw&amp;lt;/code&amp;gt; bit:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+rw myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3. Set default ownership within the group. This guarantees that any new files or folders created within the shared folder are owned by default owned by your team group:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+s myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In case the contents of the shared are sensitive or private, and should be accessed by your team, you can block access from any other users with the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R o-rw myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Further reading ==&lt;br /&gt;
&lt;br /&gt;
[https://www.digitalocean.com/community/tutorials/an-introduction-to-linux-permissions An Introduction to Linux Permissions]&lt;br /&gt;
&lt;br /&gt;
[https://www.linode.com/docs/tools-reference/linux-users-and-groups/ Linux Users and Groups]&lt;br /&gt;
&lt;br /&gt;
[https://www.digitalocean.com/community/tutorials/linux-permissions-basics-and-how-to-use-umask-on-a-vps#types-of-permissions Linux Permissions Basics and How to Use Umask on a VPS]&lt;br /&gt;
&lt;br /&gt;
[http://www.yolinux.com/TUTORIALS/LinuxTutorialManagingGroups.html Linux Tutorial - Managing Group Access on Linux and UNIX]&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1921</id>
		<title>Shared folders</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1921"/>
		<updated>2019-01-30T11:18:02Z</updated>

		<summary type="html">&lt;p&gt;Duque004: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
= Working with shared folders in the Lustre file system =&lt;br /&gt;
&lt;br /&gt;
If you work in a group or team and use large volumes of data, it is useful to work within a shared space. User can thus share inputs to their models and make their outputs also easily available. This article explains how to do so within the Lustre file system, that presently supports the HPC.&lt;br /&gt;
&lt;br /&gt;
== Identifying a suitable user group ==&lt;br /&gt;
&lt;br /&gt;
Users access the HPC cluster with their WUR-wide account. This means that all the membership information is also available in the HPC system. To check of which groups is your user a member of, use the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups &amp;lt;username&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can result in a rather long list, reflecting permissions in the overall WUR systems. Within these groups you must then identify the one that is closer to match the team or group with which you wish to collaborate.&lt;br /&gt;
&lt;br /&gt;
For instance, if I wish to work together with colleagues at ISRIC, I can search within my groups an appropriate match:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups duque004 | grep isric&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In my case the group des-isric-users looked appropriate. Then next step is to confirm if the other users in my team are also members of the group.&lt;br /&gt;
&lt;br /&gt;
== Creating a shared folder with correct permissions ==&lt;br /&gt;
&lt;br /&gt;
The Lustre file system is accessible in the &amp;lt;code&amp;gt;/lustre&amp;lt;/code&amp;gt; folder and then divided into the &amp;lt;code&amp;gt;/backup&amp;lt;/code&amp;gt; and &amp;lt;code&amp;gt;/nobackup&amp;lt;/code&amp;gt; sections (corresponding to the different usage plans). Inside each of these folders there is a sub-folder named &amp;lt;code&amp;gt;SHARED&amp;lt;/code&amp;gt; in which users are to create their own assets.&lt;br /&gt;
&lt;br /&gt;
You start by creating a folder in this space; it is probably better if it matches the name of your group or team, e.g.:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;mkdir /lustre/nobackup/SHARED/myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Or in alternative:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cd /lustre/nobackup/SHARED&lt;br /&gt;
&lt;br /&gt;
mkdir myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
== Setting permissions ==&lt;br /&gt;
&lt;br /&gt;
Three basic steps are involved in stepping permissions correctly:&lt;br /&gt;
&lt;br /&gt;
1. Pass the ownership of the group to the team. In the example below it is applied recursively to all sub-folder and files that may exist:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chgrp -R my-team-group myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. Concede read/write permissions to the group. This allows other members of the group to read and write in the shared folder. If you wish other team members to only read from the folder then remove the &amp;lt;code&amp;gt;w&amp;lt;/code&amp;gt; character from the &amp;lt;code&amp;gt;+rw&amp;lt;/code&amp;gt; bit:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+rw myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3. Set default ownership within the group. This guarantees that any new files or folders created within the shared folder are owned by default owned by your team group:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+s myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In case the contents of the shared are sensitive or private, and should be accessed by your team, you can block access from any other users with the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R o-rw myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Further reading ==&lt;br /&gt;
&lt;br /&gt;
[https://www.digitalocean.com/community/tutorials/an-introduction-to-linux-permissions An Introduction to Linux Permissions]&lt;br /&gt;
&lt;br /&gt;
[https://www.linode.com/docs/tools-reference/linux-users-and-groups/ Linux Users and Groups]&lt;br /&gt;
&lt;br /&gt;
[https://www.digitalocean.com/community/tutorials/linux-permissions-basics-and-how-to-use-umask-on-a-vps#types-of-permissions Linux Permissions Basics and How to Use Umask on a VPS]&lt;br /&gt;
&lt;br /&gt;
[http://www.yolinux.com/TUTORIALS/LinuxTutorialManagingGroups.html Linux Tutorial - Managing Group Access on Linux and UNIX]&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1920</id>
		<title>Shared folders</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1920"/>
		<updated>2019-01-30T11:16:28Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Further reading */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Working with shared folders in the Lustre file system ==&lt;br /&gt;
&lt;br /&gt;
If you work in a group or team and use large volumes of data, it is useful to work within a shared space. User can thus share inputs to their models and make their outputs also easily available. This article explains how to do so within the Lustre file system, that presently supports the HPC.&lt;br /&gt;
&lt;br /&gt;
=== Identifying a suitable user group ===&lt;br /&gt;
&lt;br /&gt;
Users access the HPC cluster with their WUR-wide account. This means that all the membership information is also available in the HPC system. To check of which groups is your user a member of, use the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups &amp;lt;username&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can result in a rather long list, reflecting permissions in the overall WUR systems. Within these groups you must then identify the one that is closer to match the team or group with which you wish to collaborate.&lt;br /&gt;
&lt;br /&gt;
For instance, if I wish to work together with colleagues at ISRIC, I can search within my groups an appropriate match:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups duque004 | grep isric&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In my case the group des-isric-users looked appropriate. Then next step is to confirm if the other users in my team are also members of the group.&lt;br /&gt;
&lt;br /&gt;
=== Creating a shared folder with correct permissions ===&lt;br /&gt;
&lt;br /&gt;
The Lustre file system is accessible in the &amp;lt;code&amp;gt;/lustre&amp;lt;/code&amp;gt; folder and then divided into the &amp;lt;code&amp;gt;/backup&amp;lt;/code&amp;gt; and &amp;lt;code&amp;gt;/nobackup&amp;lt;/code&amp;gt; sections (corresponding to the different usage plans). Inside each of these folders there is a sub-folder named &amp;lt;code&amp;gt;SHARED&amp;lt;/code&amp;gt; in which users are to create their own assets.&lt;br /&gt;
&lt;br /&gt;
You start by creating a folder in this space; it is probably better if it matches the name of your group or team, e.g.:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;mkdir /lustre/nobackup/SHARED/myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Or in alternative:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cd /lustre/nobackup/SHARED&lt;br /&gt;
&lt;br /&gt;
mkdir myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
=== Setting permissions ===&lt;br /&gt;
&lt;br /&gt;
Three basic steps are involved in stepping permissions correctly:&lt;br /&gt;
&lt;br /&gt;
1. Pass the ownership of the group to the team. In the example below it is applied recursively to all sub-folder and files that may exist:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chgrp -R my-team-group myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. Concede read/write permissions to the group. This allows other members of the group to read and write in the shared folder. If you wish other team members to only read from the folder then remove the &amp;lt;code&amp;gt;w&amp;lt;/code&amp;gt; character from the &amp;lt;code&amp;gt;+rw&amp;lt;/code&amp;gt; bit:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+rw myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3. Set default ownership within the group. This guarantees that any new files or folders created within the shared folder are owned by default owned by your team group:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+s myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In case the contents of the shared are sensitive or private, and should be accessed by your team, you can block access from any other users with the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R o-rw myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further reading ===&lt;br /&gt;
&lt;br /&gt;
[https://www.digitalocean.com/community/tutorials/an-introduction-to-linux-permissions An Introduction to Linux Permissions]&lt;br /&gt;
&lt;br /&gt;
[https://www.linode.com/docs/tools-reference/linux-users-and-groups/ Linux Users and Groups]&lt;br /&gt;
&lt;br /&gt;
[https://www.digitalocean.com/community/tutorials/linux-permissions-basics-and-how-to-use-umask-on-a-vps#types-of-permissions Linux Permissions Basics and How to Use Umask on a VPS]&lt;br /&gt;
&lt;br /&gt;
[http://www.yolinux.com/TUTORIALS/LinuxTutorialManagingGroups.html Linux Tutorial - Managing Group Access on Linux and UNIX]&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1919</id>
		<title>Shared folders</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1919"/>
		<updated>2019-01-30T11:05:35Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Setting permissions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Working with shared folders in the Lustre file system ==&lt;br /&gt;
&lt;br /&gt;
If you work in a group or team and use large volumes of data, it is useful to work within a shared space. User can thus share inputs to their models and make their outputs also easily available. This article explains how to do so within the Lustre file system, that presently supports the HPC.&lt;br /&gt;
&lt;br /&gt;
=== Identifying a suitable user group ===&lt;br /&gt;
&lt;br /&gt;
Users access the HPC cluster with their WUR-wide account. This means that all the membership information is also available in the HPC system. To check of which groups is your user a member of, use the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups &amp;lt;username&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can result in a rather long list, reflecting permissions in the overall WUR systems. Within these groups you must then identify the one that is closer to match the team or group with which you wish to collaborate.&lt;br /&gt;
&lt;br /&gt;
For instance, if I wish to work together with colleagues at ISRIC, I can search within my groups an appropriate match:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups duque004 | grep isric&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In my case the group des-isric-users looked appropriate. Then next step is to confirm if the other users in my team are also members of the group.&lt;br /&gt;
&lt;br /&gt;
=== Creating a shared folder with correct permissions ===&lt;br /&gt;
&lt;br /&gt;
The Lustre file system is accessible in the &amp;lt;code&amp;gt;/lustre&amp;lt;/code&amp;gt; folder and then divided into the &amp;lt;code&amp;gt;/backup&amp;lt;/code&amp;gt; and &amp;lt;code&amp;gt;/nobackup&amp;lt;/code&amp;gt; sections (corresponding to the different usage plans). Inside each of these folders there is a sub-folder named &amp;lt;code&amp;gt;SHARED&amp;lt;/code&amp;gt; in which users are to create their own assets.&lt;br /&gt;
&lt;br /&gt;
You start by creating a folder in this space; it is probably better if it matches the name of your group or team, e.g.:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;mkdir /lustre/nobackup/SHARED/myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Or in alternative:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cd /lustre/nobackup/SHARED&lt;br /&gt;
&lt;br /&gt;
mkdir myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
=== Setting permissions ===&lt;br /&gt;
&lt;br /&gt;
Three basic steps are involved in stepping permissions correctly:&lt;br /&gt;
&lt;br /&gt;
1. Pass the ownership of the group to the team. In the example below it is applied recursively to all sub-folder and files that may exist:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chgrp -R my-team-group myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. Concede read/write permissions to the group. This allows other members of the group to read and write in the shared folder. If you wish other team members to only read from the folder then remove the &amp;lt;code&amp;gt;w&amp;lt;/code&amp;gt; character from the &amp;lt;code&amp;gt;+rw&amp;lt;/code&amp;gt; bit:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+rw myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3. Set default ownership within the group. This guarantees that any new files or folders created within the shared folder are owned by default owned by your team group:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+s myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In case the contents of the shared are sensitive or private, and should be accessed by your team, you can block access from any other users with the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R o-rw myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further reading ===&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1918</id>
		<title>Shared folders</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1918"/>
		<updated>2019-01-30T10:53:57Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Creating a shared folder with correct permissions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Working with shared folders in the Lustre file system ==&lt;br /&gt;
&lt;br /&gt;
If you work in a group or team and use large volumes of data, it is useful to work within a shared space. User can thus share inputs to their models and make their outputs also easily available. This article explains how to do so within the Lustre file system, that presently supports the HPC.&lt;br /&gt;
&lt;br /&gt;
=== Identifying a suitable user group ===&lt;br /&gt;
&lt;br /&gt;
Users access the HPC cluster with their WUR-wide account. This means that all the membership information is also available in the HPC system. To check of which groups is your user a member of, use the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups &amp;lt;username&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can result in a rather long list, reflecting permissions in the overall WUR systems. Within these groups you must then identify the one that is closer to match the team or group with which you wish to collaborate.&lt;br /&gt;
&lt;br /&gt;
For instance, if I wish to work together with colleagues at ISRIC, I can search within my groups an appropriate match:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups duque004 | grep isric&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In my case the group des-isric-users looked appropriate. Then next step is to confirm if the other users in my team are also members of the group.&lt;br /&gt;
&lt;br /&gt;
=== Creating a shared folder with correct permissions ===&lt;br /&gt;
&lt;br /&gt;
The Lustre file system is accessible in the &amp;lt;code&amp;gt;/lustre&amp;lt;/code&amp;gt; folder and then divided into the &amp;lt;code&amp;gt;/backup&amp;lt;/code&amp;gt; and &amp;lt;code&amp;gt;/nobackup&amp;lt;/code&amp;gt; sections (corresponding to the different usage plans). Inside each of these folders there is a sub-folder named &amp;lt;code&amp;gt;SHARED&amp;lt;/code&amp;gt; in which users are to create their own assets.&lt;br /&gt;
&lt;br /&gt;
You start by creating a folder in this space; it is probably better if it matches the name of your group or team, e.g.:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;mkdir /lustre/nobackup/SHARED/myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Or in alternative:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cd /lustre/nobackup/SHARED&lt;br /&gt;
&lt;br /&gt;
mkdir myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
=== Setting permissions ===&lt;br /&gt;
&lt;br /&gt;
Three basic steps are involved in stepping permissions correctly:&lt;br /&gt;
&lt;br /&gt;
1. Pass the ownership of the group to the team. In the example below it is applied recursively to all sub-folder and files that may exist:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chgrp -R my-team-group myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. Concede read/write permissions to the group. This allows other members of the group to read and write in the shared folder. If you wish other team members to only read from the folder then remove the &amp;lt;code&amp;gt;w&amp;lt;/code&amp;gt; character from the &amp;lt;code&amp;gt;+rw&amp;lt;/code&amp;gt; bit:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+rw myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3. Set default ownership within the group. This guarantees that any new files or folders created within the folder are owned by your team group by default:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+s myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further reading ===&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1917</id>
		<title>Shared folders</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1917"/>
		<updated>2019-01-30T10:53:04Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Creating a shared folder in Lustre */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Working with shared folders in the Lustre file system ==&lt;br /&gt;
&lt;br /&gt;
If you work in a group or team and use large volumes of data, it is useful to work within a shared space. User can thus share inputs to their models and make their outputs also easily available. This article explains how to do so within the Lustre file system, that presently supports the HPC.&lt;br /&gt;
&lt;br /&gt;
=== Identifying a suitable user group ===&lt;br /&gt;
&lt;br /&gt;
Users access the HPC cluster with their WUR-wide account. This means that all the membership information is also available in the HPC system. To check of which groups is your user a member of, use the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups &amp;lt;username&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can result in a rather long list, reflecting permissions in the overall WUR systems. Within these groups you must then identify the one that is closer to match the team or group with which you wish to collaborate.&lt;br /&gt;
&lt;br /&gt;
For instance, if I wish to work together with colleagues at ISRIC, I can search within my groups an appropriate match:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups duque004 | grep isric&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In my case the group des-isric-users looked appropriate. Then next step is to confirm if the other users in my team are also members of the group.&lt;br /&gt;
&lt;br /&gt;
=== Creating a shared folder with correct permissions ===&lt;br /&gt;
&lt;br /&gt;
The Lustre file system is accessible in the &amp;lt;code&amp;gt;/lustre&amp;lt;/code&amp;gt; folder and then divided into the &amp;lt;code&amp;gt;/backup&amp;lt;/code&amp;gt; and &amp;lt;code&amp;gt;/nobackup&amp;lt;/code&amp;gt; sections (corresponding to the different usage plans). Inside each of these folders there is a sub-folder named &amp;lt;code&amp;gt;SHARED&amp;lt;/code&amp;gt; in which users are to create their own assets.&lt;br /&gt;
&lt;br /&gt;
You start by creating a folder in this space; it is probably better if it matches the name of your group or team, e.g.:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;mkdir /lustre/nobackup/SHARED/myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Or in alternative:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cd /lustre/nobackup/SHARED&lt;br /&gt;
&lt;br /&gt;
mkdir myTeamWorkspace&amp;lt;/code&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Now come the permissions. Three basic steps are involved:&lt;br /&gt;
&lt;br /&gt;
1. Pass the ownership of the group to the team. In the example below it is applied recursively to all sub-folder and files that may exist:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chgrp -R my-team-group myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. Concede read/write permissions to the group. This allows other members of the group to read and write in the shared folder. If you wish other team members to only read from the folder then remove the &amp;lt;code&amp;gt;w&amp;lt;/code&amp;gt; character from the &amp;lt;code&amp;gt;+rw&amp;lt;/code&amp;gt; bit:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+rw myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3. Set default ownership within the group. This guarantees that any new files or folders created within the folder are owned by your team group by default:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;chmod -R g+s myTeamWorkspace&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further reading ===&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1916</id>
		<title>Shared folders</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1916"/>
		<updated>2019-01-30T10:37:06Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Identifying a suitable user group */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Working with shared folders in the Lustre file system ==&lt;br /&gt;
&lt;br /&gt;
If you work in a group or team and use large volumes of data, it is useful to work within a shared space. User can thus share inputs to their models and make their outputs also easily available. This article explains how to do so within the Lustre file system, that presently supports the HPC.&lt;br /&gt;
&lt;br /&gt;
=== Identifying a suitable user group ===&lt;br /&gt;
&lt;br /&gt;
Users access the HPC cluster with their WUR-wide account. This means that all the membership information is also available in the HPC system. To check of which groups is your user a member of, use the following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups &amp;lt;username&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can result in a rather long list, reflecting permissions in the overall WUR systems. Within these groups you must then identify the one that is closer to match the team or group with which you wish to collaborate.&lt;br /&gt;
&lt;br /&gt;
For instance, if I wish to work together with colleagues at ISRIC, I can search within my groups an appropriate match:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;groups duque004 | grep isric&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In my case the group des-isric-users looked appropriate. Then next step is to confirm if the other users in my team are also members of the group.&lt;br /&gt;
&lt;br /&gt;
=== Creating a shared folder in Lustre ===&lt;br /&gt;
&lt;br /&gt;
The&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1915</id>
		<title>Shared folders</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Shared_folders&amp;diff=1915"/>
		<updated>2019-01-30T10:36:21Z</updated>

		<summary type="html">&lt;p&gt;Duque004: Created page with &amp;quot; == Working with shared folders in the Lustre file system ==  If you work in a group or team and use large volumes of data, it is useful to work within a shared space. User ca...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Working with shared folders in the Lustre file system ==&lt;br /&gt;
&lt;br /&gt;
If you work in a group or team and use large volumes of data, it is useful to work within a shared space. User can thus share inputs to their models and make their outputs also easily available. This article explains how to do so within the Lustre file system, that presently supports the HPC.&lt;br /&gt;
&lt;br /&gt;
=== Identifying a suitable user group ===&lt;br /&gt;
&lt;br /&gt;
Users access the HPC cluster with their WUR-wide account. This means that all the membership information is also available in the HPC system. To check of which groups is your user a member of, use the following command:&lt;br /&gt;
&lt;br /&gt;
$ groups &amp;lt;username&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can result in a rather long list, reflecting permissions in the overall WUR systems. Within these groups you must then identify the one that is closer to match the team or group with which you wish to collaborate.&lt;br /&gt;
&lt;br /&gt;
For instance, if I wish to work together with colleagues at ISRIC, I can search within my groups an appropriate match:&lt;br /&gt;
&lt;br /&gt;
$ groups duque004 | grep isric&lt;br /&gt;
&lt;br /&gt;
In my case the group des-isric-users looked appropriate. Then next step is to confirm if the other users in my team are also members of the group.&lt;br /&gt;
&lt;br /&gt;
=== Creating a shared folder in Lustre ===&lt;br /&gt;
&lt;br /&gt;
The&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Main_Page&amp;diff=1914</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Main_Page&amp;diff=1914"/>
		<updated>2019-01-30T10:16:43Z</updated>

		<summary type="html">&lt;p&gt;Duque004: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The Agrogenomics cluster is a [http://en.wikipedia.org/wiki/High-performance_computing High Performance Computer] (HPC) infrastructure hosted by [http://www.wageningenur.nl/nl/activiteit/Opening-High-Performance-Computing-cluster-HPC.htm Wageningen University &amp;amp; Research Centre]. It is open for use for all WUR research groups as well as other organizations, including companies, that have collaborative projects with WUR. &lt;br /&gt;
&lt;br /&gt;
The Agrogenomics HPC was an initiative of the [http://www.breed4food.com/en/breed4food.htm Breed4Food] (B4F) consortium, consisting of the [[About_ABGC | Animal Breeding and Genomics Centre]] (WU-Animal Breeding and Genomics and Wageningen Livestock Research) and four major breeding companies: [http://www.cobb-vantress.com Cobb-Vantress], [https://www.crv4all.nl CRV], [http://www.hendrix-genetics.com Hendrix Genetics], and [http://www.topigs.com TOPIGS]. Currently, in addition to the original partners, the HPC (HPC-Ag) is used by other groups from Wageningen UR (Bioinformatics, Centre for Crop Systems Analysis, Environmental Sciences Group, and Plant Research International) and plant breeding industry (Rijk Zwaan). &lt;br /&gt;
&lt;br /&gt;
= Using the HPC-Ag =&lt;br /&gt;
== Gaining access to the HPC-Ag ==&lt;br /&gt;
Access to the cluster and file transfer are traditionally done via [http://en.wikipedia.org/wiki/Secure_Shell SSH and SFTP].&lt;br /&gt;
* [[log_in_to_B4F_cluster | Logging into cluster using ssh and file transfer]]&lt;br /&gt;
* [[Services | Alternative access methods, and extra features and services on the HPC]]&lt;br /&gt;
* [[Filesystems | Accessible storage methods on the HPC]]&lt;br /&gt;
&lt;br /&gt;
== Access Policy ==&lt;br /&gt;
[[Access_Policy | Main Article: Access Policy]]&lt;br /&gt;
&lt;br /&gt;
Access needs to be granted actively (by creation of an account on the cluster by FB-IT). Use of resources is limited by the scheduler. Depending on availability of queues (&#039;partitions&#039;) granted to a user, priority to the system&#039;s resources is regulated. Note that the use of the HPC-Ag is not free of charge. List price of CPU time and storage, and possible discounts on that list price for your organisation, can be retrieved from CAT-AGRO or FB-ICT.&lt;br /&gt;
&lt;br /&gt;
= Events =&lt;br /&gt;
* [[Courses]] that have happened and are happening&lt;br /&gt;
* [[Downtime]] that will affect all users&lt;br /&gt;
* [[Meetings]] that may affect the policies of the HPC&lt;br /&gt;
&lt;br /&gt;
= Other Software =&lt;br /&gt;
&lt;br /&gt;
== Cluster Management Software and Scheduler ==&lt;br /&gt;
The HPC-Ag uses Bright Cluster Manager software for overall cluster management, and Slurm as job scheduler.&lt;br /&gt;
* [[BCM_on_B4F_cluster | Monitor cluster status with BCM]]&lt;br /&gt;
* [[SLURM_on_B4F_cluster | Submit jobs with Slurm]]&lt;br /&gt;
* [[node_usage_graph | Be aware of how much work the cluster is under right now with &#039;node_usage_graph&#039;]]&lt;br /&gt;
* [[SLURM_Compare | Rosetta Stone of Workload Managers]]&lt;br /&gt;
&lt;br /&gt;
== Installation of software by users ==&lt;br /&gt;
&lt;br /&gt;
* [[Domain_specific_software_on_B4Fcluster_installation_by_users | Installing domain specific software: installation by users]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
* [[Virtual_environment_Python_3.4_or_higher | Setting up and using a virtual environment for Python3.4 or higher ]]&lt;br /&gt;
* [[Installing WRF and WPS]]&lt;br /&gt;
&lt;br /&gt;
== Installed software ==&lt;br /&gt;
&lt;br /&gt;
* [[Globally_installed_software | Globally installed software]]&lt;br /&gt;
* [[ABGC_modules | ABGC specific modules]]&lt;br /&gt;
&lt;br /&gt;
= Useful Notes = &lt;br /&gt;
&lt;br /&gt;
== Being in control of Environment parameters ==&lt;br /&gt;
&lt;br /&gt;
* [[Using_environment_modules | Using environment modules]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Setting_TMPDIR | Set a custom temporary directory location]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
&lt;br /&gt;
== Controlling costs ==&lt;br /&gt;
&lt;br /&gt;
* [[SACCT | using SACCT to see your costs]]&lt;br /&gt;
* [[get_my_bill | using the &amp;quot;get_my_bill&amp;quot; script to estimate costs]]&lt;br /&gt;
&lt;br /&gt;
== Management ==&lt;br /&gt;
[[HPC_management | Main Article: HPC management]]&lt;br /&gt;
&lt;br /&gt;
Project Leader of the HPC is Stephen Janssen (Wageningen UR,FB-IT, Service Management). [[User:lith010 | Jan van Lith (Wageningen UR,FB-IT, Infrastructure)]] and [[User:dawes001 | Gwen Dawes (Wageningen UR, FB-IT, Infrastructure)]] are responsible for [[Maintenance_and_Management | Maintenance and Management]].&lt;br /&gt;
&lt;br /&gt;
= Miscellaneous =&lt;br /&gt;
* [[Mailinglist | Electronic mail discussion lists]]&lt;br /&gt;
* [[History_of_the_Cluster | Historical information on the startup of the HPC]]&lt;br /&gt;
* [[Bioinformatics_tips_tricks_workflows | Bioinformatics tips, tricks, and workflows]]&lt;br /&gt;
* [[Parallel_R_code_on_SLURM | Running parallel R code on SLURM]]&lt;br /&gt;
* [[Convert_between_MediaWiki_and_other_formats | Convert between MediaWiki format and other formats]]&lt;br /&gt;
* [[Manual GitLab | GitLab: Create projects and add scripts]]&lt;br /&gt;
* [[Monitoring_executions | Monitoring job execution]]&lt;br /&gt;
* [[Shared_folders | Working with shared folders in the Lustre file system]]&lt;br /&gt;
&lt;br /&gt;
= See also =&lt;br /&gt;
* [[Maintenance_and_Management | Maintenance and Management]]&lt;br /&gt;
* [[BCData | BCData]]&lt;br /&gt;
* [[Mailinglist | Electronic mail discussion lists]]&lt;br /&gt;
* [[About_ABGC | About ABGC]]&lt;br /&gt;
* [[Computer_cluster | High Performance Computing @ABGC]]&lt;br /&gt;
* [[Lustre_PFS_layout | Lustre Parallel File System layout]]&lt;br /&gt;
&lt;br /&gt;
= External links =&lt;br /&gt;
{| width=&amp;quot;90%&amp;quot;&lt;br /&gt;
|- valign=&amp;quot;top&amp;quot;&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [http://www.breed4food.com/en/show/Breed4Food-initiative-reinforces-the-Netherlands-position-as-an-innovative-country-in-animal-breeding-and-genomics.htm Breed4Food programme]&lt;br /&gt;
* [http://www.wageningenur.nl/en/Expertise-Services/Facilities/CATAgroFood-3/CATAgroFood-3/Our-facilities/Show/High-Performance-Computing-Cluster-HPC.htm CATAgroFood offers a HPC facilty]&lt;br /&gt;
* [http://www.cobb-vantress.com Cobb-Vantress homepage]&lt;br /&gt;
&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [https://www.crv4all.nl CRV homepage]&lt;br /&gt;
* [http://www.hendrix-genetics.com Hendrix Genetics homepage]&lt;br /&gt;
* [http://www.topigs.com TOPIGS homepage]&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Scientific_Linux Scientific Linux]&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Help:Cheatsheet Help with editing Wiki pages]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1903</id>
		<title>Monitoring job execution</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1903"/>
		<updated>2018-08-27T13:26:26Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Output stream redirection */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Output stream redirection ==&lt;br /&gt;
&lt;br /&gt;
The primary way to monitor job execution is through the &#039;&#039;stdout&#039;&#039; and &#039;&#039;stderr&#039;&#039; streams. These are redirected to [https://wiki.hpcagrogenomics.wur.nl/index.php/Creating_sbatch_script#output_.28stderr.2Cstdout.29_directed_to_file text files specified in the SLURM script].&lt;br /&gt;
&lt;br /&gt;
For this purpose the &amp;lt;code&amp;gt;tail&amp;lt;/code&amp;gt; command is particularly useful. To continuously follow the output to a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -f output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To obtain the last X lines of a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -n X output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Replacing X by the desired number of lines.&lt;br /&gt;
&lt;br /&gt;
If the output file gets too long and you wish to read from the begining you may combine the commands &amp;lt;code&amp;gt;cat&amp;lt;/code&amp;gt; and &amp;lt;code&amp;gt;less&amp;lt;/code&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cat output_987654.txt | less&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Use the Q key to exit &amp;lt;code&amp;gt;less&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Monitoring resource usage ==&lt;br /&gt;
&lt;br /&gt;
While the output streams may suffice in most cases, certain programmes might not provide much feedback. This could be the case with a programme that rellies on modules that are not verbose. In such situations it is best to monitor resource usage to gauge job execution. Two possible options are described below.&lt;br /&gt;
&lt;br /&gt;
=== Using sstat ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; is a SLURM tool that can be used to obtain instantaneous information on resource usage, CPU load, memory, etc. To use it start by changing your SLURM script so that your programme is launched with the &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; command (this should be the last line in the script):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note down the job number. During execution you can then use &amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt;, passing the job number with the &amp;lt;code&amp;gt;-j&amp;lt;/code&amp;gt; flag:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat --format=AveCPU,AveRSS,MaxRSS -P -j 987654&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; can provide information on many different variables, for more details check [https://slurm.schedmd.com/sstat.html the manual].&lt;br /&gt;
&lt;br /&gt;
=== Using top ===&lt;br /&gt;
&lt;br /&gt;
Another way is to log the output of the [http://man7.org/linux/man-pages/man1/top.1.html &amp;lt;code&amp;gt;top&amp;lt;/code&amp;gt;] command. This requires adding an extra command to your SLURM script, again using &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; (where &amp;quot;user001&amp;quot; should be replaced by your user name): &lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun --overcommit --ntasks=1 top -b -u user001 &amp;amp;&lt;br /&gt;
python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This will log the output of &amp;lt;code&amp;gt;top&amp;lt;/code&amp;gt; to the sdtout stream file specific in the SLURM script every 3 seconds. Using the &amp;lt;code&amp;gt;tail&amp;lt;/code&amp;gt; command you will be able to see logs like:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;top - 18:09:12 up 53 days, 22:53,  0 users,  load average: 27,04, 27,63, 26,53&lt;br /&gt;
Tasks: 1068 total,   4 running, 1064 sleeping,   0 stopped,   0 zombie&lt;br /&gt;
%Cpu(s): 41,9 us,  0,2 sy,  0,0 ni, 57,9 id,  0,0 wa,  0,0 hi,  0,0 si,  0,0 st&lt;br /&gt;
KiB Mem : 10439453+total, 18650486+free, 43690281+used, 42053763+buff/cache&lt;br /&gt;
KiB Swap: 26214400+total, 26214393+free,       52 used. 54182425+avail Mem &lt;br /&gt;
&lt;br /&gt;
  PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND&lt;br /&gt;
29260 user001   20   0   15,6g  15,2g  17392 R  1777  1,5   4014:16 R&lt;br /&gt;
29447 user001   20   0  178604  12900   1676 R   1,6  0,0   6:33.17 top&lt;br /&gt;
28627 user001   20   0  113184   1520   1260 S   0,0  0,0   0:00.01 slurm_scri+&lt;br /&gt;
29253 user001   20   0  245096   4792   1976 S   0,0  0,0   0:00.66 srun&lt;br /&gt;
29357 user001   20   0   36124    688     16 S   0,0  0,0   0:00.00 srun&amp;lt;/pre&amp;gt;&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1902</id>
		<title>Monitoring job execution</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1902"/>
		<updated>2018-08-27T13:25:24Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Using top */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Output stream redirection ==&lt;br /&gt;
&lt;br /&gt;
The primary way to monitor job execution is through the stdout and stderr streams. These are redirected to [https://wiki.hpcagrogenomics.wur.nl/index.php/Creating_sbatch_script#output_.28stderr.2Cstdout.29_directed_to_file text files specified in the SLURM script].&lt;br /&gt;
&lt;br /&gt;
For this purpose the &amp;lt;code&amp;gt;tail&amp;lt;/code&amp;gt; command is particularly useful. To continuously follow the output to a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -f output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To obtain the last X lines of a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -n X output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Replacing X by the desired number of lines.&lt;br /&gt;
&lt;br /&gt;
If the output file gets too long and you wish to read from the begining you may combine the commands cat and less:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cat output_987654.txt | less&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Use the Q key to exit less.&lt;br /&gt;
&lt;br /&gt;
== Monitoring resource usage ==&lt;br /&gt;
&lt;br /&gt;
While the output streams may suffice in most cases, certain programmes might not provide much feedback. This could be the case with a programme that rellies on modules that are not verbose. In such situations it is best to monitor resource usage to gauge job execution. Two possible options are described below.&lt;br /&gt;
&lt;br /&gt;
=== Using sstat ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; is a SLURM tool that can be used to obtain instantaneous information on resource usage, CPU load, memory, etc. To use it start by changing your SLURM script so that your programme is launched with the &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; command (this should be the last line in the script):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note down the job number. During execution you can then use &amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt;, passing the job number with the &amp;lt;code&amp;gt;-j&amp;lt;/code&amp;gt; flag:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat --format=AveCPU,AveRSS,MaxRSS -P -j 987654&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; can provide information on many different variables, for more details check [https://slurm.schedmd.com/sstat.html the manual].&lt;br /&gt;
&lt;br /&gt;
=== Using top ===&lt;br /&gt;
&lt;br /&gt;
Another way is to log the output of the [http://man7.org/linux/man-pages/man1/top.1.html &amp;lt;code&amp;gt;top&amp;lt;/code&amp;gt;] command. This requires adding an extra command to your SLURM script, again using &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; (where &amp;quot;user001&amp;quot; should be replaced by your user name): &lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun --overcommit --ntasks=1 top -b -u user001 &amp;amp;&lt;br /&gt;
python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This will log the output of &amp;lt;code&amp;gt;top&amp;lt;/code&amp;gt; to the sdtout stream file specific in the SLURM script every 3 seconds. Using the &amp;lt;code&amp;gt;tail&amp;lt;/code&amp;gt; command you will be able to see logs like:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;top - 18:09:12 up 53 days, 22:53,  0 users,  load average: 27,04, 27,63, 26,53&lt;br /&gt;
Tasks: 1068 total,   4 running, 1064 sleeping,   0 stopped,   0 zombie&lt;br /&gt;
%Cpu(s): 41,9 us,  0,2 sy,  0,0 ni, 57,9 id,  0,0 wa,  0,0 hi,  0,0 si,  0,0 st&lt;br /&gt;
KiB Mem : 10439453+total, 18650486+free, 43690281+used, 42053763+buff/cache&lt;br /&gt;
KiB Swap: 26214400+total, 26214393+free,       52 used. 54182425+avail Mem &lt;br /&gt;
&lt;br /&gt;
  PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND&lt;br /&gt;
29260 user001   20   0   15,6g  15,2g  17392 R  1777  1,5   4014:16 R&lt;br /&gt;
29447 user001   20   0  178604  12900   1676 R   1,6  0,0   6:33.17 top&lt;br /&gt;
28627 user001   20   0  113184   1520   1260 S   0,0  0,0   0:00.01 slurm_scri+&lt;br /&gt;
29253 user001   20   0  245096   4792   1976 S   0,0  0,0   0:00.66 srun&lt;br /&gt;
29357 user001   20   0   36124    688     16 S   0,0  0,0   0:00.00 srun&amp;lt;/pre&amp;gt;&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1901</id>
		<title>Monitoring job execution</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1901"/>
		<updated>2018-08-27T13:24:15Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Using sstat */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Output stream redirection ==&lt;br /&gt;
&lt;br /&gt;
The primary way to monitor job execution is through the stdout and stderr streams. These are redirected to [https://wiki.hpcagrogenomics.wur.nl/index.php/Creating_sbatch_script#output_.28stderr.2Cstdout.29_directed_to_file text files specified in the SLURM script].&lt;br /&gt;
&lt;br /&gt;
For this purpose the &amp;lt;code&amp;gt;tail&amp;lt;/code&amp;gt; command is particularly useful. To continuously follow the output to a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -f output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To obtain the last X lines of a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -n X output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Replacing X by the desired number of lines.&lt;br /&gt;
&lt;br /&gt;
If the output file gets too long and you wish to read from the begining you may combine the commands cat and less:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cat output_987654.txt | less&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Use the Q key to exit less.&lt;br /&gt;
&lt;br /&gt;
== Monitoring resource usage ==&lt;br /&gt;
&lt;br /&gt;
While the output streams may suffice in most cases, certain programmes might not provide much feedback. This could be the case with a programme that rellies on modules that are not verbose. In such situations it is best to monitor resource usage to gauge job execution. Two possible options are described below.&lt;br /&gt;
&lt;br /&gt;
=== Using sstat ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; is a SLURM tool that can be used to obtain instantaneous information on resource usage, CPU load, memory, etc. To use it start by changing your SLURM script so that your programme is launched with the &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; command (this should be the last line in the script):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note down the job number. During execution you can then use &amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt;, passing the job number with the &amp;lt;code&amp;gt;-j&amp;lt;/code&amp;gt; flag:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat --format=AveCPU,AveRSS,MaxRSS -P -j 987654&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; can provide information on many different variables, for more details check [https://slurm.schedmd.com/sstat.html the manual].&lt;br /&gt;
&lt;br /&gt;
=== Using top ===&lt;br /&gt;
&lt;br /&gt;
Another way it to log the output of the [http://man7.org/linux/man-pages/man1/top.1.html &amp;lt;code&amp;gt;top&amp;lt;/code&amp;gt;] command. This requires adding an extra command to your SLURM script, again using &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; (where &amp;quot;user001&amp;quot; should be replaced by your user name): &lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun --overcommit --ntasks=1 top -b -u user001 &amp;amp;&lt;br /&gt;
python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This will log the output of top to the sdtout stream file specific in the SLURM script every 3 seconds. Using the tail command you will be able to see logs like:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;top - 18:09:12 up 53 days, 22:53,  0 users,  load average: 27,04, 27,63, 26,53&lt;br /&gt;
Tasks: 1068 total,   4 running, 1064 sleeping,   0 stopped,   0 zombie&lt;br /&gt;
%Cpu(s): 41,9 us,  0,2 sy,  0,0 ni, 57,9 id,  0,0 wa,  0,0 hi,  0,0 si,  0,0 st&lt;br /&gt;
KiB Mem : 10439453+total, 18650486+free, 43690281+used, 42053763+buff/cache&lt;br /&gt;
KiB Swap: 26214400+total, 26214393+free,       52 used. 54182425+avail Mem &lt;br /&gt;
&lt;br /&gt;
  PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND&lt;br /&gt;
29260 user001   20   0   15,6g  15,2g  17392 R  1777  1,5   4014:16 R&lt;br /&gt;
29447 user001   20   0  178604  12900   1676 R   1,6  0,0   6:33.17 top&lt;br /&gt;
28627 user001   20   0  113184   1520   1260 S   0,0  0,0   0:00.01 slurm_scri+&lt;br /&gt;
29253 user001   20   0  245096   4792   1976 S   0,0  0,0   0:00.66 srun&lt;br /&gt;
29357 user001   20   0   36124    688     16 S   0,0  0,0   0:00.00 srun&amp;lt;/pre&amp;gt;&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1900</id>
		<title>Monitoring job execution</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1900"/>
		<updated>2018-08-27T13:22:52Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Output stream redirection */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Output stream redirection ==&lt;br /&gt;
&lt;br /&gt;
The primary way to monitor job execution is through the stdout and stderr streams. These are redirected to [https://wiki.hpcagrogenomics.wur.nl/index.php/Creating_sbatch_script#output_.28stderr.2Cstdout.29_directed_to_file text files specified in the SLURM script].&lt;br /&gt;
&lt;br /&gt;
For this purpose the &amp;lt;code&amp;gt;tail&amp;lt;/code&amp;gt; command is particularly useful. To continuously follow the output to a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -f output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To obtain the last X lines of a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -n X output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Replacing X by the desired number of lines.&lt;br /&gt;
&lt;br /&gt;
If the output file gets too long and you wish to read from the begining you may combine the commands cat and less:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cat output_987654.txt | less&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Use the Q key to exit less.&lt;br /&gt;
&lt;br /&gt;
== Monitoring resource usage ==&lt;br /&gt;
&lt;br /&gt;
While the output streams may suffice in most cases, certain programmes might not provide much feedback. This could be the case with a programme that rellies on modules that are not verbose. In such situations it is best to monitor resource usage to gauge job execution. Two possible options are described below.&lt;br /&gt;
&lt;br /&gt;
=== Using sstat ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; is a SLURM tool that can be used to obtain instantaneous information on resource usage, CPU load, memory, etc. To use it starting by changing your SLURM script so that your programme starts with the &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; command (this should be the last line in the script):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note down the job number. During execution you can then use &amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt;, passing the job number with the &amp;lt;code&amp;gt;-j&amp;lt;/code&amp;gt; flag:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat --format=AveCPU,AveRSS,MaxRSS -P -j 987654&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; can provide information on many different variables, for more details check [https://slurm.schedmd.com/sstat.html the manual].&lt;br /&gt;
&lt;br /&gt;
=== Using top ===&lt;br /&gt;
&lt;br /&gt;
Another way it to log the output of the [http://man7.org/linux/man-pages/man1/top.1.html &amp;lt;code&amp;gt;top&amp;lt;/code&amp;gt;] command. This requires adding an extra command to your SLURM script, again using &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; (where &amp;quot;user001&amp;quot; should be replaced by your user name): &lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun --overcommit --ntasks=1 top -b -u user001 &amp;amp;&lt;br /&gt;
python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This will log the output of top to the sdtout stream file specific in the SLURM script every 3 seconds. Using the tail command you will be able to see logs like:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;top - 18:09:12 up 53 days, 22:53,  0 users,  load average: 27,04, 27,63, 26,53&lt;br /&gt;
Tasks: 1068 total,   4 running, 1064 sleeping,   0 stopped,   0 zombie&lt;br /&gt;
%Cpu(s): 41,9 us,  0,2 sy,  0,0 ni, 57,9 id,  0,0 wa,  0,0 hi,  0,0 si,  0,0 st&lt;br /&gt;
KiB Mem : 10439453+total, 18650486+free, 43690281+used, 42053763+buff/cache&lt;br /&gt;
KiB Swap: 26214400+total, 26214393+free,       52 used. 54182425+avail Mem &lt;br /&gt;
&lt;br /&gt;
  PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND&lt;br /&gt;
29260 user001   20   0   15,6g  15,2g  17392 R  1777  1,5   4014:16 R&lt;br /&gt;
29447 user001   20   0  178604  12900   1676 R   1,6  0,0   6:33.17 top&lt;br /&gt;
28627 user001   20   0  113184   1520   1260 S   0,0  0,0   0:00.01 slurm_scri+&lt;br /&gt;
29253 user001   20   0  245096   4792   1976 S   0,0  0,0   0:00.66 srun&lt;br /&gt;
29357 user001   20   0   36124    688     16 S   0,0  0,0   0:00.00 srun&amp;lt;/pre&amp;gt;&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_executions&amp;diff=1899</id>
		<title>Monitoring executions</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_executions&amp;diff=1899"/>
		<updated>2018-08-23T09:38:24Z</updated>

		<summary type="html">&lt;p&gt;Duque004: Duque004 moved page Monitoring executions to Monitoring job execution&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Monitoring job execution]]&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1898</id>
		<title>Monitoring job execution</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1898"/>
		<updated>2018-08-23T09:38:24Z</updated>

		<summary type="html">&lt;p&gt;Duque004: Duque004 moved page Monitoring executions to Monitoring job execution&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Output stream redirection ==&lt;br /&gt;
&lt;br /&gt;
The primary way to monitor job execution is through the stdout and stderr streams. These are redirected to [https://wiki.hpcagrogenomics.wur.nl/index.php/Creating_sbatch_script#output_.28stderr.2Cstdout.29_directed_to_file text files specified in the SLURM script].&lt;br /&gt;
&lt;br /&gt;
For this purpose the &amp;lt;code&amp;gt;tail&amp;lt;/code&amp;gt; command is particularly useful. To continuous follow the output to a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -f output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To obtain the last X lines of a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -n X output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Replacing X by the desired number of lines.&lt;br /&gt;
&lt;br /&gt;
If the output file gets too long and you wish to read from the begining you may combine the commands cat and less:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cat output_987654.txt | less&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Use the Q key to exit less.&lt;br /&gt;
&lt;br /&gt;
== Monitoring resource usage ==&lt;br /&gt;
&lt;br /&gt;
While the output streams may suffice in most cases, certain programmes might not provide much feedback. This could be the case with a programme that rellies on modules that are not verbose. In such situations it is best to monitor resource usage to gauge job execution. Two possible options are described below.&lt;br /&gt;
&lt;br /&gt;
=== Using sstat ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; is a SLURM tool that can be used to obtain instantaneous information on resource usage, CPU load, memory, etc. To use it starting by changing your SLURM script so that your programme starts with the &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; command (this should be the last line in the script):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note down the job number. During execution you can then use &amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt;, passing the job number with the &amp;lt;code&amp;gt;-j&amp;lt;/code&amp;gt; flag:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat --format=AveCPU,AveRSS,MaxRSS -P -j 987654&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; can provide information on many different variables, for more details check [https://slurm.schedmd.com/sstat.html the manual].&lt;br /&gt;
&lt;br /&gt;
=== Using top ===&lt;br /&gt;
&lt;br /&gt;
Another way it to log the output of the [http://man7.org/linux/man-pages/man1/top.1.html &amp;lt;code&amp;gt;top&amp;lt;/code&amp;gt;] command. This requires adding an extra command to your SLURM script, again using &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; (where &amp;quot;user001&amp;quot; should be replaced by your user name): &lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun --overcommit --ntasks=1 top -b -u user001 &amp;amp;&lt;br /&gt;
python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This will log the output of top to the sdtout stream file specific in the SLURM script every 3 seconds. Using the tail command you will be able to see logs like:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;top - 18:09:12 up 53 days, 22:53,  0 users,  load average: 27,04, 27,63, 26,53&lt;br /&gt;
Tasks: 1068 total,   4 running, 1064 sleeping,   0 stopped,   0 zombie&lt;br /&gt;
%Cpu(s): 41,9 us,  0,2 sy,  0,0 ni, 57,9 id,  0,0 wa,  0,0 hi,  0,0 si,  0,0 st&lt;br /&gt;
KiB Mem : 10439453+total, 18650486+free, 43690281+used, 42053763+buff/cache&lt;br /&gt;
KiB Swap: 26214400+total, 26214393+free,       52 used. 54182425+avail Mem &lt;br /&gt;
&lt;br /&gt;
  PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND&lt;br /&gt;
29260 user001   20   0   15,6g  15,2g  17392 R  1777  1,5   4014:16 R&lt;br /&gt;
29447 user001   20   0  178604  12900   1676 R   1,6  0,0   6:33.17 top&lt;br /&gt;
28627 user001   20   0  113184   1520   1260 S   0,0  0,0   0:00.01 slurm_scri+&lt;br /&gt;
29253 user001   20   0  245096   4792   1976 S   0,0  0,0   0:00.66 srun&lt;br /&gt;
29357 user001   20   0   36124    688     16 S   0,0  0,0   0:00.00 srun&amp;lt;/pre&amp;gt;&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1897</id>
		<title>Monitoring job execution</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1897"/>
		<updated>2018-08-23T09:37:17Z</updated>

		<summary type="html">&lt;p&gt;Duque004: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Output stream redirection ==&lt;br /&gt;
&lt;br /&gt;
The primary way to monitor job execution is through the stdout and stderr streams. These are redirected to [https://wiki.hpcagrogenomics.wur.nl/index.php/Creating_sbatch_script#output_.28stderr.2Cstdout.29_directed_to_file text files specified in the SLURM script].&lt;br /&gt;
&lt;br /&gt;
For this purpose the &amp;lt;code&amp;gt;tail&amp;lt;/code&amp;gt; command is particularly useful. To continuous follow the output to a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -f output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To obtain the last X lines of a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -n X output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Replacing X by the desired number of lines.&lt;br /&gt;
&lt;br /&gt;
If the output file gets too long and you wish to read from the begining you may combine the commands cat and less:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cat output_987654.txt | less&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Use the Q key to exit less.&lt;br /&gt;
&lt;br /&gt;
== Monitoring resource usage ==&lt;br /&gt;
&lt;br /&gt;
While the output streams may suffice in most cases, certain programmes might not provide much feedback. This could be the case with a programme that rellies on modules that are not verbose. In such situations it is best to monitor resource usage to gauge job execution. Two possible options are described below.&lt;br /&gt;
&lt;br /&gt;
=== Using sstat ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; is a SLURM tool that can be used to obtain instantaneous information on resource usage, CPU load, memory, etc. To use it starting by changing your SLURM script so that your programme starts with the &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; command (this should be the last line in the script):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note down the job number. During execution you can then use &amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt;, passing the job number with the &amp;lt;code&amp;gt;-j&amp;lt;/code&amp;gt; flag:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat --format=AveCPU,AveRSS,MaxRSS -P -j 987654&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; can provide information on many different variables, for more details check [https://slurm.schedmd.com/sstat.html the manual].&lt;br /&gt;
&lt;br /&gt;
=== Using top ===&lt;br /&gt;
&lt;br /&gt;
Another way it to log the output of the [http://man7.org/linux/man-pages/man1/top.1.html &amp;lt;code&amp;gt;top&amp;lt;/code&amp;gt;] command. This requires adding an extra command to your SLURM script, again using &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; (where &amp;quot;user001&amp;quot; should be replaced by your user name): &lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun --overcommit --ntasks=1 top -b -u user001 &amp;amp;&lt;br /&gt;
python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This will log the output of top to the sdtout stream file specific in the SLURM script every 3 seconds. Using the tail command you will be able to see logs like:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;top - 18:09:12 up 53 days, 22:53,  0 users,  load average: 27,04, 27,63, 26,53&lt;br /&gt;
Tasks: 1068 total,   4 running, 1064 sleeping,   0 stopped,   0 zombie&lt;br /&gt;
%Cpu(s): 41,9 us,  0,2 sy,  0,0 ni, 57,9 id,  0,0 wa,  0,0 hi,  0,0 si,  0,0 st&lt;br /&gt;
KiB Mem : 10439453+total, 18650486+free, 43690281+used, 42053763+buff/cache&lt;br /&gt;
KiB Swap: 26214400+total, 26214393+free,       52 used. 54182425+avail Mem &lt;br /&gt;
&lt;br /&gt;
  PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND&lt;br /&gt;
29260 user001   20   0   15,6g  15,2g  17392 R  1777  1,5   4014:16 R&lt;br /&gt;
29447 user001   20   0  178604  12900   1676 R   1,6  0,0   6:33.17 top&lt;br /&gt;
28627 user001   20   0  113184   1520   1260 S   0,0  0,0   0:00.01 slurm_scri+&lt;br /&gt;
29253 user001   20   0  245096   4792   1976 S   0,0  0,0   0:00.66 srun&lt;br /&gt;
29357 user001   20   0   36124    688     16 S   0,0  0,0   0:00.00 srun&amp;lt;/pre&amp;gt;&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1896</id>
		<title>Monitoring job execution</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1896"/>
		<updated>2018-08-23T09:36:34Z</updated>

		<summary type="html">&lt;p&gt;Duque004: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Output stream redirection ==&lt;br /&gt;
&lt;br /&gt;
The primary way to monitor job execution is through the stdout and stderr streams. These are redirected to [https://wiki.hpcagrogenomics.wur.nl/index.php/Creating_sbatch_script#output_.28stderr.2Cstdout.29_directed_to_file text files specified in the SLURM script].&lt;br /&gt;
&lt;br /&gt;
For this purpose the &amp;lt;code&amp;gt;tail command is particularly useful. To continuous follow the output to a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -f output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To obtain the last X lines of a text file use:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;tail -n X output_987654.txt&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Replacing X by the desired number of lines.&lt;br /&gt;
&lt;br /&gt;
If the output file gets too long and you wish to read from the begining you may combine the commands cat and less:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;cat output_987654.txt | less&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Use the Q key to exit less.&lt;br /&gt;
&lt;br /&gt;
== Monitoring resource usage ==&lt;br /&gt;
&lt;br /&gt;
While the output streams may suffice in most cases, certain programmes might not provide much feedback. This could be the case with a programme that rellies on modules that are not verbose. In such situations it is best to monitor resource usage to gauge job execution. Two possible options are described below.&lt;br /&gt;
&lt;br /&gt;
=== Using sstat ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; is a SLURM tool that can be used to obtain instantaneous information on resource usage, CPU load, memory, etc. To use it starting by changing your SLURM script so that your programme starts with the &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; command (this should be the last line in the script):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note down the job number. During execution you can then use &amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt;, passing the job number with the &amp;lt;code&amp;gt;-j&amp;lt;/code&amp;gt; flag:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat --format=AveCPU,AveRSS,MaxRSS -P -j 987654&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;sstat&amp;lt;/code&amp;gt; can provide information on many different variables, for more details check [https://slurm.schedmd.com/sstat.html the manual].&lt;br /&gt;
&lt;br /&gt;
=== Using top ===&lt;br /&gt;
&lt;br /&gt;
Another way it to log the output of the [http://man7.org/linux/man-pages/man1/top.1.html &amp;lt;code&amp;gt;top&amp;lt;/code&amp;gt;] command. This requires adding an extra command to your SLURM script, again using &amp;lt;code&amp;gt;srun&amp;lt;/code&amp;gt; (where &amp;quot;user001&amp;quot; should be replaced by your user name): &lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;srun --overcommit --ntasks=1 top -b -u user001 &amp;amp;&lt;br /&gt;
python3 calc_pi.py&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This will log the output of top to the sdtout stream file specific in the SLURM script every 3 seconds. Using the tail command you will be able to see logs like:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;top - 18:09:12 up 53 days, 22:53,  0 users,  load average: 27,04, 27,63, 26,53&lt;br /&gt;
Tasks: 1068 total,   4 running, 1064 sleeping,   0 stopped,   0 zombie&lt;br /&gt;
%Cpu(s): 41,9 us,  0,2 sy,  0,0 ni, 57,9 id,  0,0 wa,  0,0 hi,  0,0 si,  0,0 st&lt;br /&gt;
KiB Mem : 10439453+total, 18650486+free, 43690281+used, 42053763+buff/cache&lt;br /&gt;
KiB Swap: 26214400+total, 26214393+free,       52 used. 54182425+avail Mem &lt;br /&gt;
&lt;br /&gt;
  PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND&lt;br /&gt;
29260 user001   20   0   15,6g  15,2g  17392 R  1777  1,5   4014:16 R&lt;br /&gt;
29447 user001   20   0  178604  12900   1676 R   1,6  0,0   6:33.17 top&lt;br /&gt;
28627 user001   20   0  113184   1520   1260 S   0,0  0,0   0:00.01 slurm_scri+&lt;br /&gt;
29253 user001   20   0  245096   4792   1976 S   0,0  0,0   0:00.66 srun&lt;br /&gt;
29357 user001   20   0   36124    688     16 S   0,0  0,0   0:00.00 srun&amp;lt;/code&amp;gt;&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1895</id>
		<title>Monitoring job execution</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1895"/>
		<updated>2018-08-23T09:32:26Z</updated>

		<summary type="html">&lt;p&gt;Duque004: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Output stream redirection ==&lt;br /&gt;
&lt;br /&gt;
The primary way to monitor job execution is through the stdout and stderr streams. These are redirected to [https://wiki.hpcagrogenomics.wur.nl/index.php/Creating_sbatch_script#output_.28stderr.2Cstdout.29_directed_to_file text files specified in the SLURM script].&lt;br /&gt;
&lt;br /&gt;
For this purpose the `tail`command is particularly useful. To continuous follow the output to a text file use:&lt;br /&gt;
&lt;br /&gt;
tail -f output_987654.txt&lt;br /&gt;
&lt;br /&gt;
To obtain the last X lines of a text file use:&lt;br /&gt;
&lt;br /&gt;
tail -n X output_987654.txt&lt;br /&gt;
&lt;br /&gt;
Replacing X by the desired number of lines.&lt;br /&gt;
&lt;br /&gt;
If the output file gets too long and you wish to read from the begining you may combine the commands cat and less:&lt;br /&gt;
&lt;br /&gt;
cat output_987654.txt | less&lt;br /&gt;
&lt;br /&gt;
Use the Q key to exit less.&lt;br /&gt;
&lt;br /&gt;
== Monitoring resource usage ==&lt;br /&gt;
&lt;br /&gt;
While the output streams may suffice in most cases, certain programmes might not provide much feedback. This could be the case with a programme that rellies on modules that are not verbose. In such situations it is best to monitor resource usage to gauge job execution. Two possible options are described below.&lt;br /&gt;
&lt;br /&gt;
=== Using sstat ===&lt;br /&gt;
&lt;br /&gt;
sstat is a SLURM tool that can be used to obtain instantaneous information on resource usage, CPU load, memory, etc. To use it starting by changing your SLURM script so that your programme starts with the srun command (this should be the last line in the script):&lt;br /&gt;
&lt;br /&gt;
srun python3 calc_pi.py&lt;br /&gt;
&lt;br /&gt;
Note down the job number. During execution you can then use sstat, passing the job number with the -j flag:&lt;br /&gt;
&lt;br /&gt;
sstat --format=AveCPU,AveRSS,MaxRSS -P -j 987654&lt;br /&gt;
&lt;br /&gt;
sstat can provide information on many different variables, for more details check [https://slurm.schedmd.com/sstat.html the manual].&lt;br /&gt;
&lt;br /&gt;
=== Using top ===&lt;br /&gt;
&lt;br /&gt;
Another way it to log the output of the [http://man7.org/linux/man-pages/man1/top.1.html top] command. This requires adding an extra command to your SLURM script, again using srun (where &amp;quot;user001&amp;quot; should be replaced by your user name): &lt;br /&gt;
&lt;br /&gt;
srun --overcommit --ntasks=1 top -b -u user001 &amp;amp;&lt;br /&gt;
python3 calc_pi.py&lt;br /&gt;
&lt;br /&gt;
This will log the output of top to the sdtout stream file specific in the SLURM script every 3 seconds. Using the tail command you will be able to see logs like:&lt;br /&gt;
&lt;br /&gt;
top - 18:09:12 up 53 days, 22:53,  0 users,  load average: 27,04, 27,63, 26,53&lt;br /&gt;
Tasks: 1068 total,   4 running, 1064 sleeping,   0 stopped,   0 zombie&lt;br /&gt;
%Cpu(s): 41,9 us,  0,2 sy,  0,0 ni, 57,9 id,  0,0 wa,  0,0 hi,  0,0 si,  0,0 st&lt;br /&gt;
KiB Mem : 10439453+total, 18650486+free, 43690281+used, 42053763+buff/cache&lt;br /&gt;
KiB Swap: 26214400+total, 26214393+free,       52 used. 54182425+avail Mem &lt;br /&gt;
&lt;br /&gt;
  PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND&lt;br /&gt;
29260 user001   20   0   15,6g  15,2g  17392 R  1777  1,5   4014:16 R&lt;br /&gt;
29447 user001   20   0  178604  12900   1676 R   1,6  0,0   6:33.17 top&lt;br /&gt;
28627 user001   20   0  113184   1520   1260 S   0,0  0,0   0:00.01 slurm_scri+&lt;br /&gt;
29253 user001   20   0  245096   4792   1976 S   0,0  0,0   0:00.66 srun&lt;br /&gt;
29357 user001   20   0   36124    688     16 S   0,0  0,0   0:00.00 srun&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1894</id>
		<title>Monitoring job execution</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1894"/>
		<updated>2018-08-23T09:13:02Z</updated>

		<summary type="html">&lt;p&gt;Duque004: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Output stream redirection ==&lt;br /&gt;
&lt;br /&gt;
The primary way to monitor job execution is through the stdout and stderr streams. These are redirected to [[https://wiki.hpcagrogenomics.wur.nl/index.php/Creating_sbatch_script#output_.28stderr.2Cstdout.29_directed_to_file | text files specified in the SLURM script]].&lt;br /&gt;
&lt;br /&gt;
For this purpose the `tail`command is particularly useful. To continuous follow the output to a text file use:&lt;br /&gt;
&lt;br /&gt;
tail -f output_987654.txt&lt;br /&gt;
&lt;br /&gt;
To obtain the last X lines of a text file use:&lt;br /&gt;
&lt;br /&gt;
tail -n X output_987654.txt&lt;br /&gt;
&lt;br /&gt;
Replacing X by the desired number of lines.&lt;br /&gt;
&lt;br /&gt;
If the output file gets too long and you wish to read from the begining you may combine the commands cat and less:&lt;br /&gt;
&lt;br /&gt;
cat output_987654.txt | less&lt;br /&gt;
&lt;br /&gt;
Use the Q key to exit less.&lt;br /&gt;
&lt;br /&gt;
== Monitoring resource usage ==&lt;br /&gt;
&lt;br /&gt;
While the output streams may suffice in most cases, certain programmes might not provide much feedback. This could be the case with a programme that rellies on modules that are not verbose. In such situations it is best to monitor resource usage to gauge job execution. Two possible options are described below.&lt;br /&gt;
&lt;br /&gt;
=== Using sstat ===&lt;br /&gt;
&lt;br /&gt;
sstat is a SLURM tool that can be used to obtain instantaneous information on resource usage, CPU load, memory, etc. To use it starting by changing your SLURM script so that your programme starts with the srun command (this should be the last line in the script):&lt;br /&gt;
&lt;br /&gt;
srun python3 calc_pi.py&lt;br /&gt;
&lt;br /&gt;
Note down the job number. During execution you can then use sstat, passing the job number with the -j flag:&lt;br /&gt;
&lt;br /&gt;
sstat --format=AveCPU,AveRSS,MaxRSS -P -j 987654&lt;br /&gt;
&lt;br /&gt;
sstat can provide information on many different variables, for more details check [[https://slurm.schedmd.com/sstat.html | the manual]].&lt;br /&gt;
&lt;br /&gt;
=== Logging the output of top ===&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1893</id>
		<title>Monitoring job execution</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Monitoring_job_execution&amp;diff=1893"/>
		<updated>2018-08-23T09:07:19Z</updated>

		<summary type="html">&lt;p&gt;Duque004: Created page with &amp;quot; == Output stream redirection ==  The primary way to monitor job execution is through the stdout and stderr streams. These are redirected to https://wiki.hpcagrogenomics.wur...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== Output stream redirection ==&lt;br /&gt;
&lt;br /&gt;
The primary way to monitor job execution is through the stdout and stderr streams. These are redirected to [[https://wiki.hpcagrogenomics.wur.nl/index.php/Creating_sbatch_script#output_.28stderr.2Cstdout.29_directed_to_file | text files specified in the SLURM script]].&lt;br /&gt;
&lt;br /&gt;
For this purpose the `tail`command is particularly useful. To continuous follow the output to a text file use:&lt;br /&gt;
&lt;br /&gt;
tail -f output_987654.txt&lt;br /&gt;
&lt;br /&gt;
To obtain the last X lines of a text file use:&lt;br /&gt;
&lt;br /&gt;
tail -n X output_987654.txt&lt;br /&gt;
&lt;br /&gt;
Replacing X by the desired number of lines.&lt;br /&gt;
&lt;br /&gt;
If the output file gets too long and you wish to read from the begining you may combine the commands cat and less:&lt;br /&gt;
&lt;br /&gt;
cat output_987654.txt | less&lt;br /&gt;
&lt;br /&gt;
Use the Q key to exit less.&lt;br /&gt;
&lt;br /&gt;
== Monitoring resource usage ==&lt;br /&gt;
&lt;br /&gt;
While the output streams may suffice in most cases, certain programmes might not provide much feedback. This could be the case with a programme that rellies on modules that are not verbose. In such situations it is best to monitor resource usage to gauge job execution. Two possible options are described below.&lt;br /&gt;
&lt;br /&gt;
=== Using sstat ===&lt;br /&gt;
&lt;br /&gt;
sstat is a SLURM tool that can be used to obtain instantaneous &lt;br /&gt;
&lt;br /&gt;
srun python3 calc_pi.py&lt;br /&gt;
&lt;br /&gt;
sstat --format=AveCPU,AveRSS,MaxRSS -P -j 7466208&lt;br /&gt;
&lt;br /&gt;
=== Logging the output of top ===&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Main_Page&amp;diff=1892</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Main_Page&amp;diff=1892"/>
		<updated>2018-08-23T08:40:03Z</updated>

		<summary type="html">&lt;p&gt;Duque004: Added link to monitor execution page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The Agrogenomics cluster is a [http://en.wikipedia.org/wiki/High-performance_computing High Performance Computer] (HPC) infrastructure hosted by [http://www.wageningenur.nl/nl/activiteit/Opening-High-Performance-Computing-cluster-HPC.htm Wageningen University &amp;amp; Research Centre]. It is open for use for all WUR research groups as well as other organizations, including companies, that have collaborative projects with WUR. &lt;br /&gt;
&lt;br /&gt;
The Agrogenomics HPC was an initiative of the [http://www.breed4food.com/en/breed4food.htm Breed4Food] (B4F) consortium, consisting of the [[About_ABGC | Animal Breeding and Genomics Centre]] (WU-Animal Breeding and Genomics and Wageningen Livestock Research) and four major breeding companies: [http://www.cobb-vantress.com Cobb-Vantress], [https://www.crv4all.nl CRV], [http://www.hendrix-genetics.com Hendrix Genetics], and [http://www.topigs.com TOPIGS]. Currently, in addition to the original partners, the HPC (HPC-Ag) is used by other groups from Wageningen UR (Bioinformatics, Centre for Crop Systems Analysis, Environmental Sciences Group, and Plant Research International) and plant breeding industry (Rijk Zwaan). &lt;br /&gt;
&lt;br /&gt;
= Using the HPC-Ag =&lt;br /&gt;
== Gaining access to the HPC-Ag ==&lt;br /&gt;
Access to the cluster and file transfer are traditionally done via [http://en.wikipedia.org/wiki/Secure_Shell SSH and SFTP].&lt;br /&gt;
* [[log_in_to_B4F_cluster | Logging into cluster using ssh and file transfer]]&lt;br /&gt;
* [[Services | Alternative access methods, and extra features and services on the HPC]]&lt;br /&gt;
* [[Filesystems | Accessible storage methods on the HPC]]&lt;br /&gt;
&lt;br /&gt;
== Access Policy ==&lt;br /&gt;
[[Access_Policy | Main Article: Access Policy]]&lt;br /&gt;
&lt;br /&gt;
Access needs to be granted actively (by creation of an account on the cluster by FB-IT). Use of resources is limited by the scheduler. Depending on availability of queues (&#039;partitions&#039;) granted to a user, priority to the system&#039;s resources is regulated. Note that the use of the HPC-Ag is not free of charge. List price of CPU time and storage, and possible discounts on that list price for your organisation, can be retrieved from CAT-AGRO or FB-ICT.&lt;br /&gt;
&lt;br /&gt;
= Events =&lt;br /&gt;
* [[Courses]] that have happened and are happening&lt;br /&gt;
* [[Downtime]] that will affect all users&lt;br /&gt;
* [[Meetings]] that may affect the policies of the HPC&lt;br /&gt;
&lt;br /&gt;
= Other Software =&lt;br /&gt;
&lt;br /&gt;
== Cluster Management Software and Scheduler ==&lt;br /&gt;
The HPC-Ag uses Bright Cluster Manager software for overall cluster management, and Slurm as job scheduler.&lt;br /&gt;
* [[BCM_on_B4F_cluster | Monitor cluster status with BCM]]&lt;br /&gt;
* [[SLURM_on_B4F_cluster | Submit jobs with Slurm]]&lt;br /&gt;
* [[node_usage_graph | Be aware of how much work the cluster is under right now with &#039;node_usage_graph&#039;]]&lt;br /&gt;
* [[SLURM_Compare | Rosetta Stone of Workload Managers]]&lt;br /&gt;
&lt;br /&gt;
== Installation of software by users ==&lt;br /&gt;
&lt;br /&gt;
* [[Domain_specific_software_on_B4Fcluster_installation_by_users | Installing domain specific software: installation by users]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
* [[Virtual_environment_Python_3.4_or_higher | Setting up and using a virtual environment for Python3.4 or higher ]]&lt;br /&gt;
* [[Installing WRF and WPS]]&lt;br /&gt;
&lt;br /&gt;
== Installed software ==&lt;br /&gt;
&lt;br /&gt;
* [[Globally_installed_software | Globally installed software]]&lt;br /&gt;
* [[ABGC_modules | ABGC specific modules]]&lt;br /&gt;
&lt;br /&gt;
= Useful Notes = &lt;br /&gt;
&lt;br /&gt;
== Being in control of Environment parameters ==&lt;br /&gt;
&lt;br /&gt;
* [[Using_environment_modules | Using environment modules]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Setting_TMPDIR | Set a custom temporary directory location]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
&lt;br /&gt;
== Controlling costs ==&lt;br /&gt;
&lt;br /&gt;
* [[SACCT | using SACCT to see your costs]]&lt;br /&gt;
* [[get_my_bill | using the &amp;quot;get_my_bill&amp;quot; script to estimate costs]]&lt;br /&gt;
&lt;br /&gt;
== Management ==&lt;br /&gt;
[[HPC_management | Main Article: HPC management]]&lt;br /&gt;
&lt;br /&gt;
Project Leader of the HPC is Stephen Janssen (Wageningen UR,FB-IT, Service Management). [[User:lith010 | Jan van Lith (Wageningen UR,FB-IT, Infrastructure)]] and [[User:dawes001 | Gwen Dawes (Wageningen UR, FB-IT, Infrastructure)]] are responsible for [[Maintenance_and_Management | Maintenance and Management]].&lt;br /&gt;
&lt;br /&gt;
= Miscellaneous =&lt;br /&gt;
* [[Mailinglist | Electronic mail discussion lists]]&lt;br /&gt;
* [[History_of_the_Cluster | Historical information on the startup of the HPC]]&lt;br /&gt;
* [[Bioinformatics_tips_tricks_workflows | Bioinformatics tips, tricks, and workflows]]&lt;br /&gt;
* [[Parallel_R_code_on_SLURM | Running parallel R code on SLURM]]&lt;br /&gt;
* [[Convert_between_MediaWiki_and_other_formats | Convert between MediaWiki format and other formats]]&lt;br /&gt;
* [[Manual GitLab | GitLab: Create projects and add scripts]]&lt;br /&gt;
* [[Monitoring_executions | Monitoring job execution]]&lt;br /&gt;
&lt;br /&gt;
= See also =&lt;br /&gt;
* [[Maintenance_and_Management | Maintenance and Management]]&lt;br /&gt;
* [[BCData | BCData]]&lt;br /&gt;
* [[Mailinglist | Electronic mail discussion lists]]&lt;br /&gt;
* [[About_ABGC | About ABGC]]&lt;br /&gt;
* [[Computer_cluster | High Performance Computing @ABGC]]&lt;br /&gt;
* [[Lustre_PFS_layout | Lustre Parallel File System layout]]&lt;br /&gt;
&lt;br /&gt;
= External links =&lt;br /&gt;
{| width=&amp;quot;90%&amp;quot;&lt;br /&gt;
|- valign=&amp;quot;top&amp;quot;&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [http://www.breed4food.com/en/show/Breed4Food-initiative-reinforces-the-Netherlands-position-as-an-innovative-country-in-animal-breeding-and-genomics.htm Breed4Food programme]&lt;br /&gt;
* [http://www.wageningenur.nl/en/Expertise-Services/Facilities/CATAgroFood-3/CATAgroFood-3/Our-facilities/Show/High-Performance-Computing-Cluster-HPC.htm CATAgroFood offers a HPC facilty]&lt;br /&gt;
* [http://www.cobb-vantress.com Cobb-Vantress homepage]&lt;br /&gt;
&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [https://www.crv4all.nl CRV homepage]&lt;br /&gt;
* [http://www.hendrix-genetics.com Hendrix Genetics homepage]&lt;br /&gt;
* [http://www.topigs.com TOPIGS homepage]&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Scientific_Linux Scientific Linux]&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Help:Cheatsheet Help with editing Wiki pages]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=SLURM_Compare&amp;diff=1886</id>
		<title>SLURM Compare</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=SLURM_Compare&amp;diff=1886"/>
		<updated>2018-08-01T13:45:35Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Rosetta Stone of Workload Managers */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=== Rosetta Stone of Workload Managers ===&lt;br /&gt;
&lt;br /&gt;
PBS/Torque, Slurm, LSF, SGE and LoadLeveler [http://slurm.schedmd.com/rosetta.html Rosetta Stone]&lt;br /&gt;
&lt;br /&gt;
This table lists the most common command, environment variables, and job specification options used by the major workload management systems: PBS/Torque, Slurm, LSF, SGE and LoadLeveler. Each of these workload managers has unique features, but the most commonly used functionality is available in all of these environments as listed in the table. This should be considered a work in progress and contributions to improve the document are welcome.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!User Commands!!PBS/Torque!!Slurm!!LSF!!SGE!!LoadLeveler&lt;br /&gt;
|-&lt;br /&gt;
||Job submission|| qsub [script_file]|| sbatch [script_file]|| bsub [script_file]|| qsub [script_file]|| llsubmit [script_file] &lt;br /&gt;
|-&lt;br /&gt;
||Job deletion ||qdel [job_id]|| scancel [job_id]|| bkill [job_id]|| qdel [job_id]|| llcancel [job_id] &lt;br /&gt;
|-&lt;br /&gt;
||Job status (by job)|| qstat [job_id]|| squeue --job [job_id]|| bjobs [job_id]|| qstat -u \* [-j job_id]|| llq -u [username] &lt;br /&gt;
|-&lt;br /&gt;
||Job status (by user)|| qstat -u [user_name]|| squeue -u [user_name]|| bjobs -u [user_name]|| qstat [-u user_name]|| llq -u [user_name] &lt;br /&gt;
|-&lt;br /&gt;
||Job hold ||qhold [job_id]|| scontrol hold [job_id]|| bstop [job_id]|| qhold [job_id]|| llhold -r [job_id] &lt;br /&gt;
|-&lt;br /&gt;
||Job release|| qrls [job_id]|| scontrol release [job_id]|| bresume [job_id]|| qrls [job_id]|| llhold -r [job_id] &lt;br /&gt;
|-&lt;br /&gt;
||Queue list|| qstat -Q|| squeue ||bqueues ||qconf -sql|| llclass &lt;br /&gt;
|-&lt;br /&gt;
||Node list ||pbsnodes -l|| sinfo -N OR scontrol show nodes|| bhosts|| qhost|| llstatus -L machine &lt;br /&gt;
|-&lt;br /&gt;
||Cluster status|| qstat -a|| sinfo|| bqueues|| qhost -q|| llstatus -L cluster &lt;br /&gt;
|-&lt;br /&gt;
||GUI|| xpbsmon|| sview|| xlsf OR xlsbatch|| qmon|| xload &lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
||&#039;&#039;&#039;Environment&#039;&#039;&#039;||&#039;&#039;&#039;PBS/Torque&#039;&#039;&#039;||&#039;&#039;&#039;Slurm&#039;&#039;&#039;||&#039;&#039;&#039;LSF&#039;&#039;&#039;||&#039;&#039;&#039;SGE&#039;&#039;&#039;||&#039;&#039;&#039;LoadLeveler&#039;&#039;&#039; &lt;br /&gt;
|-&lt;br /&gt;
||Job ID|| $PBS_JOBID|| $SLURM_JOBID|| $LSB_JOBID|| $JOB_ID|| $LOAD_STEP_ID &lt;br /&gt;
|-&lt;br /&gt;
||Submit Directory|| $PBS_O_WORKDIR|| $SLURM_SUBMIT_DIR|| $LSB_SUBCWD|| $SGE_O_WORKDIR|| $LOADL_STEP_INITDIR &lt;br /&gt;
|-&lt;br /&gt;
||Submit Host|| $PBS_O_HOST|| $SLURM_SUBMIT_HOST|| $LSB_SUB_HOST|| $SGE_O_HOST ||&lt;br /&gt;
|-&lt;br /&gt;
||Node List|| $PBS_NODEFILE|| $SLURM_JOB_NODELIST|| $LSB_HOSTS/LSB_MCPU_HOST|| $PE_HOSTFILE|| $LOADL_PROCESSOR_LIST &lt;br /&gt;
|-&lt;br /&gt;
||Job Array Index|| $PBS_ARRAYID|| $SLURM_ARRAY_TASK_ID|| $LSB_JOBINDEX|| $SGE_TASK_ID ||&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Using_Slurm&amp;diff=1861</id>
		<title>Using Slurm</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Using_Slurm&amp;diff=1861"/>
		<updated>2018-04-19T13:05:31Z</updated>

		<summary type="html">&lt;p&gt;Duque004: /* Batch script */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The resource allocation / scheduling software on the B4F Cluster is [http://en.wikipedia.org/wiki/Simple_Linux_Utility_for_Resource_Management SLURM]: &#039;&#039;&#039;S&#039;&#039;&#039;imple &#039;&#039;&#039;L&#039;&#039;&#039;inux &#039;&#039;&#039;U&#039;&#039;&#039;tility for &#039;&#039;&#039;R&#039;&#039;&#039;esource &#039;&#039;&#039;M&#039;&#039;&#039;anagement.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Queues and defaults ==&lt;br /&gt;
&lt;br /&gt;
=== Queues ===&lt;br /&gt;
Every organization has 3 queues (in slurm called partitions) : a high, a standard and a low priority queue.&amp;lt;br&amp;gt;&lt;br /&gt;
The High queue provides the highest priority to jobs (20) then the standard queue (10). In the low priority queue (0)&amp;lt;br&amp;gt;&lt;br /&gt;
jobs will be resubmitted if a job with higer priority needs cluster resources and those resoruces are occupied by a Low queue jobs.&lt;br /&gt;
To find out which queues your account has been authorized for, type sinfo:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
PARTITION       AVAIL  TIMELIMIT  NODES  STATE NODELIST&lt;br /&gt;
ABGC_High      up   infinite     12  down* node[043-048,055-060]&lt;br /&gt;
ABGC_High      up   infinite      6    mix fat[001-002],node[002-005]&lt;br /&gt;
ABGC_High      up   infinite     44   idle node[001,006-042,049-054]&lt;br /&gt;
ABGC_Std       up   infinite     12  down* node[043-048,055-060]&lt;br /&gt;
ABGC_Std       up   infinite      6    mix fat[001-002],node[002-005]&lt;br /&gt;
ABGC_Std       up   infinite     44   idle node[001,006-042,049-054]&lt;br /&gt;
ABGC_Low       up   infinite     12  down* node[043-048,055-060]&lt;br /&gt;
ABGC_Low       up   infinite      6    mix fat[001-002],node[002-005]&lt;br /&gt;
ABGC_Low       up   infinite     44   idle node[001,006-042,049-054]&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Defaults ===&lt;br /&gt;
There is no default queue, so you need to specify which queue to use when submitting a job.&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;The default run time for a job is 1 hour!&#039;&#039;&#039; &amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Default memory limit is 100MB per node!&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Submitting jobs: sbatch ==&lt;br /&gt;
&lt;br /&gt;
=== Example ===&lt;br /&gt;
Consider this simple python3 script that should calculate Pi to 1 million digits:&lt;br /&gt;
&amp;lt;source lang=&#039;python&#039;&amp;gt;&lt;br /&gt;
from decimal import *&lt;br /&gt;
D=Decimal&lt;br /&gt;
getcontext().prec=10000000&lt;br /&gt;
p=sum(D(1)/16**k*(D(4)/(8*k+1)-D(2)/(8*k+4)-D(1)/(8*k+5)-D(1)/(8*k+6))for k in range(411))&lt;br /&gt;
print(str(p)[:10000002])&lt;br /&gt;
&amp;lt;/source&amp;gt; &lt;br /&gt;
&lt;br /&gt;
=== Loading modules ===&lt;br /&gt;
In order for this script to run, the first thing that is needed is that Python3, which is not the default Python version on the cluster, is load into your environment. Availability of (different versions of) software can be checked by the following command:&lt;br /&gt;
  module avail&lt;br /&gt;
&lt;br /&gt;
In the list you should note that python3 is indeed available to be loaded, which then can be loaded with the following command:&lt;br /&gt;
  module load python/3.3.3&lt;br /&gt;
&lt;br /&gt;
=== Batch script ===&lt;br /&gt;
[[Creating_sbatch_script | Main Article: Creating a sbatch script]]&lt;br /&gt;
&lt;br /&gt;
The following shell/slurm script can then be used to schedule the job using the sbatch command:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
#!/bin/bash&lt;br /&gt;
#SBATCH --account=773320000&lt;br /&gt;
#SBATCH --time=1200&lt;br /&gt;
#SBATCH --mem=2048&lt;br /&gt;
#SBATCH --ntasks=1&lt;br /&gt;
#SBATCH --output=output_%j.txt&lt;br /&gt;
#SBATCH --error=error_output_%j.txt&lt;br /&gt;
#SBATCH --job-name=calc_pi.py&lt;br /&gt;
#SBATCH --partition=ABGC_Std&lt;br /&gt;
#SBATCH --mail-type=ALL&lt;br /&gt;
#SBATCH --mail-user=email@org.nl&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
time python3 calc_pi.py&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Submitting ===&lt;br /&gt;
The script, assuming it was named &#039;run_calc_pi.sh&#039;, can then be posted using the following command:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
sbatch run_calc_pi.sh&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Submitting multiple jobs (simple) ===&lt;br /&gt;
Assuming there are 10 job scripts, name runscript_1.sh through runscript_10.sh, all these scripts can be submitted using the following line of shell code:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;for i in `seq 1 10`; do echo $i; sbatch runscript_$i.sh;done&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Submitting array jobs ===&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
#SBATCH --array=0-10%4&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
SLURM allows you to submit multiple jobs using the same template. Further information about this can be found [[Array_jobs|here]].&lt;br /&gt;
&lt;br /&gt;
=== Using /tmp ===&lt;br /&gt;
There is a local disk of ~300G that can be used to temporarily stage some of your workload attached to each node. This is free to use, but please remember to clean up your data after usage.&lt;br /&gt;
&lt;br /&gt;
In order to be sure that you&#039;re able to use space in /tmp, you can add&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
#SBATCH --tmp=&amp;lt;required size&amp;gt;&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
To your sbatch script. This will prevent your job from being run on nodes where there is no free space, or it&#039;s aimed to be used by another job at the same time.&lt;br /&gt;
&lt;br /&gt;
== Monitoring submitted jobs ==&lt;br /&gt;
Once a job is submitted, the status can be monitored using the &amp;lt;code&amp;gt;squeue&amp;lt;/code&amp;gt; command. The &amp;lt;code&amp;gt;squeue&amp;lt;/code&amp;gt; command has a number of parameters for monitoring specific properties of the jobs such as time limit.&lt;br /&gt;
&lt;br /&gt;
=== Generic monitoring of all running jobs ===&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
  squeue&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
You should then get a list of jobs that are running at that time on the cluster, for the example on how to submit using the &#039;sbatch&#039; command, it may look like so:&lt;br /&gt;
    JOBID PARTITION     NAME     USER  ST       TIME  NODES NODELIST(REASON)&lt;br /&gt;
   3396      ABGC BOV-WUR- megen002   R      27:26      1 node004&lt;br /&gt;
   3397      ABGC BOV-WUR- megen002   R      27:26      1 node005&lt;br /&gt;
   3398      ABGC BOV-WUR- megen002   R      27:26      1 node006&lt;br /&gt;
   3399      ABGC BOV-WUR- megen002   R      27:26      1 node007&lt;br /&gt;
   3400      ABGC BOV-WUR- megen002   R      27:26      1 node008&lt;br /&gt;
   3401      ABGC BOV-WUR- megen002   R      27:26      1 node009&lt;br /&gt;
   3385  research BOV-WUR- megen002   R      44:38      1 node049&lt;br /&gt;
   3386  research BOV-WUR- megen002   R      44:38      1 node050&lt;br /&gt;
   3387  research BOV-WUR- megen002   R      44:38      1 node051&lt;br /&gt;
   3388  research BOV-WUR- megen002   R      44:38      1 node052&lt;br /&gt;
   3389  research BOV-WUR- megen002   R      44:38      1 node053&lt;br /&gt;
   3390  research BOV-WUR- megen002   R      44:38      1 node054&lt;br /&gt;
   3391  research BOV-WUR- megen002   R      44:38      3 node[049-051]&lt;br /&gt;
   3392  research BOV-WUR- megen002   R      44:38      3 node[052-054]&lt;br /&gt;
   3393  research BOV-WUR- megen002   R      44:38      1 node001&lt;br /&gt;
   3394  research BOV-WUR- megen002   R      44:38      1 node002&lt;br /&gt;
   3395  research BOV-WUR- megen002   R      44:38      1 node003&lt;br /&gt;
&lt;br /&gt;
=== Monitoring time limit set for a specific job ===&lt;br /&gt;
The default time limit is set at one hour. Estimated run times need to be specified when running jobs. To see what the time limit is that is set for a certain job, this can be done using the &amp;lt;code&amp;gt;squeue&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
squeue -l -j 3532&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
Information similar to the following should appear:&lt;br /&gt;
  Fri Nov 29 15:41:00 2013&lt;br /&gt;
   JOBID PARTITION     NAME     USER    STATE       TIME TIMELIMIT  NODES NODELIST(REASON)&lt;br /&gt;
   3532      ABGC BOV-WUR- megen002  RUNNING    2:47:03 3-08:00:00      1 node054&lt;br /&gt;
&lt;br /&gt;
=== Query a specific active job: scontrol ===&lt;br /&gt;
Show all the details of a currently active job, so not a completed job.&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
nfs01 ~]$ scontrol show jobid 4241&lt;br /&gt;
JobId=4241 Name=WB20F06&lt;br /&gt;
   UserId=megen002(16795409) GroupId=domain users(16777729)&lt;br /&gt;
   Priority=1 Account=(null) QOS=normal&lt;br /&gt;
   JobState=RUNNING Reason=None Dependency=(null)&lt;br /&gt;
   Requeue=1 Restarts=0 BatchFlag=1 ExitCode=0:0&lt;br /&gt;
   RunTime=02:55:25 TimeLimit=3-08:00:00 TimeMin=N/A&lt;br /&gt;
   SubmitTime=2013-12-09T13:37:29 EligibleTime=2013-12-09T13:37:29&lt;br /&gt;
   StartTime=2013-12-09T13:37:29 EndTime=2013-12-12T21:37:29&lt;br /&gt;
   PreemptTime=None SuspendTime=None SecsPreSuspend=0&lt;br /&gt;
   Partition=research AllocNode:Sid=nfs01:21799&lt;br /&gt;
   ReqNodeList=(null) ExcNodeList=(null)&lt;br /&gt;
   NodeList=node023&lt;br /&gt;
   BatchHost=node023&lt;br /&gt;
   NumNodes=1 NumCPUs=4 CPUs/Task=1 ReqS:C:T=*:*:*&lt;br /&gt;
   MinCPUsNode=1 MinMemoryNode=0 MinTmpDiskNode=0&lt;br /&gt;
   Features=(null) Gres=(null) Reservation=(null)&lt;br /&gt;
   Shared=OK Contiguous=0 Licenses=(null) Network=(null)&lt;br /&gt;
   Command=/lustre/scratch/WUR/ABGC/...&lt;br /&gt;
   WorkDir=/lustre/scratch/WUR/ABGC/...&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Check on a pending job ===&lt;br /&gt;
A submitted job could result in a pending state when there are not enough resources available to this job.&lt;br /&gt;
In this example I sumbit a job, check the status and after finding out is it &#039;&#039;&#039;pending&#039;&#039;&#039; I&#039;ll check when is probably will start.&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
[@nfs01 jobs]$ sbatch hpl_student.job&lt;br /&gt;
 Submitted batch job 740338&lt;br /&gt;
&lt;br /&gt;
[@nfs01 jobs]$ squeue -l -j 740338&lt;br /&gt;
 Fri Feb 21 15:32:31 2014&lt;br /&gt;
  JOBID PARTITION     NAME     USER    STATE       TIME TIMELIMIT  NODES NODELIST(REASON)&lt;br /&gt;
 740338 ABGC_Stud HPLstude bohme999  PENDING       0:00 1-00:00:00      1 (ReqNodeNotAvail)&lt;br /&gt;
&lt;br /&gt;
[@nfs01 jobs]$ squeue --start -j 740338&lt;br /&gt;
  JOBID PARTITION     NAME     USER  ST           START_TIME  NODES NODELIST(REASON)&lt;br /&gt;
 740338 ABGC_Stud HPLstude bohme999  PD  2014-02-22T15:31:48      1 (ReqNodeNotAvail)&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
So it seems this job will problably start the next day, but&#039;s thats no guarantee it will start indeed.&lt;br /&gt;
&lt;br /&gt;
== Removing jobs from a list: scancel ==&lt;br /&gt;
If for some reason you want to delete a job that is either in the queue or already running, you can remove it using the &#039;scancel&#039; command. The &#039;scancel&#039; command takes the jobid as a parameter. For the example above, this would be done using the following code:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
scancel 3401&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Allocating resources interactively: salloc ==&lt;br /&gt;
It&#039;s possible to set up an interactive session using salloc. Run salloc as follows:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
salloc -p &amp;lt;partition, say, ABGC_Low&amp;gt;&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
And because of the magic of SallocDefaultCommand, you will immediately be transported to a new prompt.&lt;br /&gt;
&lt;br /&gt;
Here, run &#039;hostname&#039; to see which node your shell has been transported to.&lt;br /&gt;
&lt;br /&gt;
If you don&#039;t want your shell to be transported but want a new remote shell, do:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
salloc -p ABGC_Low $SHELL&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
Now your shell will stay on nfs01, but you can do:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
srun &amp;lt;command&amp;gt; &amp;amp;&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
To submit tasks to this new shell!&lt;br /&gt;
&lt;br /&gt;
Be aware that the time limit of salloc is default 1 hour. If you intend to run jobs for longer times than this, you need to edit the settings for it. See: https://computing.llnl.gov/linux/slurm/salloc.html&lt;br /&gt;
&lt;br /&gt;
== Get overview of past and current jobs: sacct ==&lt;br /&gt;
To do some accounting on past and present jobs, and to see whether they ran to completion, you can do:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
sacct&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
This should provide information similar to the following:&lt;br /&gt;
&lt;br /&gt;
         JobID    JobName  Partition    Account  AllocCPUS      State ExitCode &lt;br /&gt;
  ------------ ---------- ---------- ---------- ---------- ---------- -------- &lt;br /&gt;
  3385         BOV-WUR-58   research                    12  COMPLETED      0:0 &lt;br /&gt;
  3385.batch        batch                                1  COMPLETED      0:0 &lt;br /&gt;
  3386         BOV-WUR-59   research                    12 CANCELLED+      0:0 &lt;br /&gt;
  3386.batch        batch                                1  CANCELLED     0:15 &lt;br /&gt;
  3528         BOV-WUR-59       ABGC                    16    RUNNING      0:0 &lt;br /&gt;
  3529         BOV-WUR-60       ABGC                    16    RUNNING      0:0&lt;br /&gt;
&lt;br /&gt;
Or in more detail for a specific job:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
sacct --format=jobid,jobname,account,partition,ntasks,alloccpus,elapsed,state,exitcode -j 4220&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
This should provide information about job id 4220:&lt;br /&gt;
&lt;br /&gt;
       JobID    JobName    Account  Partition   NTasks  AllocCPUS    Elapsed      State ExitCode &lt;br /&gt;
  ------------ ---------- ---------- ---------- -------- ---------- ---------- ---------- -------- &lt;br /&gt;
  4220         PreProces+              research                   3   00:30:52  COMPLETED      0:0 &lt;br /&gt;
  4220.batch        batch                              1          1   00:30:52  COMPLETED      0:0&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Job Status Codes&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Typically your job will be either in the Running state of PenDing state. However here is a breakdown of all the states that your job could be in.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
!Code!!State!!Description&lt;br /&gt;
|-&lt;br /&gt;
|CA	||CANCELLED||	Job was explicitly cancelled by the user or system administrator. The job may or may not have been initiated.&lt;br /&gt;
|-&lt;br /&gt;
|CD||	COMPLETED||	Job has terminated all processes on all nodes.&lt;br /&gt;
|-&lt;br /&gt;
|CF||	CONFIGURING||	Job has been allocated resources, but are waiting for them to become ready for use (e.g. booting).&lt;br /&gt;
|-&lt;br /&gt;
|CG||	COMPLETING||	Job is in the process of completing. Some processes on some nodes may still be active.&lt;br /&gt;
|-&lt;br /&gt;
|F||	FAILED||	Job terminated with non-zero exit code or other failure condition.&lt;br /&gt;
|-&lt;br /&gt;
|NF||	NODE_FAIL||	Job terminated due to failure of one or more allocated nodes.&lt;br /&gt;
|-&lt;br /&gt;
|PD||	PENDING||	Job is awaiting resource allocation.&lt;br /&gt;
|-&lt;br /&gt;
|R||	RUNNING||	Job currently has an allocation.&lt;br /&gt;
|-&lt;br /&gt;
|S||	SUSPENDED||	Job has an allocation, but execution has been suspended.&lt;br /&gt;
|-&lt;br /&gt;
|TO||	TIMEOUT||	Job terminated upon reaching its time limit.&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Running MPI jobs on B4F cluster ==&lt;br /&gt;
&lt;br /&gt;
[[MPI_on_B4F_cluster | Main article: MPI on B4F Cluster]]&lt;br /&gt;
&amp;lt; text here &amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Understanding which resources are available to you: sinfo ==&lt;br /&gt;
By using the &#039;sinfo&#039; command you can retrieve information on which &#039;Partitions&#039; are available to you. A &#039;Partition&#039; using SLURM is similar to the &#039;queue&#039; when submitting using the Sun Grid Engine (&#039;qsub&#039;). The different Partitions grant different levels of resource allocation. Not all defined Partitions will be available to any given person. E.g., Master students will only have the Student Partition available, researchers at the ABGC will have &#039;student&#039;, &#039;research&#039;, and &#039;ABGC&#039; partitions available. The higher the level of  resource allocation, though, the higher the cost per compute-hour. The default Partition is the &#039;student&#039; partition. A full list of Partitions can be found from the Bright Cluster Manager webpage.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
sinfo&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  PARTITION AVAIL  TIMELIMIT  NODES  STATE NODELIST&lt;br /&gt;
  student*     up   infinite     12  down* node[043-048,055-060]&lt;br /&gt;
  student*     up   infinite     50   idle fat[001-002],node[001-042,049-054]&lt;br /&gt;
  research     up   infinite     12  down* node[043-048,055-060]&lt;br /&gt;
  research     up   infinite     50   idle fat[001-002],node[001-042,049-054]&lt;br /&gt;
  ABGC         up   infinite     12  down* node[043-048,055-060]&lt;br /&gt;
  ABGC         up   infinite     50   idle fat[001-002],node[001-042,049-054]&lt;br /&gt;
&lt;br /&gt;
== See also ==&lt;br /&gt;
* [[B4F_cluster | B4F Cluster]]&lt;br /&gt;
* [[BCM_on_B4F_cluster | BCM on B4F cluster]]&lt;br /&gt;
* [[SLURM_Compare | SLURM compared to other common schedulers]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
&lt;br /&gt;
== External links ==&lt;br /&gt;
* [http://slurm.schedmd.com Slurm official documentation]&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Simple_Linux_Utility_for_Resource_Management Slurm on Wikipedia]&lt;br /&gt;
* [http://www.youtube.com/watch?v=axWffyrk3aY Slurm Tutorial on Youtube]&lt;/div&gt;</summary>
		<author><name>Duque004</name></author>
	</entry>
</feed>