<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.anunna.wur.nl/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Dawes0011</id>
	<title>HPCwiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.anunna.wur.nl/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Dawes0011"/>
	<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php/Special:Contributions/Dawes0011"/>
	<updated>2026-04-18T01:38:53Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.1</generator>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Old_binaries&amp;diff=2244</id>
		<title>Old binaries</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Old_binaries&amp;diff=2244"/>
		<updated>2023-06-19T09:27:57Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;On Linux, most binaries rely on shared libraries that are loaded by the OS before the binary is executed.&lt;br /&gt;
&lt;br /&gt;
This allows them to share common runtimes, and avoid having to redo the same work multiple times. This is useful for optimisation and security, as debugging only has to be done in one place for multiple applications.&lt;br /&gt;
&lt;br /&gt;
However, a significant number of these libraries are automatically inserted by the OS, and as the cluster upgrades and changes some of these will no longer be automatically available. This will make older installs fail to execute, giving, for instance, errors related to glibc, or missing symbols.&lt;br /&gt;
&lt;br /&gt;
Thankfully, the entire library from previous OS builds is made available on the shared filesystem, granting older binaries the ability to access them whilst also running on a younger OS.&lt;br /&gt;
&lt;br /&gt;
There is a special environment variable: &amp;lt;code&amp;gt;LD_LIBRARY_PATH&amp;lt;/code&amp;gt; which tells the linker (ld) where extra to go find libraries. We use this extensively in the underlying module system on Anunna to control which libraries can be used. One such module, &amp;lt;code&amp;gt;sl7_libs&amp;lt;/code&amp;gt; automatically adds the older libraries in to older binaries, and has been added as a prerequisite for most older installs.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If you&#039;ve compiled something yourself that isn&#039;t a module, you may find yourself needing this module. Simply:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
module load sl7-libs/main&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To add this path.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Jupyter ==&lt;br /&gt;
&lt;br /&gt;
This may also extend to custom-built kernels on Jupyter. Here, however, you can&#039;t add this module, as instantiating a kernel is done though &amp;lt;code&amp;gt;kernel.json&amp;lt;/code&amp;gt; instead of bash.&lt;br /&gt;
&lt;br /&gt;
To add the needed libraries, you need to modify your notebook to look like the following:&lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
 &amp;quot;language&amp;quot;: &amp;quot;python&amp;quot;,&lt;br /&gt;
 &amp;quot;argv&amp;quot;: [&lt;br /&gt;
  &amp;quot;/path/to/my/venv/bin/python&amp;quot;,&lt;br /&gt;
  &amp;quot;-m&amp;quot;,&lt;br /&gt;
  &amp;quot;ipykernel&amp;quot;,&lt;br /&gt;
  &amp;quot;-f&amp;quot;,&lt;br /&gt;
  &amp;quot;{connection_file}&amp;quot;&lt;br /&gt;
 ],&lt;br /&gt;
 &amp;quot;display_name&amp;quot;: &amp;quot;myvenv&amp;quot;,&lt;br /&gt;
 &amp;quot;env&amp;quot; : {&amp;quot;LD_LIBRARY_PATH&amp;quot;: &amp;quot;/usr/lib:/usr/lib64:/usr/lib/x86_64-linux-gnu:/shared/legacyapps/sl7-libs/lib:/shared/legacyapps/sl7-libs/lib64&amp;quot; }&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When editing JSON, please be mindful that all elements in an object must be separated by commas, and that only doublequotes may be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To test your JSON, try the following:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
cat ~/.local.share/jupyter/kernels/mykernel/kernel.json | jq&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If &amp;lt;code&amp;gt;jq&amp;lt;/code&amp;gt; repeats your JSON, then it can parse it and you are good to go.&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Old_binaries&amp;diff=2243</id>
		<title>Old binaries</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Old_binaries&amp;diff=2243"/>
		<updated>2023-06-19T09:27:26Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;On Linux, most binaries rely on shared libraries that are loaded by the OS before the binary is executed.&lt;br /&gt;
&lt;br /&gt;
This allows them to share common runtimes, and avoid having to redo the same work multiple times. This is useful for optimisation and security, as debugging only has to be done in one place for multiple applications.&lt;br /&gt;
&lt;br /&gt;
However, a significant number of these libraries are automatically inserted by the OS, and as the cluster upgrades and changes some of these will no longer be automatically available. This will make older installs fail to execute, giving, for instance, errors related to glibc, or missing symbols.&lt;br /&gt;
&lt;br /&gt;
Thankfully, the entire library from previous OS builds is made available on the shared filesystem, granting older binaries the ability to access them whilst also running on a younger OS.&lt;br /&gt;
&lt;br /&gt;
There is a special environment variable: &amp;lt;code&amp;gt;LD_LIBRARY_PATH&amp;lt;/code&amp;gt; which tells the linker (ld) where extra to go find libraries. We use this extensively in the underlying module system on Anunna to control which libraries can be used. One such module, &amp;lt;code&amp;gt;sl7_libs&amp;lt;/code&amp;gt; automatically adds the older libraries in to older binaries, and has been added as a prerequisite for most older installs.&lt;br /&gt;
&lt;br /&gt;
If you&#039;ve compiled something yourself that isn&#039;t a module, you may find yourself needing this module. Simply:&lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
module load sl7-libs/main&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To add this path.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Jupyter ==&lt;br /&gt;
&lt;br /&gt;
This may also extend to custom-built kernels on Jupyter. Here, however, you can&#039;t add this module, as instantiating a kernel is done though &amp;lt;code&amp;gt;kernel.json&amp;lt;/code&amp;gt; instead of bash.&lt;br /&gt;
&lt;br /&gt;
To add the needed libraries, you need to modify your notebook to look like the following:&lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
 &amp;quot;language&amp;quot;: &amp;quot;python&amp;quot;,&lt;br /&gt;
 &amp;quot;argv&amp;quot;: [&lt;br /&gt;
  &amp;quot;/path/to/my/venv/bin/python&amp;quot;,&lt;br /&gt;
  &amp;quot;-m&amp;quot;,&lt;br /&gt;
  &amp;quot;ipykernel&amp;quot;,&lt;br /&gt;
  &amp;quot;-f&amp;quot;,&lt;br /&gt;
  &amp;quot;{connection_file}&amp;quot;&lt;br /&gt;
 ],&lt;br /&gt;
 &amp;quot;display_name&amp;quot;: &amp;quot;myvenv&amp;quot;,&lt;br /&gt;
 &amp;quot;env&amp;quot; : {&amp;quot;LD_LIBRARY_PATH&amp;quot;: &amp;quot;/usr/lib:/usr/lib64:/usr/lib/x86_64-linux-gnu:/shared/legacyapps/sl7-libs/lib:/shared/legacyapps/sl7-libs/lib64&amp;quot; }&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When editing JSON, please be mindful that all elements in an object must be separated by commas, and that only doublequotes may be used.&lt;br /&gt;
&lt;br /&gt;
To test your JSON, try the following:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
cat ~/.local.share/jupyter/kernels/mykernel/kernel.json | jq&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If &amp;lt;code&amp;gt;jq&amp;lt;/code&amp;gt; repeats your JSON, then it can parse it and you are good to go.&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Old_binaries&amp;diff=2242</id>
		<title>Old binaries</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Old_binaries&amp;diff=2242"/>
		<updated>2023-06-19T09:27:09Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: Created page with &amp;quot;On Linux, most binaries rely on shared libraries that are loaded by the OS before the binary is executed.  This allows them to share common runtimes, and avoid having to redo the same work multiple times. This is useful for optimisation and security, as debugging only has to be done in one place for multiple applications.  However, a significant number of these libraries are automatically inserted by the OS, and as the cluster upgrades and changes some of these will no l...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;On Linux, most binaries rely on shared libraries that are loaded by the OS before the binary is executed.&lt;br /&gt;
&lt;br /&gt;
This allows them to share common runtimes, and avoid having to redo the same work multiple times. This is useful for optimisation and security, as debugging only has to be done in one place for multiple applications.&lt;br /&gt;
&lt;br /&gt;
However, a significant number of these libraries are automatically inserted by the OS, and as the cluster upgrades and changes some of these will no longer be automatically available. This will make older installs fail to execute, giving, for instance, errors related to glibc, or missing symbols.&lt;br /&gt;
&lt;br /&gt;
Thankfully, the entire library from previous OS builds is made available on the shared filesystem, granting older binaries the ability to access them whilst also running on a younger OS.&lt;br /&gt;
&lt;br /&gt;
There is a special environment variable: &amp;lt;code&amp;gt;LD_LIBRARY_PATH&amp;lt;/code&amp;gt; which tells the linker (ld) where extra to go find libraries. We use this extensively in the underlying module system on Anunna to control which libraries can be used. One such module, &amp;lt;code&amp;gt;sl7_libs&amp;lt;/code&amp;gt; automatically adds the older libraries in to older binaries, and has been added as a prerequisite for most older installs.&lt;br /&gt;
&lt;br /&gt;
If you&#039;ve compiled something yourself that isn&#039;t a module, you may find yourself needing this module. Simply:&lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
module load sl7-libs/main&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To add this path.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Jupyter ==&lt;br /&gt;
&lt;br /&gt;
This may also extend to custom-built kernels on Jupyter. Here, however, you can&#039;t add this module, as instantiating a kernel is done though &amp;lt;code&amp;gt;kernel.json&amp;lt;/code&amp;gt; instead of bash.&lt;br /&gt;
&lt;br /&gt;
To add the needed libraries, you need to modify your notebook to look like the following:&lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
 &amp;quot;language&amp;quot;: &amp;quot;python&amp;quot;,&lt;br /&gt;
 &amp;quot;argv&amp;quot;: [&lt;br /&gt;
  &amp;quot;/path/to/my/venv/bin/python&amp;quot;,&lt;br /&gt;
  &amp;quot;-m&amp;quot;,&lt;br /&gt;
  &amp;quot;ipykernel&amp;quot;,&lt;br /&gt;
  &amp;quot;-f&amp;quot;,&lt;br /&gt;
  &amp;quot;{connection_file}&amp;quot;&lt;br /&gt;
 ],&lt;br /&gt;
 &amp;quot;display_name&amp;quot;: &amp;quot;myvenv&amp;quot;,&lt;br /&gt;
 &amp;quot;env&amp;quot; : {&amp;quot;LD_LIBRARY_PATH&amp;quot;: &amp;quot;/usr/lib:/usr/lib64:/usr/lib/x86_64-linux-gnu:/shared/legacyapps/sl7-libs/lib:/shared/legacyapps/sl7-libs/lib64&amp;quot; }&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When editing JSON, please be mindful that all elements in an object must be separated by commas, and that only doublequotes may be used.&lt;br /&gt;
&lt;br /&gt;
To test your JSON, try the following:&lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
cat ~/.local.share/jupyter/kernels/mykernel/kernel.json | jq&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If &amp;lt;code&amp;gt;jq&amp;lt;/code&amp;gt; repeats your JSON, then it can parse it and you are good to go.&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Main_Page&amp;diff=2241</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Main_Page&amp;diff=2241"/>
		<updated>2023-06-19T09:11:41Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: /* Miscellaneous */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Anunna is a [http://en.wikipedia.org/wiki/High-performance_computing High Performance Computer] (HPC) infrastructure hosted by [http://www.wageningenur.nl/nl/activiteit/Opening-High-Performance-Computing-cluster-HPC.htm Wageningen University &amp;amp; Research Centre]. It is open for use for all WUR research groups as well as other organizations, including companies, that have collaborative projects with WUR. &lt;br /&gt;
&lt;br /&gt;
= Using Anunna =&lt;br /&gt;
* [[Tariffs | Costs associated with resource usage]]&lt;br /&gt;
&lt;br /&gt;
== Gaining access to Anunna==&lt;br /&gt;
Access to the cluster and file transfer are traditionally done via [http://en.wikipedia.org/wiki/Secure_Shell SSH and SFTP].&lt;br /&gt;
* [[log_in_to_B4F_cluster | Logging into cluster using ssh]]&lt;br /&gt;
* [[file_transfer | File transfer options]]&lt;br /&gt;
* [[Services | Alternative access methods, and extra features and services on Anunna]]&lt;br /&gt;
* [[Filesystems | Data storage methods on Anunna]]&lt;br /&gt;
&lt;br /&gt;
== Access Policy ==&lt;br /&gt;
[[Access_Policy | Main Article: Access Policy]]&lt;br /&gt;
&lt;br /&gt;
Access needs to be granted actively (by creation of an account on the cluster by FB-IT). Use of resources is limited by the scheduler. Depending on availability of queues (&#039;partitions&#039;) granted to a user, priority to the system&#039;s resources is regulated. Note that the use of Anunna is not free of charge. List price of CPU time and storage, and possible discounts on that list price for your organisation, can be retrieved from Shared Research Facilities or FB-IT.&lt;br /&gt;
&lt;br /&gt;
= Events =&lt;br /&gt;
&lt;br /&gt;
* [[Courses]] that have happened and are happening&lt;br /&gt;
* [[Downtime]] that will affect all users&lt;br /&gt;
* [[Meetings]] that may affect the policies of Anunna&lt;br /&gt;
&lt;br /&gt;
= Other Software =&lt;br /&gt;
&lt;br /&gt;
== Cluster Management Software and Scheduler ==&lt;br /&gt;
Anunna uses Bright Cluster Manager software for overall cluster management, and Slurm as job scheduler.&lt;br /&gt;
* [[BCM_on_B4F_cluster | Monitor cluster status with BCM]]&lt;br /&gt;
* [[Using_Slurm | Submit jobs with Slurm]]&lt;br /&gt;
* [[node_usage_graph | Be aware of how much work the cluster is under right now with &#039;node_usage_graph&#039;]]&lt;br /&gt;
* [[SLURM_Compare | Rosetta Stone of Workload Managers]]&lt;br /&gt;
&lt;br /&gt;
== Installation of software by users ==&lt;br /&gt;
&lt;br /&gt;
* [[Domain_specific_software_on_B4Fcluster_installation_by_users | Installing domain specific software: installation by users]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
* [[Virtual_environment_Python_3.4_or_higher | Setting up and using a virtual environment for Python3.4 or higher ]]&lt;br /&gt;
* [[Installing WRF and WPS]]&lt;br /&gt;
* [[Running scripts on a fixed timeschedule (cron)]]&lt;br /&gt;
&lt;br /&gt;
== Installed software ==&lt;br /&gt;
&lt;br /&gt;
* [[Globally_installed_software | Globally installed software]]&lt;br /&gt;
* [[ABGC_modules | ABGC specific modules]]&lt;br /&gt;
&lt;br /&gt;
= Useful Notes = &lt;br /&gt;
&lt;br /&gt;
== Being in control of Environment parameters ==&lt;br /&gt;
&lt;br /&gt;
* [[Using_environment_modules | Using environment modules]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Setting_TMPDIR | Set a custom temporary directory location]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
&lt;br /&gt;
== Controlling costs ==&lt;br /&gt;
&lt;br /&gt;
* [[SACCT | using SACCT to see your costs]]&lt;br /&gt;
* [[get_my_bill | using the &amp;quot;get_my_bill&amp;quot; script to estimate costs]]&lt;br /&gt;
&lt;br /&gt;
== Management ==&lt;br /&gt;
Product Owner of Anunna is Alexander van Ittersum (Wageningen UR,FB-IT, C&amp;amp;PS). [[User:dawes001 | Gwen Dawes (Wageningen UR, FB-IT, C&amp;amp;PS)]] and [[User:haars001 | Jan van Haarst (Wageningen UR,FB-IT, C&amp;amp;PS)]] are responsible for [[Maintenance_and_Management | Maintenance and Management]] of the cluster.&lt;br /&gt;
&lt;br /&gt;
* [[Roadmap | Ambitions regarding innovation, support and administration of Anunna ]]&lt;br /&gt;
&lt;br /&gt;
= Miscellaneous =&lt;br /&gt;
* [[Mailinglist | Electronic mail discussion lists]]&lt;br /&gt;
* [[History_of_the_Cluster | Historical information on the startup of Anunna]]&lt;br /&gt;
* [[Bioinformatics_tips_tricks_workflows | Bioinformatics tips, tricks, and workflows]]&lt;br /&gt;
* [[Parallel_R_code_on_SLURM | Running parallel R code on SLURM]]&lt;br /&gt;
* [[Convert_between_MediaWiki_and_other_formats | Convert between MediaWiki format and other formats]]&lt;br /&gt;
* [[Manual GitLab | GitLab: Create projects and add scripts]]&lt;br /&gt;
* [[Monitoring_executions | Monitoring job execution]]&lt;br /&gt;
* [[Shared_folders | Working with shared folders in the Lustre file system]]&lt;br /&gt;
* [[Old_binaries | Running older binaries on the updated OS]]&lt;br /&gt;
&lt;br /&gt;
= See also =&lt;br /&gt;
* [[Maintenance_and_Management | Maintenance and Management]]&lt;br /&gt;
* [[BCData | BCData]]&lt;br /&gt;
* [[Mailinglist | Electronic mail discussion lists]]&lt;br /&gt;
* [[About_ABGC | About ABGC]]&lt;br /&gt;
* [[Computer_cluster | High Performance Computing @ABGC]]&lt;br /&gt;
* [[Lustre_PFS_layout | Lustre Parallel File System layout]]&lt;br /&gt;
&lt;br /&gt;
= External links =&lt;br /&gt;
{| width=&amp;quot;90%&amp;quot;&lt;br /&gt;
|- valign=&amp;quot;top&amp;quot;&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [https://www.wur.nl/en/Value-Creation-Cooperation/Facilities/Wageningen-Shared-Research-Facilities/Our-facilities/Show/High-Performance-Computing-Cluster-HPC-Anunna.htm SRF offers a HPC facilty]&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Scientific_Linux Scientific Linux]&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Help:Cheatsheet Help with editing Wiki pages]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Main_Page&amp;diff=2240</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Main_Page&amp;diff=2240"/>
		<updated>2023-06-19T09:11:27Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: /* Miscellaneous */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Anunna is a [http://en.wikipedia.org/wiki/High-performance_computing High Performance Computer] (HPC) infrastructure hosted by [http://www.wageningenur.nl/nl/activiteit/Opening-High-Performance-Computing-cluster-HPC.htm Wageningen University &amp;amp; Research Centre]. It is open for use for all WUR research groups as well as other organizations, including companies, that have collaborative projects with WUR. &lt;br /&gt;
&lt;br /&gt;
= Using Anunna =&lt;br /&gt;
* [[Tariffs | Costs associated with resource usage]]&lt;br /&gt;
&lt;br /&gt;
== Gaining access to Anunna==&lt;br /&gt;
Access to the cluster and file transfer are traditionally done via [http://en.wikipedia.org/wiki/Secure_Shell SSH and SFTP].&lt;br /&gt;
* [[log_in_to_B4F_cluster | Logging into cluster using ssh]]&lt;br /&gt;
* [[file_transfer | File transfer options]]&lt;br /&gt;
* [[Services | Alternative access methods, and extra features and services on Anunna]]&lt;br /&gt;
* [[Filesystems | Data storage methods on Anunna]]&lt;br /&gt;
&lt;br /&gt;
== Access Policy ==&lt;br /&gt;
[[Access_Policy | Main Article: Access Policy]]&lt;br /&gt;
&lt;br /&gt;
Access needs to be granted actively (by creation of an account on the cluster by FB-IT). Use of resources is limited by the scheduler. Depending on availability of queues (&#039;partitions&#039;) granted to a user, priority to the system&#039;s resources is regulated. Note that the use of Anunna is not free of charge. List price of CPU time and storage, and possible discounts on that list price for your organisation, can be retrieved from Shared Research Facilities or FB-IT.&lt;br /&gt;
&lt;br /&gt;
= Events =&lt;br /&gt;
&lt;br /&gt;
* [[Courses]] that have happened and are happening&lt;br /&gt;
* [[Downtime]] that will affect all users&lt;br /&gt;
* [[Meetings]] that may affect the policies of Anunna&lt;br /&gt;
&lt;br /&gt;
= Other Software =&lt;br /&gt;
&lt;br /&gt;
== Cluster Management Software and Scheduler ==&lt;br /&gt;
Anunna uses Bright Cluster Manager software for overall cluster management, and Slurm as job scheduler.&lt;br /&gt;
* [[BCM_on_B4F_cluster | Monitor cluster status with BCM]]&lt;br /&gt;
* [[Using_Slurm | Submit jobs with Slurm]]&lt;br /&gt;
* [[node_usage_graph | Be aware of how much work the cluster is under right now with &#039;node_usage_graph&#039;]]&lt;br /&gt;
* [[SLURM_Compare | Rosetta Stone of Workload Managers]]&lt;br /&gt;
&lt;br /&gt;
== Installation of software by users ==&lt;br /&gt;
&lt;br /&gt;
* [[Domain_specific_software_on_B4Fcluster_installation_by_users | Installing domain specific software: installation by users]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
* [[Virtual_environment_Python_3.4_or_higher | Setting up and using a virtual environment for Python3.4 or higher ]]&lt;br /&gt;
* [[Installing WRF and WPS]]&lt;br /&gt;
* [[Running scripts on a fixed timeschedule (cron)]]&lt;br /&gt;
&lt;br /&gt;
== Installed software ==&lt;br /&gt;
&lt;br /&gt;
* [[Globally_installed_software | Globally installed software]]&lt;br /&gt;
* [[ABGC_modules | ABGC specific modules]]&lt;br /&gt;
&lt;br /&gt;
= Useful Notes = &lt;br /&gt;
&lt;br /&gt;
== Being in control of Environment parameters ==&lt;br /&gt;
&lt;br /&gt;
* [[Using_environment_modules | Using environment modules]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Setting_TMPDIR | Set a custom temporary directory location]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
&lt;br /&gt;
== Controlling costs ==&lt;br /&gt;
&lt;br /&gt;
* [[SACCT | using SACCT to see your costs]]&lt;br /&gt;
* [[get_my_bill | using the &amp;quot;get_my_bill&amp;quot; script to estimate costs]]&lt;br /&gt;
&lt;br /&gt;
== Management ==&lt;br /&gt;
Product Owner of Anunna is Alexander van Ittersum (Wageningen UR,FB-IT, C&amp;amp;PS). [[User:dawes001 | Gwen Dawes (Wageningen UR, FB-IT, C&amp;amp;PS)]] and [[User:haars001 | Jan van Haarst (Wageningen UR,FB-IT, C&amp;amp;PS)]] are responsible for [[Maintenance_and_Management | Maintenance and Management]] of the cluster.&lt;br /&gt;
&lt;br /&gt;
* [[Roadmap | Ambitions regarding innovation, support and administration of Anunna ]]&lt;br /&gt;
&lt;br /&gt;
= Miscellaneous =&lt;br /&gt;
* [[Mailinglist | Electronic mail discussion lists]]&lt;br /&gt;
* [[History_of_the_Cluster | Historical information on the startup of Anunna]]&lt;br /&gt;
* [[Bioinformatics_tips_tricks_workflows | Bioinformatics tips, tricks, and workflows]]&lt;br /&gt;
* [[Parallel_R_code_on_SLURM | Running parallel R code on SLURM]]&lt;br /&gt;
* [[Convert_between_MediaWiki_and_other_formats | Convert between MediaWiki format and other formats]]&lt;br /&gt;
* [[Manual GitLab | GitLab: Create projects and add scripts]]&lt;br /&gt;
* [[Monitoring_executions | Monitoring job execution]]&lt;br /&gt;
* [[Shared_folders | Working with shared folders in the Lustre file system]]&lt;br /&gt;
* [[Old_binaries | running older binaries on the updated OS]]&lt;br /&gt;
&lt;br /&gt;
= See also =&lt;br /&gt;
* [[Maintenance_and_Management | Maintenance and Management]]&lt;br /&gt;
* [[BCData | BCData]]&lt;br /&gt;
* [[Mailinglist | Electronic mail discussion lists]]&lt;br /&gt;
* [[About_ABGC | About ABGC]]&lt;br /&gt;
* [[Computer_cluster | High Performance Computing @ABGC]]&lt;br /&gt;
* [[Lustre_PFS_layout | Lustre Parallel File System layout]]&lt;br /&gt;
&lt;br /&gt;
= External links =&lt;br /&gt;
{| width=&amp;quot;90%&amp;quot;&lt;br /&gt;
|- valign=&amp;quot;top&amp;quot;&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [https://www.wur.nl/en/Value-Creation-Cooperation/Facilities/Wageningen-Shared-Research-Facilities/Our-facilities/Show/High-Performance-Computing-Cluster-HPC-Anunna.htm SRF offers a HPC facilty]&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Scientific_Linux Scientific Linux]&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Help:Cheatsheet Help with editing Wiki pages]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Filesystems&amp;diff=2239</id>
		<title>Filesystems</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Filesystems&amp;diff=2239"/>
		<updated>2023-06-16T09:04:11Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Anunna currently has multiple filesystem mounts that are available cluster-wide:&lt;br /&gt;
&lt;br /&gt;
== Global ==&lt;br /&gt;
* /home - This mount uses NFS to mount the home directories directly from nfs01. Each user has a 200G quota for this filesystem, as it is regularly backed up to tape, and can reliably be restored from up to a week&#039;s history.&lt;br /&gt;
&lt;br /&gt;
* /shared - This mount provides a consistent set of binaries for the entire cluster.&lt;br /&gt;
&lt;br /&gt;
* /lustre - This large mount uses the Lustre filesystem to provide files from multiple redundant servers. Access is provided per group, thus:&lt;br /&gt;
 /lustre/[level]/[partner]/[unit]&lt;br /&gt;
e.g.&lt;br /&gt;
 /lustre/backup/WUR/ABGC/&lt;br /&gt;
It comprises of two major parts (and some minor):&lt;br /&gt;
* /lustre/backup - In case of disaster, this data is stored a second time on a separate machine. Whilst this backup is purely in case of complete tragedy (such as some immense filesystem error, or multiple component failure), it can potentially be used to revert mistakes if you are very fast about reporting them. There is however no guarantee of this service.&lt;br /&gt;
* /lustre/nobackup - This is the &#039;normal&#039; filesystem for Lustre - no backups, just stored on the filesystem. Without having a backup needed, the cost of data here is not as much as under /lustre/backup, but in case of disaster cannot be recivered.&lt;br /&gt;
* /lustre/shared - Same as /lustre/backup, except publicly available. This is where truly shared data lives that isn&#039;t assigned to a specific group.&lt;br /&gt;
&lt;br /&gt;
And additionally:&lt;br /&gt;
* /lustre/scratch - A separated, low resilience filesystem. Files here may be removed after some time if the filesystem gets too full (Typically 30 days). You should tidy up this data yourself once work is complete.&lt;br /&gt;
&lt;br /&gt;
=== Private shared directories ===&lt;br /&gt;
If you are working with a group of users on a similar project, you might consider making a [[Shared_folders|Shared directory]] to coordinate. Information on how to do so is in the linked article.&lt;br /&gt;
&lt;br /&gt;
== Local ==&lt;br /&gt;
Specific to certain machines are some other filesystems that are available to you:&lt;br /&gt;
* /archive - an archive mount only accessible from the login nodes. Files here are sent to the Isilon for deeper storage. The cost of storing data here is much less than on the Lustre, but it cannot be used for compute work. This location is only available to WUR users. Files are able to be reverted via snapshot, and there is a separated backup, however this only comes in fortnightly (14 day) intervals.&lt;br /&gt;
&lt;br /&gt;
* /tmp - On each worker node there is a /tmp mount that can be used for temporary local caching. Be advised that you should clean this up, lest your files become a hindrance to other users. You can request a node with free space in your sbatch script like so:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
#SBATCH --tmp=&amp;lt;required space&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* /dev/shm - On each worker you may also create a virtual filesystem directly into memory, for extremely fast data access. Be advised that this will count against the memory used for your job, but it is also the fastest available filesystem if needed.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== See also ==&lt;br /&gt;
* [[Tariffs | Costs associated with resource usage]]&lt;br /&gt;
&lt;br /&gt;
== External links ==&lt;br /&gt;
* [http://wiki.lustre.org/index.php/Main_Page Lustre website]&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Setting_up_Python_virtualenv&amp;diff=2238</id>
		<title>Setting up Python virtualenv</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Setting_up_Python_virtualenv&amp;diff=2238"/>
		<updated>2023-06-16T09:02:11Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;With many Python packages available, which are often in conflict or requiring different versions depending on application, installing and controlling packages and versions is not always easy. In addition, so many packages are often used only occasionally, that it is questionable whether a system administrator of a centralized server system or a High Performance Compute (HPC) infrastructure can be expected to resolve all issues posed by users of the infrastructure. Even on a local system with full administrative rights managing versions, dependencies, and package collisions is often very difficult. The solution is to use a virtual environment, in which a specific set of packages can then be installed. As many different virtual environments can be created, and used side-by-side, as is necessary. &lt;br /&gt;
&lt;br /&gt;
NOTE: as of Python 3.3 virtual environment support is built-in. See this page for an [[virtual_environment_Python_3.4_or_higher | alternative set-up of your virtual environment if using Python 3.4 or higher]].&lt;br /&gt;
&lt;br /&gt;
== Creating a new virtual environment ==&lt;br /&gt;
It is assumed that the appropriate &amp;lt;code&amp;gt;virtualenv&amp;lt;/code&amp;gt; executable for the Python version of choice is installed. A new virtual environment, in this case called &amp;lt;code&amp;gt;newenv&amp;lt;/code&amp;gt; is created like so:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
module load python/my-favourite-version (e.g. 2.7.12)&lt;br /&gt;
virtualenv newenv&lt;br /&gt;
OR&lt;br /&gt;
pyvenv newenv (For versions &amp;gt;3.4)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When the new environment is created, one will see a message similar to this:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;  New python executable in newenv/bin/python3&lt;br /&gt;
  Also creating executable in newenv/bin/python&lt;br /&gt;
  Installing Setuptools.........................................................................done.&lt;br /&gt;
  Installing Pip................................................................................done.&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Activating a virtual environment ==&lt;br /&gt;
Once the environment is created, each time the environment needs to be activated, the following command needs to be issued:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
source newenv/bin/activate&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This assumes that the folder that contains the virtual environment documents (in this case called &amp;lt;code&amp;gt;newenv&amp;lt;/code&amp;gt;), is in the present working directory.&lt;br /&gt;
When working on the virtual environment, the virtual environment name will be between brackets in front of the &amp;lt;code&amp;gt;user-host-prompt&amp;lt;/code&amp;gt; string.&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;  (newenv)user@host:~$&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Installing modules on the virtual environment ==&lt;br /&gt;
Installing modules is the same as usual. The difference is that modules are in &amp;lt;code&amp;gt;/path/to/virtenv/lib&amp;lt;/code&amp;gt;, which may be living somewhere on your home directory. When working from the virtual environment, the default &amp;lt;code&amp;gt;pip&amp;lt;/code&amp;gt; will belong to the python version that is currently active. This means that the executable in &amp;lt;code&amp;gt;/path/to/virtenv/bin&amp;lt;/code&amp;gt; are in fact the first in the &amp;lt;code&amp;gt;$PATH&amp;lt;/code&amp;gt;.&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
pip install numpy&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Similarly, installing packages from source works exactly the same as usual.&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
python setup.py install&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== deactivating a virtual environment ==&lt;br /&gt;
Quitting a virtual environment can be done by using the command &amp;lt;code&amp;gt;deactivate&amp;lt;/code&amp;gt;, which was loaded using the &amp;lt;code&amp;gt;source&amp;lt;/code&amp;gt; command upon activating the virtual environment.&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
deactivate&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Virtualenv kernels in Jupyter ==&lt;br /&gt;
Want your own virtualenv kernel in a notebook? This can be done by making your own kernel specifications:&lt;br /&gt;
&lt;br /&gt;
(an alternative way to the manual way (using conda) is described [[Using conda to install a new kernel into your notebook|here ]])&lt;br /&gt;
&lt;br /&gt;
* Make sure you have the ipykernel module in your venv. Activate it and pip install it:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;source ~/path/to/my/virtualenv/bin/activate &amp;amp;&amp;amp; pip install ipykernel&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Create the following directory path in your homedir if it doesn&#039;t already exist:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;mkdir -p ~/.local/share/jupyter/kernels/&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Think of a nice descriptive name that doesn&#039;t clash with one of the already present kernels. I&#039;ll use &#039;testing&#039;. Create this folder:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;mkdir ~/.local/share/jupyter/kernels/testing/&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Add this file to this folder:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;vi ~/.local/share/jupyter/kernels/testing/kernel.json &lt;br /&gt;
{&lt;br /&gt;
 &amp;quot;language&amp;quot;: &amp;quot;python&amp;quot;,&lt;br /&gt;
 &amp;quot;argv&amp;quot;: [&lt;br /&gt;
  &amp;quot;/home/myhome/path/to/my/virtualenv/bin/python&amp;quot;,&lt;br /&gt;
  &amp;quot;-m&amp;quot;,&lt;br /&gt;
  &amp;quot;ipykernel&amp;quot;,&lt;br /&gt;
  &amp;quot;-f&amp;quot;,&lt;br /&gt;
  &amp;quot;{connection_file}&amp;quot;&lt;br /&gt;
 ],&lt;br /&gt;
 &amp;quot;display_name&amp;quot;: &amp;quot;testing&amp;quot;&lt;br /&gt;
}&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Reload Jupyterhub page. testing should now exist in your kernels list.&lt;br /&gt;
&lt;br /&gt;
You can do more complex things with this, such as construct your own Spark environment. This relies on having the module findspark installed:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt; vi ~/.local/share/jupyter/kernels/mysparkkernel/kernel.json &lt;br /&gt;
{&lt;br /&gt;
 &amp;quot;language&amp;quot;: &amp;quot;python&amp;quot;,&lt;br /&gt;
 &amp;quot;env&amp;quot;: {&lt;br /&gt;
   &amp;quot;SPARK_HOME&amp;quot;:&lt;br /&gt;
     &amp;quot;/shared/apps/spark/my-spark-version&amp;quot;&lt;br /&gt;
 },&lt;br /&gt;
 &amp;quot;argv&amp;quot;: [&lt;br /&gt;
  &amp;quot;/home/myhome/my/spark/venv/bin/python&amp;quot;,&lt;br /&gt;
  &amp;quot;-m&amp;quot;,&lt;br /&gt;
  &amp;quot;ipykernel&amp;quot;,&lt;br /&gt;
  &amp;quot;-c&amp;quot;, &amp;quot;import findspark; findspark.init()&amp;quot;,&lt;br /&gt;
  &amp;quot;-f&amp;quot;,&lt;br /&gt;
  &amp;quot;{connection_file}&amp;quot;&lt;br /&gt;
 ],&lt;br /&gt;
 &amp;quot;display_name&amp;quot;: &amp;quot;My Spark kernel&amp;quot;&lt;br /&gt;
}&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
(You&#039;ll want to make sure your spark cluster has the same environment - start it after activating this venv inside your sbatch script)&lt;br /&gt;
&lt;br /&gt;
== Make IPython work under virtualenv ==&lt;br /&gt;
IPython may not work initially under a virtual environment. It may produce an error message like below:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;    File &amp;quot;/usr/bin/ipython&amp;quot;, line 11&lt;br /&gt;
    print &amp;quot;Could not start qtconsole. Please install ipython-qtconsole&amp;quot;&lt;br /&gt;
                                                                      ^&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can be resolved by adding a soft link with the name &amp;lt;code&amp;gt;ipython&amp;lt;/code&amp;gt; to the &amp;lt;code&amp;gt;bin&amp;lt;/code&amp;gt; directory in the virtual environment folder.&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
ln -s /path/to/virtenv/bin/ipython3 /path/to/virtenv/bin/ipython&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== External links ==&lt;br /&gt;
* [https://pypi.python.org/pypi/virtualenv Python3 documentation for virtualenv]&lt;br /&gt;
* [http://cemcfarland.wordpress.com/2013/03/09/getting-ipython3-working-inside-your-virtualenv/ Solving the IPython hickup under virtual environment]&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=MPI_on_B4F_cluster&amp;diff=2237</id>
		<title>MPI on B4F cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=MPI_on_B4F_cluster&amp;diff=2237"/>
		<updated>2023-06-16T09:01:59Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== A simple &#039;Hello World&#039; example ==&lt;br /&gt;
Consider the following simple MPI version, in C, of the &#039;Hello World&#039; example:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source lang=&#039;cpp&#039;&amp;gt;&lt;br /&gt;
#include &amp;lt;stdio.h&amp;gt;&lt;br /&gt;
#include &amp;lt;mpi.h&amp;gt;&lt;br /&gt;
int main(int argc, char ** argv) {&lt;br /&gt;
  int size,rank,namelen;&lt;br /&gt;
  char processor_name[MPI_MAX_PROCESSOR_NAME];&lt;br /&gt;
  MPI_Init(&amp;amp;argc, &amp;amp;argv);&lt;br /&gt;
  MPI_Comm_rank(MPI_COMM_WORLD,&amp;amp;rank);&lt;br /&gt;
  MPI_Comm_size(MPI_COMM_WORLD,&amp;amp;size);&lt;br /&gt;
  MPI_Get_processor_name(processor_name, &amp;amp;namelen);&lt;br /&gt;
  printf(&amp;quot;Hello MPI! Process %d of %d on %s\n&amp;quot;, rank, size, processor_name);&lt;br /&gt;
  MPI_Finalize();&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Before compiling, make sure that the compilers that are required available.&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
module list&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To avoid conflicts between libraries, the safest way is purging all modules:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
module purge&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The load both gcc and openmpi libraries. If modules were purged, then slurm needs to be reloaded too.&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
module load gcc/4.8.1 openmpi/gcc/64/1.6.5 slurm/2.5.7&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Compile the &amp;lt;code&amp;gt;hello_mpi.c&amp;lt;/code&amp;gt; code.&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
mpicc hello_mpi.c -o test_hello_world&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If desired, a list of libraries compiled into the executable can be viewed:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
ldd test_hello_world&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  linux-vdso.so.1 =&amp;gt;  (0x00002aaaaaacb000)&lt;br /&gt;
  libmpi.so.1 =&amp;gt; /shared/apps/openmpi/gcc/64/1.6.5/lib64/libmpi.so.1 (0x00002aaaaaccd000)&lt;br /&gt;
  libdl.so.2 =&amp;gt; /lib64/libdl.so.2 (0x00002aaaab080000)&lt;br /&gt;
  libm.so.6 =&amp;gt; /lib64/libm.so.6 (0x00002aaaab284000)&lt;br /&gt;
  libnuma.so.1 =&amp;gt; /usr/lib64/libnuma.so.1 (0x0000003e29400000)&lt;br /&gt;
  librt.so.1 =&amp;gt; /lib64/librt.so.1 (0x00002aaaab509000)&lt;br /&gt;
  libnsl.so.1 =&amp;gt; /lib64/libnsl.so.1 (0x00002aaaab711000)&lt;br /&gt;
  libutil.so.1 =&amp;gt; /lib64/libutil.so.1 (0x00002aaaab92a000)&lt;br /&gt;
  libpthread.so.0 =&amp;gt; /lib64/libpthread.so.0 (0x00002aaaabb2e000)&lt;br /&gt;
  libc.so.6 =&amp;gt; /lib64/libc.so.6 (0x00002aaaabd4b000)&lt;br /&gt;
  /lib64/ld-linux-x86-64.so.2 (0x00002aaaaaaab000)&lt;br /&gt;
&lt;br /&gt;
Running the executable on two nodes, with four tasks per node, can be done like this:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
srun --nodes=2 --ntasks-per-node=4 --mpi=openmpi ./test_hello_world&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This will result in the following output:&lt;br /&gt;
  Hello MPI! Process 4 of 8 on node011&lt;br /&gt;
  Hello MPI! Process 1 of 8 on node010&lt;br /&gt;
  Hello MPI! Process 7 of 8 on node011&lt;br /&gt;
  Hello MPI! Process 6 of 8 on node011&lt;br /&gt;
  Hello MPI! Process 5 of 8 on node011&lt;br /&gt;
  Hello MPI! Process 2 of 8 on node010&lt;br /&gt;
  Hello MPI! Process 0 of 8 on node010&lt;br /&gt;
  Hello MPI! Process 3 of 8 on node010&lt;br /&gt;
&lt;br /&gt;
== A mvapich2 sbatch example ==&lt;br /&gt;
A mpi job using mvapich2 on 32 cores, using the normal compute nodes and the fast infiniband interconnect for RDMA traffic.&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
$ module load mvapich2/gcc&lt;br /&gt;
$ vim batch.sh&lt;br /&gt;
 #!/bin/sh&lt;br /&gt;
 #SBATCH --comment=projectx&lt;br /&gt;
 #SBATCH --time=30-0&lt;br /&gt;
 #SBATCH  -n 32&lt;br /&gt;
 #SBATCH --constraint=4gpercpu&lt;br /&gt;
 #SBATCH --output=output_%j.txt&lt;br /&gt;
 #SBATCH --error=error_output_%j.txt&lt;br /&gt;
 #SBATCH --job-name=MPItest&lt;br /&gt;
 #SBATCH --mail-type=ALL&lt;br /&gt;
 #SBATCH --mail-user=user@wur.nl&lt;br /&gt;
 &lt;br /&gt;
 echo &amp;quot;Starting at `date`&amp;quot;&lt;br /&gt;
 echo &amp;quot;Running on hosts: $SLURM_NODELIST&amp;quot;&lt;br /&gt;
 echo &amp;quot;Running on $SLURM_NNODES nodes.&amp;quot;&lt;br /&gt;
 echo &amp;quot;Running on $SLURM_NPROCS processors.&amp;quot;&lt;br /&gt;
 echo &amp;quot;Current working directory is `pwd`&amp;quot;&lt;br /&gt;
 # echo &amp;quot;Env var MPIR_CVAR_NEMESIS_TCP_NETWORK_IFACE is $MPIR_CVAR_NEMESIS_TCP_NETWORK_IFACE&amp;quot;&lt;br /&gt;
 # export MPIR_CVAR_NEMESIS_TCP_NETWORK_IFACE=ib0&lt;br /&gt;
&lt;br /&gt;
 mpirun -iface ib0 -np 32 ./tmf_par.out -NX 480 -NY 240 -alpha  11 -chi 1.3 -psi_b 5e-2  -beta  0.0 -zeta 3.5 -kT 0.10 &lt;br /&gt;
&lt;br /&gt;
 echo &amp;quot;Program finished with exit code $? at: `date`&amp;quot;&lt;br /&gt;
&lt;br /&gt;
$ sbatch batch.sh&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=MPI_on_B4F_cluster&amp;diff=2236</id>
		<title>MPI on B4F cluster</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=MPI_on_B4F_cluster&amp;diff=2236"/>
		<updated>2023-06-16T09:01:42Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== A simple &#039;Hello World&#039; example ==&lt;br /&gt;
Consider the following simple MPI version, in C, of the &#039;Hello World&#039; example:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source lang=&#039;cpp&#039;&amp;gt;&lt;br /&gt;
#include &amp;lt;stdio.h&amp;gt;&lt;br /&gt;
#include &amp;lt;mpi.h&amp;gt;&lt;br /&gt;
int main(int argc, char ** argv) {&lt;br /&gt;
  int size,rank,namelen;&lt;br /&gt;
  char processor_name[MPI_MAX_PROCESSOR_NAME];&lt;br /&gt;
  MPI_Init(&amp;amp;argc, &amp;amp;argv);&lt;br /&gt;
  MPI_Comm_rank(MPI_COMM_WORLD,&amp;amp;rank);&lt;br /&gt;
  MPI_Comm_size(MPI_COMM_WORLD,&amp;amp;size);&lt;br /&gt;
  MPI_Get_processor_name(processor_name, &amp;amp;namelen);&lt;br /&gt;
  printf(&amp;quot;Hello MPI! Process %d of %d on %s\n&amp;quot;, rank, size, processor_name);&lt;br /&gt;
  MPI_Finalize();&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Before compiling, make sure that the compilers that are required available.&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
module list&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To avoid conflicts between libraries, the safest way is purging all modules:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
module purge&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The load both gcc and openmpi libraries. If modules were purged, then slurm needs to be reloaded too.&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
module load gcc/4.8.1 openmpi/gcc/64/1.6.5 slurm/2.5.7&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Compile the &amp;lt;code&amp;gt;hello_mpi.c&amp;lt;/code&amp;gt; code.&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
mpicc hello_mpi.c -o test_hello_world&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If desired, a list of libraries compiled into the executable can be viewed:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
ldd test_hello_world&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  linux-vdso.so.1 =&amp;gt;  (0x00002aaaaaacb000)&lt;br /&gt;
  libmpi.so.1 =&amp;gt; /shared/apps/openmpi/gcc/64/1.6.5/lib64/libmpi.so.1 (0x00002aaaaaccd000)&lt;br /&gt;
  libdl.so.2 =&amp;gt; /lib64/libdl.so.2 (0x00002aaaab080000)&lt;br /&gt;
  libm.so.6 =&amp;gt; /lib64/libm.so.6 (0x00002aaaab284000)&lt;br /&gt;
  libnuma.so.1 =&amp;gt; /usr/lib64/libnuma.so.1 (0x0000003e29400000)&lt;br /&gt;
  librt.so.1 =&amp;gt; /lib64/librt.so.1 (0x00002aaaab509000)&lt;br /&gt;
  libnsl.so.1 =&amp;gt; /lib64/libnsl.so.1 (0x00002aaaab711000)&lt;br /&gt;
  libutil.so.1 =&amp;gt; /lib64/libutil.so.1 (0x00002aaaab92a000)&lt;br /&gt;
  libpthread.so.0 =&amp;gt; /lib64/libpthread.so.0 (0x00002aaaabb2e000)&lt;br /&gt;
  libc.so.6 =&amp;gt; /lib64/libc.so.6 (0x00002aaaabd4b000)&lt;br /&gt;
  /lib64/ld-linux-x86-64.so.2 (0x00002aaaaaaab000)&lt;br /&gt;
&lt;br /&gt;
Running the executable on two nodes, with four tasks per node, can be done like this:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
srun --nodes=2 --ntasks-per-node=4 --mpi=openmpi ./test_hello_world&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This will result in the following output:&lt;br /&gt;
  Hello MPI! Process 4 of 8 on node011&lt;br /&gt;
  Hello MPI! Process 1 of 8 on node010&lt;br /&gt;
  Hello MPI! Process 7 of 8 on node011&lt;br /&gt;
  Hello MPI! Process 6 of 8 on node011&lt;br /&gt;
  Hello MPI! Process 5 of 8 on node011&lt;br /&gt;
  Hello MPI! Process 2 of 8 on node010&lt;br /&gt;
  Hello MPI! Process 0 of 8 on node010&lt;br /&gt;
  Hello MPI! Process 3 of 8 on node010&lt;br /&gt;
&lt;br /&gt;
== A mvapich2 sbatch example ==&lt;br /&gt;
A mpi job using mvapich2 on 32 cores, using the normal compute nodes and the fast infiniband interconnect for RDMA traffic.&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
$ module load mvapich2/gcc&lt;br /&gt;
$ vim batch.sh&lt;br /&gt;
 #!/bin/sh&lt;br /&gt;
 #SBATCH --comment=projectx&lt;br /&gt;
 #SBATCH --time=30-0&lt;br /&gt;
 #SBATCH  -n 32&lt;br /&gt;
 #SBATCH --constraint=4gpercpu&lt;br /&gt;
 #SBATCH --output=output_%j.txt&lt;br /&gt;
 #SBATCH --error=error_output_%j.txt&lt;br /&gt;
 #SBATCH --job-name=MPItest&lt;br /&gt;
 #SBATCH --mail-type=ALL&lt;br /&gt;
 #SBATCH --mail-user=user@wur.nl&lt;br /&gt;
 &lt;br /&gt;
 echo &amp;quot;Starting at `date`&amp;quot;&lt;br /&gt;
 echo &amp;quot;Running on hosts: $SLURM_NODELIST&amp;quot;&lt;br /&gt;
 echo &amp;quot;Running on $SLURM_NNODES nodes.&amp;quot;&lt;br /&gt;
 echo &amp;quot;Running on $SLURM_NPROCS processors.&amp;quot;&lt;br /&gt;
 echo &amp;quot;Current working directory is `pwd`&amp;quot;&lt;br /&gt;
 # echo &amp;quot;Env var MPIR_CVAR_NEMESIS_TCP_NETWORK_IFACE is $MPIR_CVAR_NEMESIS_TCP_NETWORK_IFACE&amp;quot;&lt;br /&gt;
 # export MPIR_CVAR_NEMESIS_TCP_NETWORK_IFACE=ib0&lt;br /&gt;
&lt;br /&gt;
 mpirun -iface ib0 -np 32 ./tmf_par.out -NX 480 -NY 240 -alpha  11 -chi 1.3 -psi_b 5e-2  -beta  0.0 -zeta 3.5 -kT 0.10 &lt;br /&gt;
&lt;br /&gt;
 echo &amp;quot;Program finished with exit code $? at: `date`&amp;quot;&lt;br /&gt;
&lt;br /&gt;
$ sbatch batch.sh&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=JBrowse&amp;diff=2235</id>
		<title>JBrowse</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=JBrowse&amp;diff=2235"/>
		<updated>2023-06-16T09:01:30Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=== Typical commands used to set up a JBrowse === &lt;br /&gt;
&lt;br /&gt;
Author: Martijn Derks&lt;br /&gt;
&lt;br /&gt;
* JBrowse is available for multiple species:&lt;br /&gt;
** https://jbrowse.hpcagrogenomics.wur.nl/pig/&lt;br /&gt;
** https://jbrowse.hpcagrogenomics.wur.nl/chicken/&lt;br /&gt;
** https://jbrowse.hpcagrogenomics.wur.nl/cattle/&lt;br /&gt;
** https://jbrowse.hpcagrogenomics.wur.nl/turkey/&lt;br /&gt;
** https://jbrowse.hpcagrogenomics.wur.nl/Cyprinus_carpio/&lt;br /&gt;
* Users are free to add usefull commands to this tutorial&lt;br /&gt;
&lt;br /&gt;
=== Install JBrowse ===&lt;br /&gt;
&lt;br /&gt;
Download the latest JBrowse here: http://jbrowse.org/&lt;br /&gt;
&lt;br /&gt;
Make a directory in &amp;lt;code&amp;gt;/shared/apps/jbrowse/&amp;lt;/code&amp;gt; for your species of interested (e.g. &amp;lt;code&amp;gt;mkdir Cyprinus_carpio&amp;lt;/code&amp;gt;). Move the downloaded JBrowse source files there. All further procedures detailed in this Wiki page assume working from that directory (NOTE: if your species of interest is already there, contact the maintainer of that JBrowse instance).&lt;br /&gt;
Run the setup script to install perl dependencies and required modules&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
unzip JBrowse-1.12.0.zip&lt;br /&gt;
mv JBrowse-1.12.0/* $PWD&lt;br /&gt;
./setup.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Add reference sequence ===&lt;br /&gt;
&lt;br /&gt;
Example code for chicken genome&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
bin/prepare-refseqs.pl --fasta /lustre/nobackup/WUR/ABGC/shared/public_data_store/genomes/chicken/Ensembl74/Gallus_gallus.Galgal4.74.dna.toplevel.fa&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To remove tracks use following command:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
bin/remove-track.pl -D --trackLabel &#039;trackname&#039;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Add annotation files (GFF/BED)===&lt;br /&gt;
&lt;br /&gt;
Data can be downloaded from the Ensembl FTP site: http://www.ensembl.org/info/data/ftp/index.html&lt;br /&gt;
&lt;br /&gt;
Add gene features:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
bin/flatfile-to-json.pl --key &amp;quot;Genes&amp;quot; --type gene --config &#039;{ &amp;quot;category&amp;quot;: &amp;quot;GalGal4.83 Annotation&amp;quot; }&#039; --trackLabel Genes --gff ../ensembl_data/Gallus_gallus.Galgal4.83.gff3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Add corresponding transcripts:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
bin/flatfile-to-json.pl --key &amp;quot;Transcripts&amp;quot; --className transcript --subfeatureClasses &#039;{&amp;quot;exon&amp;quot;: &amp;quot;exon&amp;quot;, &amp;quot;CDS&amp;quot;: &amp;quot;CDS&amp;quot;, &amp;quot;five_prime_UTR&amp;quot;: &amp;quot;five_prime_UTR&amp;quot;, &amp;quot;three_prime_UTR&amp;quot;: &amp;quot;three_prime_UTR&amp;quot;}&#039; --config &#039;{ &amp;quot;category&amp;quot;: &amp;quot;GalGal4.83 Annotation&amp;quot; }&#039; --type transcript --trackLabel Transcripts --gff ../ensembl_data/Gallus_gallus.Galgal4.83.gff3&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Alignment tracks (BAM)===&lt;br /&gt;
&lt;br /&gt;
You can load single BAM-files by following command:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
bin/add-bam-track --label &amp;lt;label&amp;gt; --bam_url &amp;lt;url&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To load multiple BAM files present in a certain directory use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
for bam in /&amp;lt;dir&amp;gt;*.bam; do&lt;br /&gt;
        ln -s $bam track_symlinks/ ## Make symlinks from the BAM files&lt;br /&gt;
        ln -s $bam.bai track_symlinks/ ## Make symlinks to the BAM index files&lt;br /&gt;
        tissue=`echo $bam | rev | cut -c 5- | cut -d&#039;/&#039; -f1 | rev` ## USe the name of the file without .bam as trackLabel&lt;br /&gt;
        &lt;br /&gt;
        ## Add BAM in alignment mode (Alignments2)&lt;br /&gt;
        echo &#039;{&lt;br /&gt;
                &amp;quot;label&amp;quot; : &amp;quot;&#039;${tissue}&#039;_alignment&amp;quot;,&lt;br /&gt;
                &amp;quot;key&amp;quot; : &amp;quot;&#039;${tissue}&#039;_alignment&amp;quot;,&lt;br /&gt;
                &amp;quot;storeClass&amp;quot; : &amp;quot;JBrowse/Store/SeqFeature/BAM&amp;quot;,&lt;br /&gt;
                &amp;quot;urlTemplate&amp;quot; : &amp;quot;../track_symlinks/&#039;${tissue}&#039;&amp;quot;,&lt;br /&gt;
                &amp;quot;category&amp;quot; : &amp;quot;3. RNA-seq alignments&amp;quot;,&lt;br /&gt;
                &amp;quot;type&amp;quot; : &amp;quot;Alignments2&amp;quot;&lt;br /&gt;
        }&#039; | bin/add-track-json.pl data/trackList.json&lt;br /&gt;
&lt;br /&gt;
        ## Add BAM in coverage mode (SNPCoverage)&lt;br /&gt;
        echo &#039;{&lt;br /&gt;
                &amp;quot;label&amp;quot; : &amp;quot;&#039;${tissue}&#039;_coverage&amp;quot;,&lt;br /&gt;
                &amp;quot;key&amp;quot; : &amp;quot;&#039;${tissue}&#039;_coverage&amp;quot;,&lt;br /&gt;
                &amp;quot;storeClass&amp;quot; : &amp;quot;JBrowse/Store/SeqFeature/BAM&amp;quot;,&lt;br /&gt;
                &amp;quot;urlTemplate&amp;quot; : &amp;quot;../track_symlinks/&#039;${tissue}&#039;&amp;quot;,&lt;br /&gt;
                &amp;quot;category&amp;quot; : &amp;quot;3. RNA-seq alignments&amp;quot;,&lt;br /&gt;
                &amp;quot;type&amp;quot; : &amp;quot;SNPCoverage&amp;quot;&lt;br /&gt;
        }&#039; | bin/add-track-json.pl data/trackList.json&lt;br /&gt;
&lt;br /&gt;
done&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Make sure the BAM file can be read by a everybody if not use:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
chmod +r &amp;lt;BAM_file&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Make sure that all directoryies in the full path of the BAMfile are executable:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
chmod +x &amp;lt;dir&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Variant tracks (VCF)===&lt;br /&gt;
&lt;br /&gt;
To load a VCF file in JBrowse make sure the file is gzipped and indexed&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
tabix -p vcf Gallus_gallus_incl_consequences.vcf.gz&lt;br /&gt;
&lt;br /&gt;
echo &#039; {&lt;br /&gt;
       &amp;quot;label&amp;quot; : &amp;quot;Gallus_gallus_incl_consequences&amp;quot;,&lt;br /&gt;
       &amp;quot;key&amp;quot; : &amp;quot;Gallus_gallus_incl_consequences&amp;quot;,&lt;br /&gt;
       &amp;quot;storeClass&amp;quot; : &amp;quot;JBrowse/Store/SeqFeature/VCFTabix&amp;quot;,&lt;br /&gt;
       &amp;quot;urlTemplate&amp;quot; : &amp;quot;../../ensembl_data/VCF/Gallus_gallus_incl_consequences.vcf.gz&amp;quot;,&lt;br /&gt;
       &amp;quot;category&amp;quot; : &amp;quot;2. Variants&amp;quot;,&lt;br /&gt;
       &amp;quot;type&amp;quot; : &amp;quot;HTMLVariants&amp;quot;&lt;br /&gt;
     } &#039; | bin/add-track-json.pl data/trackList.json&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Wiggle/BigWig tracks (WIG)===&lt;br /&gt;
&lt;br /&gt;
You can load single BigWig-files by following command:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
bin/add-bw-track --label &amp;lt;label&amp;gt; --bw_url &amp;lt;url&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Evidence tracks===&lt;br /&gt;
&lt;br /&gt;
Evidence tracks can be loaded in bed, gff and gbk format using &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
bin/flatfile-to-json.pl&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Examples are given above.&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Whole_genome_alignment_pipeline&amp;diff=2234</id>
		<title>Whole genome alignment pipeline</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Whole_genome_alignment_pipeline&amp;diff=2234"/>
		<updated>2023-06-16T09:01:16Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Author: Carolina Pita Barros &amp;lt;br /&amp;gt; &lt;br /&gt;
Contact: carolina.pitabarros@wur.nl  &amp;lt;br /&amp;gt;&lt;br /&gt;
ABG&amp;lt;br /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
For up-to-date documentation see [https://github.com/CarolinaPB/whole-genome-alignment here]&lt;br /&gt;
&lt;br /&gt;
= Whole genome alignment pipeline =&lt;br /&gt;
&lt;br /&gt;
Path to pipeline: /lustre/nobackup/WUR/ABGC/shared/PIPELINES/whole-genome-alignment&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;first-follow-the-instructions-here&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
== First follow the instructions here: ==&lt;br /&gt;
&lt;br /&gt;
[https://carolinapb.github.io/2021-06-23-how-to-run-my-pipelines/ Step by step guide on how to use my pipelines]&amp;lt;br /&amp;gt;&lt;br /&gt;
Click [https://github.com/CarolinaPB/snakemake-template/blob/master/Short%20introduction%20to%20Snakemake.pdf here] for an introduction to Snakemake&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;about&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
== ABOUT ==&lt;br /&gt;
&lt;br /&gt;
This pipeline aligns one or more genomes to a specified genome and plots the alignment.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;tools-used&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
==== Tools used: ====&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/lh3/minimap2 minimap2]&lt;br /&gt;
* R&lt;br /&gt;
** [https://github.com/tpoorten/dotPlotly/blob/master/pafCoordsDotPlotly.R pafCoordsDotPlotly.R]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
!align=&amp;quot;center&amp;quot;| [[File:whole-genome-alignment-workflow.png]]&lt;br /&gt;
|-&lt;br /&gt;
|align=&amp;quot;center&amp;quot;| &#039;&#039;Pipeline workflow&#039;&#039;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;edit-config.yaml-with-the-paths-to-your-files&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
=== Edit config.yaml with the paths to your files ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;syntaxhighlight lang=&amp;quot;yaml&amp;quot;&amp;gt;# genome alignment parameters:&lt;br /&gt;
GENOME: /path/to/genome #genome fasta to be compared&lt;br /&gt;
COMPARISON_GENOME: &lt;br /&gt;
  &amp;lt;genome1&amp;gt;: path/to/genome1.fasta&lt;br /&gt;
  &amp;lt;genome2&amp;gt;: path/to/genome2.fasta&lt;br /&gt;
  &amp;lt;genome3&amp;gt;: path/to/genome3.fasta&lt;br /&gt;
&lt;br /&gt;
# filter alignments less than cutoff X bp&lt;br /&gt;
MIN_ALIGNMENT_LENGTH: 10000&lt;br /&gt;
MIN_QUERY_LENGTH: 50000&lt;br /&gt;
&lt;br /&gt;
PREFIX: &amp;lt;prefix&amp;gt;&lt;br /&gt;
&lt;br /&gt;
OUTDIR: /path/to/outdir&amp;lt;/syntaxhighlight&amp;gt;&lt;br /&gt;
* GENOME: path to the genome fasta file (can be compressed). This is the genome that you want to be compared to all the others&lt;br /&gt;
* COMPARISON_GENOME: genome fasta (can be compressed) for whole genome comparison. Add your species name and the path to the fasta file. ex: chicken: /path/to/chicken.fna.gz. You can add several genomes, one on each line.&lt;br /&gt;
* MIN_ALIGNMENT_LENGTH and MIN_QUERY_LENGTH - parameters for plotting. If your plot is coming out blank or if there’s an error with the plotting step, try lowering these thresholds. This happens because the alignments are not large enough.&lt;br /&gt;
* PREFIX: name of your species (ex: turkey)&lt;br /&gt;
* OUTDIR: directory where snakemake will run and where the results will be written to&lt;br /&gt;
&lt;br /&gt;
If you want the results to be written to this directory (not to a new directory), comment out or remove&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;OUTDIR: /path/to/outdir&amp;lt;/pre&amp;gt;&lt;br /&gt;
&amp;lt;span id=&amp;quot;additional-set-up&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
== ADDITIONAL SET UP ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;installing-r-packages&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
=== Installing R packages ===&lt;br /&gt;
&lt;br /&gt;
First load R: &amp;lt;code&amp;gt;module load R/4.0.2&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Enter the R environment by writing &amp;lt;code&amp;gt;R&amp;lt;/code&amp;gt; and clicking enter. Install the packages:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;list.of.packages &amp;amp;lt;- c(&amp;amp;quot;optparse&amp;amp;quot;, &amp;amp;quot;data.table&amp;amp;quot;, &amp;amp;quot;ggplot2&amp;amp;quot;)&lt;br /&gt;
new.packages &amp;amp;lt;- list.of.packages[!(list.of.packages %in% installed.packages()[,&amp;amp;quot;Package&amp;amp;quot;])]&lt;br /&gt;
if(length(new.packages)) install.packages(new.packages)&amp;lt;/pre&amp;gt;&lt;br /&gt;
If you get an error like this:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;Warning in install.packages(new.packages) :&lt;br /&gt;
&#039;lib = &amp;amp;quot;/shared/apps/R/3.6.2/lib64/R/library&amp;amp;quot;&#039; is not writable&amp;lt;/pre&amp;gt;&lt;br /&gt;
Follow the instructions on how to install R packages locally [https://wiki.anunna.wur.nl/index.php/Installing_R_packages_locally here] and try to install the packages again.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;results&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
== RESULTS ==&lt;br /&gt;
&lt;br /&gt;
The most important files and directories are:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;run_date&amp;gt;_files.txt&#039;&#039;&#039; dated file with an overview of the files used to run the pipeline (for documentation purposes)&lt;br /&gt;
* &#039;&#039;&#039;genome_alignment/{prefix}&#039;&#039;vs&#039;&#039;{species}.paf&#039;&#039;&#039; paf format file w&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Get_my_bill&amp;diff=2233</id>
		<title>Get my bill</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Get_my_bill&amp;diff=2233"/>
		<updated>2023-06-16T09:01:06Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;To estimate costs over a certain time-period the following script can be invoked:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
module load anunna&lt;br /&gt;
get_my_bill&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The script will, by default, report the cost over the current month, until present. The output will look similar to this:&lt;br /&gt;
 &lt;br /&gt;
  !!! These results are advisory only !!!&lt;br /&gt;
  User: user001&lt;br /&gt;
  Currently run jobs this month: 1038&lt;br /&gt;
  Total cost so far: 25.19 EUR&lt;br /&gt;
  For account: 12345&lt;br /&gt;
  Jobs: 0 Cost: 0.00 EUR&lt;br /&gt;
  For account: project2&lt;br /&gt;
  Jobs: 104 Cost: 4.61 EUR&lt;br /&gt;
  For account: 56789&lt;br /&gt;
  Jobs: 74 Cost: 6.59 EUR&lt;br /&gt;
  For account: project 4&lt;br /&gt;
  Jobs: 19 Cost: 2.09 EUR&lt;br /&gt;
  For account: project5&lt;br /&gt;
  Jobs: 80 Cost: 0.04 EUR&lt;br /&gt;
  For account: project6&lt;br /&gt;
  Jobs: 738 Cost: 11.86 EUR&lt;br /&gt;
  For account: project7&lt;br /&gt;
  Jobs: 1 Cost: 0.01 EUR&lt;br /&gt;
  For account: project8&lt;br /&gt;
  Jobs: 22 Cost: 0.00 EUR&lt;br /&gt;
  Type            Time                                Current Use Current Cost EUR                &lt;br /&gt;
  home            2015-02-13 23:54:53                    7.131 GB 0.00                            &lt;br /&gt;
  backup          2015-02-06 06:02:43                    4.000 kB 0.00                            &lt;br /&gt;
  nobackup        2015-02-10 10:45:29                    5.348 TB 29.48                           &lt;br /&gt;
  scratch         2015-02-08 13:48:40                    0.233 TB 0.00                            &lt;br /&gt;
  Total this month: 54.67 EUR&lt;br /&gt;
&lt;br /&gt;
The script provides options in addition to the defaults, which can be invoked as follows:&lt;br /&gt;
Options:&lt;br /&gt;
  -h - Show this help message&lt;br /&gt;
  -g - Show results for your entire group&lt;br /&gt;
  -d [disk|compute] - get extra detail on disk/compute usages&lt;br /&gt;
  -t YYYY-MM - Show results for specific month&lt;br /&gt;
&lt;br /&gt;
== See also ==&lt;br /&gt;
* [[Main_Page | Main page AgHPC Wiki]]&lt;br /&gt;
* [[Main_Page#Controlling_costs | Controlling costs @ AgHPC Wiki]]&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Hadoop&amp;diff=2232</id>
		<title>Hadoop</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Hadoop&amp;diff=2232"/>
		<updated>2023-06-16T09:00:40Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Hadoop is another kind of architecture to promote writing code that can be executed in a parallel fashion. It&#039;s become more of an environment than a piece of software these days, as it typically contains:&lt;br /&gt;
&lt;br /&gt;
* HBase, flat table storage&lt;br /&gt;
* Hive, a form of data warehousing&lt;br /&gt;
* MapReduce, a style of coding to promote largely non-interacting workflows in order to leverage slow interconnects between workers&lt;br /&gt;
* HDFS, a distributed filesystem between workers&lt;br /&gt;
And many other components.&lt;br /&gt;
&lt;br /&gt;
This install of Hadoop has been optimised for executing a standalone HDFS filesystem, primarily for teaching purposes.&lt;br /&gt;
&lt;br /&gt;
== WARNING ==&lt;br /&gt;
HDFS is incredibly insecure. As soon as anyone knows the location of your nameserver, they may spoof into it to access your data, as any user. This can be prevented by using Kerberos tickets, however that is currently beyond our level of complexity. Do not store anything you care about here.&lt;br /&gt;
&lt;br /&gt;
=== Usage ===&lt;br /&gt;
load up the module, and java:&lt;br /&gt;
&amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
module load hadoop&lt;br /&gt;
module load java&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This will set your environment variables correctly.&lt;br /&gt;
&lt;br /&gt;
There is an example submission script at:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;/shared/apps/hadoop/current/wur/example_sbatch.sh&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
Copy this to your preferred location and submit it.&lt;br /&gt;
&lt;br /&gt;
Edit this script and adjust the settings accordingly. You can override the location that HDFS will store its blocks by altering the environment variable HADOOP_TMP_DIR, which will default to /lustre/scratch.&lt;br /&gt;
&lt;br /&gt;
Once it&#039;s run, the output log will tell you the location. I&#039;d recommend setting a static location in your job script so that it&#039;s reproducibly located, but the HADOOP_CONF_DIR will point to the correct fs.defaultFS once you&#039;ve started it.&lt;br /&gt;
&lt;br /&gt;
=== Mechanism details ===&lt;br /&gt;
The example script will run an executable called start-hdfs.sh. This simply starts by creating ~/.hadoop/$SLURM_JOB_ID, then assigning that as HADOOP_LOG_DIR. Next, it creates a subfolder conf, assigns that to HADOOP_CONF_DIR, and writes the core-site.xml and hdfs-site.xml files into it. Last, it creates a symlink from ~/.hadoop/current to point to your currently running HDFS instance. On the client side, all HADOOP_CONF_DIR&#039;s point to ~/.hadoop/current/conf so that whichever job is submitted, this symlink will point to the currently running config for you.&lt;br /&gt;
&lt;br /&gt;
If you&#039;re trying to share your HDFS instance with other users, you should:&lt;br /&gt;
* Copy the ~/.hadoop/current/conf/* to somewhere shared&lt;br /&gt;
* Tell them to set the environment variable HADOOP_CONF_DIR to that location&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Population_structural_variant_calling_pipeline&amp;diff=2231</id>
		<title>Population structural variant calling pipeline</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Population_structural_variant_calling_pipeline&amp;diff=2231"/>
		<updated>2023-06-16T09:00:29Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Author: Carolina Pita Barros &amp;lt;br /&amp;gt; &lt;br /&gt;
Contact: carolina.pitabarros@wur.nl  &amp;lt;br /&amp;gt;&lt;br /&gt;
ABG&amp;lt;br /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[https://github.com/CarolinaPB/population-structural-var-calling-smoove/tree/single_run Link to the repository]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== First follow the instructions here: ==&lt;br /&gt;
&lt;br /&gt;
[https://carolinapb.github.io/2021-06-23-how-to-run-my-pipelines/ Step by step guide on how to use my pipelines]&amp;lt;br /&amp;gt;&lt;br /&gt;
Click [https://github.com/CarolinaPB/snakemake-template/blob/master/Short%20introduction%20to%20Snakemake.pdf here] for an introduction to Snakemake&lt;br /&gt;
&lt;br /&gt;
== ABOUT ==&lt;br /&gt;
&lt;br /&gt;
This is a pipeline to perform structural variant calling in a population using Smoove. It also runs VEP and performs PCA. In addition to the VCF with the SVs, you also get a .tsv file with some summarized information on the SVs: it includes allele frequency per population, as well as VEP annotation and depth fold change as described in [https://github.com/brentp/duphold duphold]:  &amp;lt;br /&amp;gt; &lt;br /&gt;
&amp;amp;gt; DHBFC: fold-change for the variant depth relative to bins in the genome with similar GC-content.&amp;lt;br /&amp;gt;  &lt;br /&gt;
&amp;amp;gt; DHFFC: fold-change for the variant depth relative to Flanking regions.&lt;br /&gt;
&lt;br /&gt;
==== Tools used: ====&lt;br /&gt;
&lt;br /&gt;
* Smoove - SV calling&lt;br /&gt;
* VEP - determines the effect of the variants&lt;br /&gt;
* Plink - perform PCA&lt;br /&gt;
* R - plot PCA&lt;br /&gt;
* SURVIVOR - basic SV stats&lt;br /&gt;
* Python  &lt;br /&gt;
** PyVcf - add depth to vcf and create final table&lt;br /&gt;
** bamgroupreads.py + samblaster - create bam files with split and discordant reads&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
!align=&amp;quot;center&amp;quot;| [[File:Pop-sv-calling-workflow.png]]&lt;br /&gt;
|-&lt;br /&gt;
|align=&amp;quot;center&amp;quot;| &#039;&#039;Pipeline workflow&#039;&#039;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Edit config.yaml with the paths to your files ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;OUTDIR: /path/to/output &lt;br /&gt;
READS_DIR: /path/to/reads/ # don&#039;t add the reads files, just the directory where they are&lt;br /&gt;
SAMPLE_LIST: /path/to/file&lt;br /&gt;
REFERENCE: /path/to/assembly&lt;br /&gt;
CONTIGS_IGNORE: /path/to/file&lt;br /&gt;
SPECIES: &amp;amp;lt;species_name&amp;amp;gt;&lt;br /&gt;
PREFIX: &amp;amp;lt;output name&amp;amp;gt;&lt;br /&gt;
NUM_CHRS: &amp;amp;lt;number of chromosomes&amp;amp;gt;&lt;br /&gt;
BWA_MEM_M: &amp;amp;lt;Y/N&amp;amp;gt;&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* OUTDIR - directory where snakemake will run and where the results will be written to&lt;br /&gt;
* READS_DIR - path to the directory that contains the reads&lt;br /&gt;
* SAMPLE_LIST - three column csv with the sample name, name of the bam files to use in the second column and the name of the corresponding population on the third column. These bams should all be in the same directory (READS_DIR)&lt;br /&gt;
* Example: &lt;br /&gt;
&amp;lt;blockquote&amp;gt;sample1,sample1.bam,Pop1&amp;lt;br /&amp;gt;&lt;br /&gt;
sample2,sample2.bam,Pop1&amp;lt;br /&amp;gt;&lt;br /&gt;
sample3,sample3.bam,Pop2&amp;lt;br /&amp;gt;&lt;br /&gt;
sample4,sample4.bam,Pop2&amp;lt;br /&amp;gt;&amp;lt;/blockquote&amp;gt;&lt;br /&gt;
Tip: use the name of the bam file without the .bam extension as the sample name. Ex: from sample1.bam to sample1&lt;br /&gt;
* REFERENCE - path to the assembly file&lt;br /&gt;
* CONTIGS_IGNORE - contigs to be excluded from SV calling (usually the small contigs)&lt;br /&gt;
** If you don&#039;t want to exclude contigs you&#039;ll need to edit the Snakefile to remove this line &amp;lt;code&amp;gt;--excludechroms {params.contigs} \&amp;lt;/code&amp;gt;&lt;br /&gt;
* SPECIES - species name to be used for VEP&lt;br /&gt;
* PREFIX - prefix for the created files&lt;br /&gt;
* NUM_CHRS - number of chromosomes for your species (necessary for plink). ex: 38&lt;br /&gt;
* BWA_MEM_M - if you mapped your reads with `bwa mem` using the `-M` parameter and you want split read support in your VCF you need to run an extra step. For this write `Y`.  &lt;br /&gt;
For a more detailed explanation see here [https://carolinapb.github.io/2021-10-28-smoove-SR-support/ Smoove SR support]  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If you want the results to be written to this directory (not to a new directory), comment out or remove&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;OUTDIR: /path/to/outdir&amp;lt;/pre&amp;gt;&lt;br /&gt;
== ADDITIONAL SET UP ==&lt;br /&gt;
=== Configuring VEP ===&lt;br /&gt;
This pipeline uses VEP in offline mode, which increases performance. In order to use it in this mode, the cache for the species used needs to be installed: &lt;br /&gt;
Check if the cache file for your species already exist in &amp;lt;code&amp;gt;/lustre/nobackup/SHARED/cache/&amp;lt;/code&amp;gt;. If it doesn’t, create it with&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;/usr/bin/perl /shared/apps/SHARED/ensembl-vep/INSTALL.pl --CACHEDIR /lustre/nobackup/SHARED/cache/ --AUTO c -n --SPECIES &amp;amp;lt;species&amp;amp;gt;&amp;lt;/pre&amp;gt;&lt;br /&gt;
When multiple assemblies are found you need to run it again with &amp;lt;code&amp;gt;--ASSEMBLY &amp;amp;lt;assembly name&amp;amp;gt;&amp;lt;/code&amp;gt;, where “assembly name” is the name of the assembly you want to use.&lt;br /&gt;
&lt;br /&gt;
==== Other option: ====&lt;br /&gt;
&lt;br /&gt;
You can install VEP with&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;conda install -c bioconda ensembl-vep&amp;lt;/pre&amp;gt;&lt;br /&gt;
and install the cache with&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;vep_install --CACHEDIR &amp;amp;lt;where/to/install/cache&amp;amp;gt; --AUTO c -n --SPECIES &amp;amp;lt;species&amp;amp;gt;&amp;lt;/pre&amp;gt;&lt;br /&gt;
When multiple assemblies are found you need to run it again with &amp;lt;code&amp;gt;--ASSEMBLY &amp;amp;lt;assembly name&amp;amp;gt;&amp;lt;/code&amp;gt;, where “assembly name” is the name of the assembly you want to use.&lt;br /&gt;
&lt;br /&gt;
In the Snakefile, in rule &amp;lt;code&amp;gt;run_vep&amp;lt;/code&amp;gt;, replace &amp;lt;code&amp;gt;/shared/apps/SHARED/ensembl-vep/vep&amp;lt;/code&amp;gt; with &amp;lt;code&amp;gt;vep&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installing R packages ===&lt;br /&gt;
First load R:&lt;br /&gt;
&amp;lt;pre&amp;gt;module load R/3.6.2&amp;lt;/pre&amp;gt;&lt;br /&gt;
Enter the R environment by writing &amp;lt;pre&amp;gt;R&amp;lt;/pre&amp;gt; and clicking enter.  &lt;br /&gt;
Install the packages:&lt;br /&gt;
&amp;lt;pre&amp;gt;list.of.packages &amp;lt;- c(&amp;quot;optparse&amp;quot;, &amp;quot;data.table&amp;quot;, &amp;quot;ggplot2&amp;quot;)&lt;br /&gt;
new.packages &amp;lt;- list.of.packages[!(list.of.packages %in% installed.packages()[,&amp;quot;Package&amp;quot;])]&lt;br /&gt;
if(length(new.packages)) install.packages(new.packages)&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If you get an error like this:&lt;br /&gt;
&amp;lt;pre&amp;gt;Warning in install.packages(new.packages) :&lt;br /&gt;
  &#039;lib = &amp;quot;/shared/apps/R/3.6.2/lib64/R/library&amp;quot;&#039; is not writable&amp;lt;/pre&amp;gt;&lt;br /&gt;
Follow the instructions on how to install R packages locally [https://wiki.anunna.wur.nl/index.php/Installing_R_packages_locally here]  and try to install the packages again. &lt;br /&gt;
&lt;br /&gt;
== RESULTS ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;run_date&amp;gt;_files.txt&#039;&#039;&#039; Dated file with an overview of the files used to run the pipeline (for documentation purposes)&lt;br /&gt;
* &#039;&#039;&#039;2_merged&#039;&#039;&#039;&lt;br /&gt;
** {prefix}.smoove-counts.html - shows a summary of the number of reads before and after filtering&lt;br /&gt;
* &#039;&#039;&#039;5_postprocessing&#039;&#039;&#039; directory that contains the final VCF file containing the structural variants found. This file has been annotated with VEP&lt;br /&gt;
** {prefix}.smoove.square.vep.vcf.gz - Final VCF - with VEP annotation, not filtered for quality&lt;br /&gt;
** {prefix}.smoove.square.vep.vcf.gz_summary.html - statistics from VEP&lt;br /&gt;
** {prefix}.nosex, {prefix}.log, {prefix}.eigenvec, {prefix}.eigenval - output files from the PCA&lt;br /&gt;
** {prefix}_DUP_DEL_INV_table.tsv - table with the most important information extracted from the VCF. Contains information about the SV, allele frequency for each population, VEP annotation and depth information. The variants have been filtered with Minimum Quality score = 30&lt;br /&gt;
** {prefix}_DUP_DEL_INV.vcf - vcf file with annotated duplications, deletions and inversions. It has been filtered with Minimum Quality score = 30 and the DEPTH* field was added&lt;br /&gt;
** {prefix}_BND.vcf - vcf file with variants annotated with BND&lt;br /&gt;
* &#039;&#039;&#039;6_metrics&#039;&#039;&#039; directory that contains general stats about the number of SVs found&lt;br /&gt;
* &#039;&#039;&#039;FIGURES&#039;&#039;&#039; directory that contains the PCA plot&lt;br /&gt;
&lt;br /&gt;
What you do with the results from this structural variant calling pipeline depends on your research question: a possible next step would be to explore the &#039;&#039;&#039;{prefix}_DUP_DEL_INV_table.tsv&#039;&#039;&#039; file and look at the largest SVs found (sort by &#039;&#039;SVLEN&#039;&#039;) or at a specific effect in the ANNOTATION column, such as “frameshift_variant”.&lt;br /&gt;
&lt;br /&gt;
See [https://m.ensembl.org/info/genome/variation/prediction/predicted_data.html VEP effect descriptions] for a short description of the effects annotated by VEP.&lt;br /&gt;
&lt;br /&gt;
-----&lt;br /&gt;
&lt;br /&gt;
*The &#039;&#039;&#039;DEPTH&#039;&#039;&#039; field in the vcf has six fields, corresponding to the average depth across all samples.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;DEPTH=(DHBFC_1/1, DHBFC_0/1, DHBFC_0/0, DHFFC_1/1, DHFFC_0/1, DHFFC_0/0)&amp;lt;/pre&amp;gt;&lt;br /&gt;
Depth fold change as described in [https://github.com/brentp/duphold duphold]: &lt;br /&gt;
&amp;lt;blockquote&amp;gt;DHBFC: fold-change for the variant depth relative to bins in the genome with similar GC-content.&amp;lt;br /&amp;gt;&lt;br /&gt;
DHFFC: fold-change for the variant depth relative to Flanking regions.&amp;lt;/blockquote&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These fields are also in the &amp;lt;code&amp;gt;{prefix}_DUP_DEL_INV_table.tsv&amp;lt;/code&amp;gt; file&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Maker_protocols_Pmajor&amp;diff=2230</id>
		<title>Maker protocols Pmajor</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Maker_protocols_Pmajor&amp;diff=2230"/>
		<updated>2023-06-16T09:00:11Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page describes the various rounds of [http://www.yandell-lab.org/software/maker.html Maker]-based annotations for the [http://en.wikipedia.org/wiki/Parus_major &#039;&#039;Parus major&#039;&#039; (Great Tit)] genome.&lt;br /&gt;
&lt;br /&gt;
== Round 1 == &lt;br /&gt;
=== Rationale ===&lt;br /&gt;
For this round no P. major-based ESTs were available. Zebrafinch (T. guttata) is the closest relative for which a reasonably complete gene-model set is available. As a first pass, it was decided to let gene predictions be driven by ab-inititio predictions rather than by Zebrafinch EST. &lt;br /&gt;
&lt;br /&gt;
=== Invoking maker script ===&lt;br /&gt;
Do not forget to load the &amp;lt;code&amp;gt;maker&amp;lt;/code&amp;gt; module:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
module load maker/2.28 &lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
script submitted by SLURM (&amp;lt;code&amp;gt;sbatch&amp;lt;/code&amp;gt; command):&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
#!/bin/bash&lt;br /&gt;
#SBATCH --time=48000&lt;br /&gt;
#SBATCH --nodes=1&lt;br /&gt;
#SBATCH --ntasks=16&lt;br /&gt;
#SBATCH --output=output_%j.txt&lt;br /&gt;
#SBATCH --error=error_output_%j.txt&lt;br /&gt;
#SBATCH --job-name=test_maker&lt;br /&gt;
#SBATCH --mail-type=ALL&lt;br /&gt;
#SBATCH --mail-user=hendrik-jan.megens@wur.nl&lt;br /&gt;
maker&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Maker settings ===&lt;br /&gt;
==== content of &amp;lt;code&amp;gt;maker_opts.ctl&amp;lt;/code&amp;gt; ====&lt;br /&gt;
 #-----Genome (these are always required)&lt;br /&gt;
  genome=Pam.fa #genome sequence (fasta file or fasta embeded in GFF3 file)&lt;br /&gt;
  organism_type=eukaryotic #eukaryotic or prokaryotic. Default is eukaryotic&lt;br /&gt;
  &lt;br /&gt;
  #-----Re-annotation Using MAKER Derived GFF3&lt;br /&gt;
  maker_gff= #MAKER derived GFF3 file&lt;br /&gt;
  est_pass=0 #use ESTs in maker_gff: 1 = yes, 0 = no&lt;br /&gt;
  altest_pass=0 #use alternate organism ESTs in maker_gff: 1 = yes, 0 = no&lt;br /&gt;
  protein_pass=0 #use protein alignments in maker_gff: 1 = yes, 0 = no&lt;br /&gt;
  rm_pass=0 #use repeats in maker_gff: 1 = yes, 0 = no&lt;br /&gt;
  model_pass=0 #use gene models in maker_gff: 1 = yes, 0 = no&lt;br /&gt;
  pred_pass=0 #use ab-initio predictions in maker_gff: 1 = yes, 0 = no&lt;br /&gt;
  other_pass=0 #passthrough anyything else in maker_gff: 1 = yes, 0 = no&lt;br /&gt;
  &lt;br /&gt;
  #-----EST Evidence (for best results provide a file for at least one)&lt;br /&gt;
  est= #set of ESTs or assembled mRNA-seq in fasta format&lt;br /&gt;
  altest= Taeniopygia_guttata.taeGut3.2.4.74.cdna.all.fa #EST/cDNA sequence file in fasta format from an alternate organism&lt;br /&gt;
  est_gff= #aligned ESTs or mRNA-seq from an external GFF3 file&lt;br /&gt;
  altest_gff= #aligned ESTs from a closly relate species in GFF3 format&lt;br /&gt;
  &lt;br /&gt;
  #-----Protein Homology Evidence (for best results provide a file for at least one)&lt;br /&gt;
  protein= Taeniopygia_guttata.taeGut3.2.4.74.pep.all.fa #protein sequence file in fasta format (i.e. from mutiple oransisms)&lt;br /&gt;
  protein_gff=  #aligned protein homology evidence from an external GFF3 file&lt;br /&gt;
  &lt;br /&gt;
  #-----Repeat Masking (leave values blank to skip repeat masking)&lt;br /&gt;
  model_org=Metazoa #select a model organism for RepBase masking in RepeatMasker&lt;br /&gt;
  rmlib= #provide an organism specific repeat library in fasta format for RepeatMasker&lt;br /&gt;
  repeat_protein= #provide a fasta file of transposable element proteins for RepeatRunner&lt;br /&gt;
  rm_gff= #pre-identified repeat elements from an external GFF3 file&lt;br /&gt;
  prok_rm=0 #forces MAKER to repeatmask prokaryotes (no reason to change this), 1 = yes, 0 = no&lt;br /&gt;
  softmask=1 #use soft-masking rather than hard-masking in BLAST (i.e. seg and dust filtering)&lt;br /&gt;
  &lt;br /&gt;
  #-----Gene Prediction&lt;br /&gt;
  snaphmm= /shared/apps/WUR/ABGC/snap/snap-2013-11-29/HMM/mam54.hmm #SNAP HMM file&lt;br /&gt;
  gmhmm= #GeneMark HMM file&lt;br /&gt;
  augustus_species= chicken #Augustus gene prediction species model&lt;br /&gt;
  fgenesh_par_file= #FGENESH parameter file&lt;br /&gt;
  pred_gff= #ab-initio predictions from an external GFF3 file&lt;br /&gt;
  model_gff= #annotated gene models from an external GFF3 file (annotation pass-through)&lt;br /&gt;
  est2genome=0 #infer gene predictions directly from ESTs, 1 = yes, 0 = no&lt;br /&gt;
  protein2genome=0 #infer predictions from protein homology, 1 = yes, 0 = no&lt;br /&gt;
  unmask=0 #also run ab-initio prediction programs on unmasked sequence, 1 = yes, 0 = no&lt;br /&gt;
  &lt;br /&gt;
  #-----Other Annotation Feature Types (features MAKER doesn&#039;t recognize)&lt;br /&gt;
  other_gff= #extra features to pass-through to final MAKER generated GFF3 file&lt;br /&gt;
  &lt;br /&gt;
  #-----External Application Behavior Options&lt;br /&gt;
  alt_peptide=C #amino acid used to replace non-standard amino acids in BLAST databases&lt;br /&gt;
  cpus=16 #max number of cpus to use in BLAST and RepeatMasker (not for MPI, leave 1 when using MPI)&lt;br /&gt;
  &lt;br /&gt;
  #-----MAKER Behavior Options&lt;br /&gt;
  max_dna_len=100000 #length for dividing up contigs into chunks (increases/decreases memory usage)&lt;br /&gt;
  min_contig=1 #skip genome contigs below this length (under 10kb are often useless)&lt;br /&gt;
  &lt;br /&gt;
  pred_flank=200 #flank for extending evidence clusters sent to gene predictors&lt;br /&gt;
  pred_stats=0 #report AED and QI statistics for all predictions as well as models&lt;br /&gt;
  AED_threshold=1 #Maximum Annotation Edit Distance allowed (bound by 0 and 1)&lt;br /&gt;
  min_protein=0 #require at least this many amino acids in predicted proteins&lt;br /&gt;
  alt_splice=0 #Take extra steps to try and find alternative splicing, 1 = yes, 0 = no&lt;br /&gt;
  always_complete=0 #extra steps to force start and stop codons, 1 = yes, 0 = no&lt;br /&gt;
  map_forward=0 #map names and attributes forward from old GFF3 genes, 1 = yes, 0 = no&lt;br /&gt;
  keep_preds=0 #Concordance threshold to add unsupported gene prediction (bound by 0 and 1)&lt;br /&gt;
  &lt;br /&gt;
  split_hit=10000 #length for the splitting of hits (expected max intron size for evidence alignments)&lt;br /&gt;
  single_exon=0 #consider single exon EST evidence when generating annotations, 1 = yes, 0 = no&lt;br /&gt;
  single_length=250 #min length required for single exon ESTs if &#039;single_exon is enabled&#039;&lt;br /&gt;
  correct_est_fusion=0 #limits use of ESTs in annotation to avoid fusion genes&lt;br /&gt;
  &lt;br /&gt;
  tries=2 #number of times to try a contig if there is a failure for some reason&lt;br /&gt;
  clean_try=1 #remove all data from previous run before retrying, 1 = yes, 0 = no&lt;br /&gt;
  clean_up=1 #removes theVoid directory with individual analysis files, 1 = yes, 0 = no&lt;br /&gt;
  TMP= #specify a directory other than the system default temporary directory for temporary files&lt;br /&gt;
&lt;br /&gt;
==== content of &amp;lt;code&amp;gt;maker_exe.ctl&amp;lt;/code&amp;gt; ====&lt;br /&gt;
  #-----Location of Executables Used by MAKER/EVALUATOR&lt;br /&gt;
  makeblastdb=/shared/apps/WUR/ABGC/blast/ncbi-blast-2.2.28+/bin/makeblastdb #location of NCBI+ makeblastdb executable&lt;br /&gt;
  blastn=/shared/apps/WUR/ABGC/blast/ncbi-blast-2.2.28+/bin/blastn #location of NCBI+ blastn executable&lt;br /&gt;
  blastx=/shared/apps/WUR/ABGC/blast/ncbi-blast-2.2.28+/bin/blastx #location of NCBI+ blastx executable&lt;br /&gt;
  tblastx=/shared/apps/WUR/ABGC/blast/ncbi-blast-2.2.28+/bin/tblastx #location of NCBI+ tblastx executable&lt;br /&gt;
  formatdb= #location of NCBI formatdb executable&lt;br /&gt;
  blastall= #location of NCBI blastall executable&lt;br /&gt;
  xdformat= #location of WUBLAST xdformat executable&lt;br /&gt;
  blasta= #location of WUBLAST blasta executable&lt;br /&gt;
  RepeatMasker=/shared/apps/WUR/ABGC/RepeatMasker/RepeatMasker-4-0-3/RepeatMasker #location of RepeatMasker executable&lt;br /&gt;
  exonerate=/shared/apps/WUR/ABGC/exonerate/exonerate-2.2.0-x86_64/bin/exonerate #location of exonerate executable&lt;br /&gt;
  &lt;br /&gt;
  #-----Ab-initio Gene Prediction Algorithms&lt;br /&gt;
  snap=/shared/apps/WUR/ABGC/snap/snap-2013-11-29/snap #location of snap executable&lt;br /&gt;
  gmhmme3= #location of eukaryotic genemark executable&lt;br /&gt;
  gmhmmp= #location of prokaryotic genemark executable&lt;br /&gt;
  augustus=/shared/apps/WUR/ABGC/augustus/augustus.2.7/src/augustus #location of augustus executable&lt;br /&gt;
  fgenesh= #location of fgenesh executable&lt;br /&gt;
  &lt;br /&gt;
  #-----Other Algorithms&lt;br /&gt;
  probuild= #location of probuild executable (required for genemark)&lt;br /&gt;
&lt;br /&gt;
==== contents of &amp;lt;code&amp;gt;maker_bopts.ctl&amp;lt;/code&amp;gt;==== &lt;br /&gt;
  #-----BLAST and Exonerate Statistics Thresholds&lt;br /&gt;
  blast_type=ncbi+ #set to &#039;ncbi+&#039;, &#039;ncbi&#039; or &#039;wublast&#039;&lt;br /&gt;
  &lt;br /&gt;
  pcov_blastn=0.8 #Blastn Percent Coverage Threhold EST-Genome Alignments&lt;br /&gt;
  pid_blastn=0.85 #Blastn Percent Identity Threshold EST-Genome Aligments&lt;br /&gt;
  eval_blastn=1e-10 #Blastn eval cutoff&lt;br /&gt;
  bit_blastn=40 #Blastn bit cutoff&lt;br /&gt;
  depth_blastn=0 #Blastn depth cutoff (0 to disable cutoff)&lt;br /&gt;
  &lt;br /&gt;
  pcov_blastx=0.5 #Blastx Percent Coverage Threhold Protein-Genome Alignments&lt;br /&gt;
  pid_blastx=0.4 #Blastx Percent Identity Threshold Protein-Genome Aligments&lt;br /&gt;
  eval_blastx=1e-06 #Blastx eval cutoff&lt;br /&gt;
  bit_blastx=30 #Blastx bit cutoff&lt;br /&gt;
  depth_blastx=0 #Blastx depth cutoff (0 to disable cutoff)&lt;br /&gt;
  &lt;br /&gt;
  pcov_tblastx=0.8 #tBlastx Percent Coverage Threhold alt-EST-Genome Alignments&lt;br /&gt;
  pid_tblastx=0.85 #tBlastx Percent Identity Threshold alt-EST-Genome Aligments&lt;br /&gt;
  eval_tblastx=1e-10 #tBlastx eval cutoff&lt;br /&gt;
  bit_tblastx=40 #tBlastx bit cutoff&lt;br /&gt;
  depth_tblastx=0 #tBlastx depth cutoff (0 to disable cutoff)&lt;br /&gt;
  &lt;br /&gt;
  pcov_rm_blastx=0.5 #Blastx Percent Coverage Threhold For Transposable Element Masking&lt;br /&gt;
  pid_rm_blastx=0.4 #Blastx Percent Identity Threshold For Transposbale Element Masking&lt;br /&gt;
  eval_rm_blastx=1e-06 #Blastx eval cutoff for transposable element masking&lt;br /&gt;
  bit_rm_blastx=30 #Blastx bit cutoff for transposable element masking&lt;br /&gt;
  &lt;br /&gt;
  ep_score_limit=20 #Exonerate protein percent of maximal score threshold&lt;br /&gt;
  en_score_limit=20 #Exonerate nucleotide percent of maximal score threshold&lt;br /&gt;
&lt;br /&gt;
== See also ==&lt;br /&gt;
[[Maker_2.2.8 | Maker pipeline as installed on Anunna]]&lt;br /&gt;
&lt;br /&gt;
== External links ==&lt;br /&gt;
* [http://www.yandell-lab.org/software/maker.html Maker homepage]&lt;br /&gt;
* [http://gmod.org/wiki/MAKER_Tutorial_2013 Maker tutorial]&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Filesystems&amp;diff=2229</id>
		<title>Filesystems</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Filesystems&amp;diff=2229"/>
		<updated>2023-06-16T08:59:44Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Anunna currently has multiple filesystem mounts that are available cluster-wide:&lt;br /&gt;
&lt;br /&gt;
== Global ==&lt;br /&gt;
* /home - This mount uses NFS to mount the home directories directly from nfs01. Each user has a 200G quota for this filesystem, as it is regularly backed up to tape, and can reliably be restored from up to a week&#039;s history.&lt;br /&gt;
&lt;br /&gt;
* /shared - This mount provides a consistent set of binaries for the entire cluster.&lt;br /&gt;
&lt;br /&gt;
* /lustre - This large mount uses the Lustre filesystem to provide files from multiple redundant servers. Access is provided per group, thus:&lt;br /&gt;
 /lustre/[level]/[partner]/[unit]&lt;br /&gt;
e.g.&lt;br /&gt;
 /lustre/backup/WUR/ABGC/&lt;br /&gt;
It comprises of three major parts (and some minor):&lt;br /&gt;
* /lustre/backup - In case of disaster, this data is stored a second time on a separate machine. Whilst this backup is purely in case of complete tragedy (such as some immense filesystem error, or multiple component failure), it can potentially be used to revert mistakes if you are very fast about reporting them. There is however no guarantee of this service.&lt;br /&gt;
* /lustre/nobackup - This is the &#039;normal&#039; filesystem for Lustre - no backups, just stored on the filesystem. Without having a backup needed, the cost of data here is not as much as under /lustre/backup, but in case of disaster cannot be recivered.&lt;br /&gt;
* /lustre/scratch - Files here may be removed after some time if the filesystem gets too full (Typically 30 days). You should tidy up this data yourself once work is complete.&lt;br /&gt;
* /lustre/shared - Same as /lustre/backup, except publicly available. This is where truly shared data lives that isn&#039;t assigned to a specific group.&lt;br /&gt;
&lt;br /&gt;
=== Private shared directories ===&lt;br /&gt;
If you are working with a group of users on a similar project, you might consider making a [[Shared_folders|Shared directory]] to coordinate. Information on how to do so is in the linked article.&lt;br /&gt;
&lt;br /&gt;
== Local ==&lt;br /&gt;
Specific to certain machines are some other filesystems that are available to you:&lt;br /&gt;
* /archive - an archive mount only accessible from the login nodes. Files here are sent to the Isilon for deeper storage. The cost of storing data here is much less than on the Lustre, but it cannot be used for compute work. This location is only available to WUR users. Files are able to be reverted via snapshot, and there is a separated backup, however this only comes in fortnightly (14 day) intervals.&lt;br /&gt;
&lt;br /&gt;
* /tmp - On each worker node there is a /tmp mount that can be used for temporary local caching. Be advised that you should clean this up, lest your files become a hindrance to other users. You can request a node with free space in your sbatch script like so:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
#SBATCH --tmp=&amp;lt;required space&amp;gt;&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* /dev/shm - On each worker you may also create a virtual filesystem directly into memory, for extremely fast data access. Be advised that this will count against the memory used for your job, but it is also the fastest available filesystem if needed.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== See also ==&lt;br /&gt;
* [[Tariffs | Costs associated with resource usage]]&lt;br /&gt;
&lt;br /&gt;
== External links ==&lt;br /&gt;
* [http://wiki.lustre.org/index.php/Main_Page Lustre website]&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Environment_Modules&amp;diff=2228</id>
		<title>Environment Modules</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Environment_Modules&amp;diff=2228"/>
		<updated>2023-06-16T08:59:32Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Preface ==&lt;br /&gt;
Environment modules [http://modules.sourceforge.net] are a smart way to provide interchangeable blocks of executables and reproducible environments for use in an HPC. It&#039;s also the only way to provide simultaneous versions of the same software without collisions, as each module is housed entirely in its own subfolder structure.&lt;br /&gt;
&lt;br /&gt;
== Using modules ==&lt;br /&gt;
The module executable is automatically provided to you upon login. Most users have some modules automatically loaded as well; to see these, use&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;module list&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
You should be able to see which modules are loaded.&lt;br /&gt;
&lt;br /&gt;
One of the most important modules to load is &#039;shared&#039; - this is Anunna specific, as it will extend the MODULEPATH environment variable to use modules present in /shared as well as /cm/local/ . Without this, many modules will not be available to you.&lt;br /&gt;
&lt;br /&gt;
== Loading modules ==&lt;br /&gt;
Availability of (different versions of) software can be checked by the following command:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;module avail&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
For example, you should be able to find the basic module slurm. This provides the path to the sbatch, srun, etc. executables for job submission. To load this, simply:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;module load slurm&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And you will see that it will automatically load the latest version - no need to write out any further.&lt;br /&gt;
&lt;br /&gt;
Many of the hand-installed programs have a path such as:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;hdf5/gcc/64/1.8.14&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
Which translates into:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;SOFTWARE/COMPILER/BITS/VERSION&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
You can elect to load this to various levels:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;module load hdf5               # loads the latest version, not caring for compiler&lt;br /&gt;
module load hdf5/gcc           # loads the latest gcc-compiled version, not caring for 32/64 bits (default 64)&lt;br /&gt;
module load hdf5/gcc/64        # loads the latest 64-bit gcc-compiled version&lt;br /&gt;
module load hdf5/gcc/64/1.8.14 # loads this specific version of hdf5&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This allows your job scripts to either automatically be upgraded when the latest executables are installed, or elect to use only one specific version of a piece of code.&lt;br /&gt;
&lt;br /&gt;
== Switching modules ==&lt;br /&gt;
&lt;br /&gt;
If you want to remove a module, simply&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;module unload module/1&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
This will remove the executable path from your environment. It&#039;ll also follow the same logic as above, i.e. you can unload all loaded slurm modules independent of version by just unloading the base module name. You can then load up a new one. You can do this in one command, with:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;module switch module/1 module/2&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Some modules will not allow themselves to be loaded when another one is loaded, for instance, for sanity reasons it&#039;s not possible to load two java modules at the same time. Trying to do this will give:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;Module &#039;module/2&#039; conflicts with the currently loaded module(s) &#039;module/1&#039;&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
If you&#039;re seeing this, you must unload or switch your modules rather than overloading them:&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;module switch module/2&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
This works if both modules have the same base path.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== External links ==&lt;br /&gt;
* http://modules.sourceforge.net&lt;br /&gt;
* http://www.admin-magazine.com/HPC/Articles/Environment-Modules&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Population_variant_calling_pipeline&amp;diff=2227</id>
		<title>Population variant calling pipeline</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Population_variant_calling_pipeline&amp;diff=2227"/>
		<updated>2023-06-16T08:59:21Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Author: Carolina Pita Barros &amp;lt;br /&amp;gt; &lt;br /&gt;
Contact: carolina.pitabarros@wur.nl  &amp;lt;br /&amp;gt;&lt;br /&gt;
ABG&amp;lt;br /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
For up-to-date documentation see [https://github.com/CarolinaPB/population-variant-calling here]&lt;br /&gt;
&lt;br /&gt;
= Population level variant calling =&lt;br /&gt;
&lt;br /&gt;
Path to pipeline: /lustre/nobackup/WUR/ABGC/shared/PIPELINES/population-variant-calling&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;first-follow-the-instructions-here&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
== First follow the instructions here ==&lt;br /&gt;
&lt;br /&gt;
[https://carolinapb.github.io/2021-06-23-how-to-run-my-pipelines/ Step by step guide on how to use my pipelines]&amp;lt;br /&amp;gt;&lt;br /&gt;
Click [https://github.com/CarolinaPB/snakemake-template/blob/master/Short%20introduction%20to%20Snakemake.pdf here] for an introduction to Snakemake&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;about&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
== ABOUT ==&lt;br /&gt;
&lt;br /&gt;
This is a pipeline that takes short reads aligned to a genome (in &amp;lt;code&amp;gt;.bam&amp;lt;/code&amp;gt; format) and performs population level variant calling with &amp;lt;code&amp;gt;Freebayes&amp;lt;/code&amp;gt;. It uses VEP to annotate the resulting VCF, calculates statistics, and calculates and plots a PCA.&lt;br /&gt;
&lt;br /&gt;
It was developed to work with the results of [https://github.com/CarolinaPB/population-mapping this population mapping pipeline]. There are a few &amp;lt;code&amp;gt;Freebayes&amp;lt;/code&amp;gt; requirements that you need to take into account if you don&#039;t use the mapping pipeline mentioned above to map your reads. You should make sure that:&lt;br /&gt;
&lt;br /&gt;
* Alignments have read groups&lt;br /&gt;
* Alignments are sorted&lt;br /&gt;
* Duplicates are marked&lt;br /&gt;
&lt;br /&gt;
See [https://github.com/freebayes/freebayes#calling-variants-from-fastq-to-vcf here] for more details.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;tools-used&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
==== Tools used ====&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/freebayes/freebayes Freebayes] - variant calling using short reads&lt;br /&gt;
* [https://samtools.github.io/bcftools/bcftools.html bcftools] - vcf statistics&lt;br /&gt;
* [https://www.cog-genomics.org/plink/ Plink] - compute PCA&lt;br /&gt;
* R - Plot PCA&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
!align=&amp;quot;center&amp;quot;| [[File:population-var-calling-workflow.png]]&lt;br /&gt;
|-&lt;br /&gt;
|align=&amp;quot;center&amp;quot;| &#039;&#039;Pipeline workflow&#039;&#039;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;edit-configyaml-with-the-paths-to-your-files&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
=== Edit config.yaml with the paths to your files ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;syntaxhighlight lang=&amp;quot;yaml&amp;quot;&amp;gt;ASSEMBLY: /path/to/fasta&lt;br /&gt;
MAPPING_DIR: /path/to/bams/dir&lt;br /&gt;
PREFIX: &amp;lt;prefix&amp;gt;&lt;br /&gt;
OUTDIR: /path/to/outdir&lt;br /&gt;
SPECIES: &amp;lt;species&amp;gt;&lt;br /&gt;
NUM_CHRS: &amp;lt;number of chromosomes&amp;gt;&amp;lt;/syntaxhighlight&amp;gt;&lt;br /&gt;
* ASSEMBLY - path to genome fasta file&lt;br /&gt;
* MAPPING_DIR - path to directory with bam files to be used&lt;br /&gt;
** the pipeline will use all bam files in the directory, if you want to use a subset of those, create a file named &amp;lt;code&amp;gt;bam_list.txt&amp;lt;/code&amp;gt; that contains the paths to the bam files you want to use. One path per line.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;syntaxhighlight lang=&amp;quot;text&amp;quot;&amp;gt;/path/to/file.bam&lt;br /&gt;
/path/to/file2.bam&amp;lt;/syntaxhighlight&amp;gt;&lt;br /&gt;
* PREFIX - prefix for the created files&lt;br /&gt;
* OUTDIR - directory where snakemake will run and where the results will be written to&amp;lt;br /&amp;gt;&lt;br /&gt;
If you want the results to be written to this directory (not to a new directory), open config.yaml and comment out &amp;lt;code&amp;gt;OUTDIR: /path/to/outdir&amp;lt;/code&amp;gt;&lt;br /&gt;
* SPECIES - species name to be used for VEP&lt;br /&gt;
* NUM_CHRS - number of chromosomes for your species (necessary for plink). ex: 38&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;additional-set-up&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
== ADDITIONAL SET UP ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;configuring-vep&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
=== Configuring VEP ===&lt;br /&gt;
&lt;br /&gt;
This pipeline uses VEP in offline mode, which increases performance. In order to use it in this mode, the cache for the species used needs to be installed:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;for-people-using-wurs-anunna&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
==== For people using WUR&#039;s Anunna: ====&lt;br /&gt;
&lt;br /&gt;
Check if the cache file for your species already exist in &amp;lt;code&amp;gt;/lustre/nobackup/SHARED/cache/&amp;lt;/code&amp;gt;. If it doesn&#039;t, create it with&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;/usr/bin/perl /shared/apps/SHARED/ensembl-vep/INSTALL.pl --CACHEDIR /lustre/nobackup/SHARED/cache/ --AUTO c -n --SPECIES &amp;amp;lt;species&amp;amp;gt;&amp;lt;/pre&amp;gt;&lt;br /&gt;
When multiple assemblies are found you need to run it again with &amp;lt;code&amp;gt;--ASSEMBLY &amp;amp;lt;assembly name&amp;amp;gt;&amp;lt;/code&amp;gt;, where &amp;amp;quot;assembly name&amp;amp;quot; is the name of the assembly you want to use.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;for-those-not-from-wur&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
==== For those not from WUR: ====&lt;br /&gt;
&lt;br /&gt;
You can install VEP with&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;conda install -c bioconda ensembl-vep&amp;lt;/pre&amp;gt;&lt;br /&gt;
and install the cache with&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;vep_install --CACHEDIR &amp;amp;lt;where/to/install/cache&amp;amp;gt; --AUTO c -n --SPECIES &amp;amp;lt;species&amp;amp;gt;&amp;lt;/pre&amp;gt;&lt;br /&gt;
When multiple assemblies are found you need to run it again with &amp;lt;code&amp;gt;--ASSEMBLY &amp;amp;lt;assembly name&amp;amp;gt;&amp;lt;/code&amp;gt;, where &amp;amp;quot;assembly name&amp;amp;quot; is the name of the assembly you want to use.&lt;br /&gt;
&lt;br /&gt;
In the Snakefile, in rule &amp;lt;code&amp;gt;run_vep&amp;lt;/code&amp;gt;, replace &amp;lt;code&amp;gt;/shared/apps/SHARED/ensembl-vep/vep&amp;lt;/code&amp;gt; with &amp;lt;code&amp;gt;vep&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;installing-r-packages&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
=== Installing R packages ===&lt;br /&gt;
&lt;br /&gt;
First load R: &amp;lt;code&amp;gt;module load R/3.6.2&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Enter the R environment by writing &amp;lt;code&amp;gt;R&amp;lt;/code&amp;gt; and clicking enter. Install the packages:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;list.of.packages &amp;amp;lt;- c(&amp;amp;quot;optparse&amp;amp;quot;, &amp;amp;quot;data.table&amp;amp;quot;, &amp;amp;quot;ggplot2&amp;amp;quot;)&lt;br /&gt;
&lt;br /&gt;
new.packages &amp;amp;lt;- list.of.packages[!(list.of.packages %in% installed.packages()[,&amp;amp;quot;Package&amp;amp;quot;])]&lt;br /&gt;
if(length(new.packages)) install.packages(new.packages)&amp;lt;/pre&amp;gt;&lt;br /&gt;
If you get an error like this:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;Warning in install.packages(new.packages) :&lt;br /&gt;
&#039;lib = &amp;amp;quot;/shared/apps/R/3.6.2/lib64/R/library&amp;amp;quot;&#039; is not writable&amp;lt;/pre&amp;gt;&lt;br /&gt;
Follow the instructions on how to install R packages locally [https://wiki.anunna.wur.nl/index.php/Installing_R_packages_locally here] and try to install the packages again.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span id=&amp;quot;results&amp;quot;&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
== RESULTS ==&lt;br /&gt;
&lt;br /&gt;
The most important files and directories are:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;amp;lt;run_date&amp;amp;gt;_files.txt&#039;&#039;&#039; dated file with an overview of the files used to run the pipeline (for documentation purposes)&lt;br /&gt;
* &#039;&#039;&#039;results&#039;&#039;&#039; directory that contains&lt;br /&gt;
** &#039;&#039;&#039;final_VCF&#039;&#039;&#039; directory with variant calling VCF files, as well as VCF stats&lt;br /&gt;
*** {prefix}.vep.vcf.gz - final VCF file&lt;br /&gt;
*** {prefix}.vep.vcf.gz.stats&lt;br /&gt;
** &#039;&#039;&#039;PCA&#039;&#039;&#039; PCA results and plot&lt;br /&gt;
*** {prefix}.eigenvec and {prefix}.eigenval - file with PCA eigenvectors and eigenvalues, respectively&lt;br /&gt;
*** {prefix}.pdf - PCA plot&lt;br /&gt;
&lt;br /&gt;
The VCF file has been filtered for &amp;lt;code&amp;gt;QUAL &amp;amp;gt; 20&amp;lt;/code&amp;gt;. Freebayes is ran with parameters &amp;lt;code&amp;gt;--use-best-n-alleles 4 --min-base-quality 10 --min-alternate-fraction 0.2 --haplotype-length 0 --ploidy 2 --min-alternate-count 2&amp;lt;/code&amp;gt;. These parameters can be changed in the Snakefile.&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Using_environment_modules&amp;diff=2226</id>
		<title>Using environment modules</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Using_environment_modules&amp;diff=2226"/>
		<updated>2023-06-16T08:59:02Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=== Environment Modules ===&lt;br /&gt;
[http://modules.sourceforge.net/ Environment modules] are a simple way to allow multiple potentially clashing programs to coexist on a large shared machine such as an HPC. It allows a user to specify exactly which programs are loaded, and even which version of each program, whilst simultaneously allowing the administrator the ability to automatically configure the appropriate environment variables for the system itself.&lt;br /&gt;
&lt;br /&gt;
== Viewing Modules ==&lt;br /&gt;
Upon logging in to Anunna, you should find that when you do:&lt;br /&gt;
  module list&lt;br /&gt;
&lt;br /&gt;
You will see something like this:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
-bash-4.1$ module list&lt;br /&gt;
Currently Loaded Modulefiles:&lt;br /&gt;
  1) shared        2) slurm/2.5.7&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is a list of all loaded modules in your shell session. To get a list of all available modules, simply&lt;br /&gt;
   module available&lt;br /&gt;
&lt;br /&gt;
And this will show you the (very exhaustive) list of modules on Anunna:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source  lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
-bash-4.1$ module avail&lt;br /&gt;
&lt;br /&gt;
---------------------------- /shared/modulefiles ----------------------------&lt;br /&gt;
acml/gcc/64/5.3.1                     netcdf/gcc/64/4.1.3&lt;br /&gt;
acml/gcc/fma4/5.3.1                   netcdf/gcc/64/4.3.0&lt;br /&gt;
acml/gcc/mp/64/5.3.1                  netcdf/gcc/64/4.3.2&lt;br /&gt;
acml/gcc/mp/fma4/5.3.1                netcdf/gcc/64/4.3.3&lt;br /&gt;
acml/gcc-int64/64/5.3.1               netcdf/gcc/64/4.3.3.1&lt;br /&gt;
acml/gcc-int64/fma4/5.3.1             netcdf/intel/64/4.1.3&lt;br /&gt;
...&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let&#039;s look at each of these module names. Each module is named for the application it provides, plus a subfolder of which compiler it was compiled with (if compiled), the number of address bits or options (if compiled), and the version.&lt;br /&gt;
&lt;br /&gt;
If you want to see a list for a specific module, you can&lt;br /&gt;
  module avail netcdf&lt;br /&gt;
&lt;br /&gt;
And the complete list of versions will be shown.&lt;br /&gt;
&lt;br /&gt;
== Loading Modules ==&lt;br /&gt;
To load a module, simply&lt;br /&gt;
  module load foo&lt;br /&gt;
&lt;br /&gt;
And the most recent version of module foo will automatically be loaded. If foo is compiled, it will automatically select the gcc version. If you want to specify a certain version, then&lt;br /&gt;
  module load foo/gcc/64/1.0.0&lt;br /&gt;
&lt;br /&gt;
Will load foo version 1, compiled with gcc. Be advised that this may not always work, as some modules are not compatible with each other, but a message will be shown if this is the case. Additionally, some modules will automatically load other modules with them for them to operate.&lt;br /&gt;
&lt;br /&gt;
== Unloading Modules ==&lt;br /&gt;
If you want to remove a module that you&#039;ve loaded, then&lt;br /&gt;
  module unload foo&lt;br /&gt;
&lt;br /&gt;
Will remove all module foo&#039;s loaded.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Example ==&lt;br /&gt;
Consider this simple python3 script that should calculate Pi to 1 million digits:&lt;br /&gt;
&amp;lt;source lang=&#039;python&#039;&amp;gt;&lt;br /&gt;
from decimal import *&lt;br /&gt;
D=Decimal&lt;br /&gt;
getcontext().prec=10000000&lt;br /&gt;
p=sum(D(1)/16**k*(D(4)/(8*k+1)-D(2)/(8*k+4)-D(1)/(8*k+5)-D(1)/(8*k+6))for k in range(411))&lt;br /&gt;
print(str(p)[:10000002])&lt;br /&gt;
&amp;lt;/source&amp;gt; &lt;br /&gt;
&lt;br /&gt;
This script will not run at all in the default 2.4 version of Python on the cluster. In order for this script to run you must use Python3. To do this, first list all versions of Python:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
-bash-4.1$ module avail python&lt;br /&gt;
&lt;br /&gt;
---------------------------- /shared/modulefiles ----------------------------&lt;br /&gt;
python/2.7.6 python/3.3.3 python/3.4.2&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then you can load the specific version you need:&lt;br /&gt;
  module load python/3.3.3&lt;br /&gt;
&lt;br /&gt;
Now you have access to the executable python3.&lt;br /&gt;
&lt;br /&gt;
== See also ==&lt;br /&gt;
* [[Environment_Modules | Environment Modules]]&lt;br /&gt;
* [[Control_R_environment_using_modules | Control R environment using modules]]&lt;br /&gt;
* [[Create_shortcut_log-in_command | Create a shortcut for the ssh log-in command]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
&lt;br /&gt;
== External links ==&lt;br /&gt;
* http://modules.sourceforge.net &lt;br /&gt;
* https://modules.readthedocs.io/en/latest/ (documentation)&lt;br /&gt;
* http://www.admin-magazine.com/HPC/Articles/Environment-Modules&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Assemble_mitochondrial_genomes_from_short_read_data&amp;diff=2225</id>
		<title>Assemble mitochondrial genomes from short read data</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Assemble_mitochondrial_genomes_from_short_read_data&amp;diff=2225"/>
		<updated>2023-06-16T08:58:46Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;A simple procedure for assembling mitochondrial genomes based on whole-genome re-sequencing data. The first step is to extract reads from the sequence library based on a closely related entirely assembled genome (e.g., for pig, the MT genome as present in the genome build, but could also be of a related species). The genome is then assembled using SOAPdenovo.&lt;br /&gt;
&lt;br /&gt;
* a reference genome of a closely related population or species.&lt;br /&gt;
* a bowtie2 index (make with bowtie2_build)&lt;br /&gt;
* a blastable db of the reference mitochondrial genome&lt;br /&gt;
* a SOAPdenovo configuration file:&lt;br /&gt;
&lt;br /&gt;
soapdenovo.config&lt;br /&gt;
  [LIB]&lt;br /&gt;
  avg_ins=450&lt;br /&gt;
  reverse_seq=0&lt;br /&gt;
  asm_flags=1&lt;br /&gt;
  rank=3&lt;br /&gt;
  q1=fq1.fq&lt;br /&gt;
  q2=fq2.fq&lt;br /&gt;
Note that the avg_ins flag may vary between libraries; may have an effect on assembly efficiency.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
#!/bin/bash&lt;br /&gt;
#SBATCH --time=1000&lt;br /&gt;
#SBATCH --mem=16000&lt;br /&gt;
#SBATCH --ntasks=8&lt;br /&gt;
#SBATCH --nodes=1&lt;br /&gt;
#SBATCH --constraint=4gpercpu&lt;br /&gt;
#SBATCH --output=output_%j.txt&lt;br /&gt;
#SBATCH --error=error_output_%j.txt&lt;br /&gt;
#SBATCH --job-name=assemble_mito&lt;br /&gt;
#SBATCH --mail-type=ALL&lt;br /&gt;
#SBATCH --mail-user=hendrik-jan.megens@wur.nl&lt;br /&gt;
module load bowtie/2-2.2.1 SOAPdenovo2/r240 BLAST+/2.2.28 MUMmer/3.23 &lt;br /&gt;
&lt;br /&gt;
bowtie2 --phred$2 --local -p 8 -x mt_pig.fa -1 $3 -2 $4 | head -2 &amp;gt;$1_mito_align.sam&lt;br /&gt;
bowtie2 --phred$2 --local -p 8 -x mt_pig.fa -1 $3 -2 $4 | awk &#039;$5&amp;gt;0&#039; | head -10000 &amp;gt;&amp;gt;$1_mito_align.sam&lt;br /&gt;
&lt;br /&gt;
java7 -jar /shared/apps/SHARED/picard-tools/picard-tools-1.109/SamToFastq.jar I=$1_mito_align.sam F=fq1.fq F2=fq2.fq INCLUDE_NON_PF_READS=True&lt;br /&gt;
&lt;br /&gt;
SOAPdenovo-63mer all -K 63 -p 4 -s soapdenovo.config -o $1_mito_assembly.fa&lt;br /&gt;
&lt;br /&gt;
blastn -query $1_mito_assembly.fa.scafSeq -db mt_pig.fa -outfmt 6&lt;br /&gt;
&lt;br /&gt;
mummer -mum -b -c mt_pig.fa $1_mito_assembly.fa.scafSeq &amp;gt; mummer.mums&lt;br /&gt;
mummerplot -postscript -p mummer mummer.mums&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Invoke like this:&lt;br /&gt;
&amp;lt;source lang=&#039;bash&#039;&amp;gt;&lt;br /&gt;
sbatch do_mtalign_bowtie_pig.sh MA01F18 33\&lt;br /&gt;
 /lustre/nobackup/WUR/ABGC/shared/Pig/ABGSA/ABGSA0071/ABGSA0071_MA01F18_R1.PF.fastq.gz\&lt;br /&gt;
 /lustre/nobackup/WUR/ABGC/shared/Pig/ABGSA/ABGSA0071/ABGSA0071_MA01F18_R2.PF.fastq.gz&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Globally_installed_software&amp;diff=2224</id>
		<title>Globally installed software</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Globally_installed_software&amp;diff=2224"/>
		<updated>2023-06-16T08:58:33Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Available as modules ==&lt;br /&gt;
* acml&lt;br /&gt;
* alphaimpute&lt;br /&gt;
* bamtools&lt;br /&gt;
* bcftools&lt;br /&gt;
* beagle&lt;br /&gt;
* blacs&lt;br /&gt;
* blas&lt;br /&gt;
* [[BLAST | BLAST+]]&lt;br /&gt;
* bonnie++&lt;br /&gt;
* boost&lt;br /&gt;
* [[bowtie2 | bowtie]]&lt;br /&gt;
* [[canu | canu]]&lt;br /&gt;
* cdo&lt;br /&gt;
* cmgui&lt;br /&gt;
* [[cuda | cuda]]&lt;br /&gt;
* diamond&lt;br /&gt;
* [[dmtcp | dmtcp]]&lt;br /&gt;
* emos&lt;br /&gt;
* fasttree&lt;br /&gt;
* ferret&lt;br /&gt;
* ffmpeg&lt;br /&gt;
* fftw2&lt;br /&gt;
* fftw3&lt;br /&gt;
* flex&lt;br /&gt;
* freebayes&lt;br /&gt;
* gcc&lt;br /&gt;
* gcta&lt;br /&gt;
* gdal&lt;br /&gt;
* gdb&lt;br /&gt;
* geos&lt;br /&gt;
* glibc&lt;br /&gt;
* glimmer&lt;br /&gt;
* glimmerHMM&lt;br /&gt;
* globalarrays&lt;br /&gt;
* grads&lt;br /&gt;
* grib&lt;br /&gt;
* gsl&lt;br /&gt;
* [[hadoop | hadoop ]]&lt;br /&gt;
* hdf4&lt;br /&gt;
* hdf5&lt;br /&gt;
* [[hmmer | hmmer]]&lt;br /&gt;
* hpl&lt;br /&gt;
* htslib&lt;br /&gt;
* hwloc&lt;br /&gt;
* ima2p&lt;br /&gt;
* intel&lt;br /&gt;
* intel-cluster-checker&lt;br /&gt;
* intel-cluster-runtime&lt;br /&gt;
* intel-tbb-oss&lt;br /&gt;
* iozone&lt;br /&gt;
* jasper&lt;br /&gt;
* java&lt;br /&gt;
* julia&lt;br /&gt;
* lapack&lt;br /&gt;
* [[matlab | MATLAB]]&lt;br /&gt;
* mixblup&lt;br /&gt;
* mpich&lt;br /&gt;
* mpiexec&lt;br /&gt;
* MRO&lt;br /&gt;
* muscle&lt;br /&gt;
* mvapich&lt;br /&gt;
* mvapich2&lt;br /&gt;
* ncbi-blast&lt;br /&gt;
* ncl&lt;br /&gt;
* nco&lt;br /&gt;
* ncview&lt;br /&gt;
* netcdf&lt;br /&gt;
* netcdf3-c++&lt;br /&gt;
* netperf&lt;br /&gt;
* octave&lt;br /&gt;
* open64&lt;br /&gt;
* openblas&lt;br /&gt;
* openlava&lt;br /&gt;
* openmpi&lt;br /&gt;
* oracle-instantclient&lt;br /&gt;
* phdf5&lt;br /&gt;
* picard&lt;br /&gt;
* plink&lt;br /&gt;
* prodigal&lt;br /&gt;
* python&lt;br /&gt;
* R&lt;br /&gt;
* ragel&lt;br /&gt;
* [[RAxML | RAxML]]&lt;br /&gt;
* RRO&lt;br /&gt;
* rstudio&lt;br /&gt;
* [[samtools | samtools]]&lt;br /&gt;
* scalapack&lt;br /&gt;
* slurm&lt;br /&gt;
* slurm-drmaa&lt;br /&gt;
* snpEff&lt;br /&gt;
* [[spark | SPARK]]&lt;br /&gt;
* szip&lt;br /&gt;
* torque&lt;br /&gt;
* [[Trinity | trinity]]&lt;br /&gt;
* udunits&lt;br /&gt;
* vcftools&lt;br /&gt;
* vim&lt;br /&gt;
* zlib&lt;br /&gt;
&lt;br /&gt;
== Globally installed on all nodes ==&lt;br /&gt;
&lt;br /&gt;
* Perl5.10&lt;br /&gt;
* pigz&lt;br /&gt;
* Python2.6&lt;br /&gt;
* BioPerl v1.61&lt;br /&gt;
* [http://samtools.sourceforge.net/tabix.shtml bgzip]&lt;br /&gt;
* [http://samtools.sourceforge.net/tabix.shtml tabix v0.2.5]&lt;br /&gt;
&lt;br /&gt;
== Available as global SHARED modules ==&lt;br /&gt;
Software can be deposited in:&lt;br /&gt;
  /shared/apps/SHARED/&lt;br /&gt;
&lt;br /&gt;
Modules can be found in:&lt;br /&gt;
  /shared/modulefiles/SHARED/&lt;br /&gt;
&lt;br /&gt;
{| width=&amp;quot;90%&amp;quot;&lt;br /&gt;
|- valign=&amp;quot;top&amp;quot;&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [[allpathslg_48961 | ALLPATHS-LG/48961]]&lt;br /&gt;
* [[allpathslg 51910 | ALLPATHS-LG/51910]]&lt;br /&gt;
* [[ANGSD_0.614 | angsd/0.614]]&lt;br /&gt;
* [[augustus_2.7 | augustus/2.7]]&lt;br /&gt;
* [[bedtools2.18 | bedtools/2.18.0]]&lt;br /&gt;
* [[BLAST | BLAST+/2.2.28]]   &lt;br /&gt;
* [[blat_v35 |blat/v35]]&lt;br /&gt;
* [[bowtie2_v2.2.1 | bowtie/2-2.2.1]]&lt;br /&gt;
* [[bowtie1_v1.0.0 | bowtie/1-1.0.0]]&lt;br /&gt;
* [[bwa_5.9 | bwa/0.5.9]]   &lt;br /&gt;
* [[bwa_7.5 | bwa/0.7.5a]]&lt;br /&gt;
* [[cegma_2.4 | cegma/2.4]]   &lt;br /&gt;
* [[Cufflinks | cufflinks/2.1.1]]&lt;br /&gt;
&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [[exonerate_2.2.0 | exonerate/2.2.0-x86_64]] &lt;br /&gt;
* [[geneid_1.4.4 | geneid/1.4.4]]&lt;br /&gt;
* [[genewise_2.2.3 | genewise/2.2.3-rc7]]     &lt;br /&gt;
* [[gmap_2014-01-21 | gmap/2014-01-21]]&lt;br /&gt;
* [[hmmer_3.1 | hmmer/3.1b1]]&lt;br /&gt;
* [[jellyfish_2.1.1 | jellyfish/2.1.1]]&lt;br /&gt;
* [[MAFFT_7.130 | MAFFT/7.130]]&lt;br /&gt;
* [[maker_2.2.8 | maker/2.28]]&lt;br /&gt;
* [[Muscle_3.8.31 | muscle/3.8.31]]     &lt;br /&gt;
* [[Plink_1.07 | Plink/1.07]]&lt;br /&gt;
* [[Plink_1.9 | Plink/1.9]]&lt;br /&gt;
* [[Provean_1.1.3 | provean/1.1.3]]  &lt;br /&gt;
* [[RepeatMasker_4.0.3 | RepeatMasker/4.0.3]]&lt;br /&gt;
&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [[RepeatModeler_1.0.7 | RepeatModeler/1.0.7]]&lt;br /&gt;
* [[RAxML8.0.0 | RAxML/8.0.0]]&lt;br /&gt;
* [[samtools v0.1.12a | samtools/0.1.12a]]&lt;br /&gt;
* [[samtools v0.1.19 | samtools/0.1.19]]&lt;br /&gt;
* [[snap | snap/2013-11-29]]&lt;br /&gt;
* [[soapdenovo2_r240 | SOAPdenovo2/r240]]&lt;br /&gt;
* [[sra_toolkit_2.3.4 | sra-toolkit/2.3.4]]&lt;br /&gt;
* [[TopHat_2.0.11 | tophat/2.0.11]]&lt;br /&gt;
* [[Trinity_r20131110 | Trinity/r20131110]]&lt;br /&gt;
* [[wgs_assembler_8.1 | wgs-assembler/8.1]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Adding a module to your current session ==&lt;br /&gt;
Use &amp;lt;code&amp;gt;module apropos&amp;lt;/code&amp;gt; to find the module you wish to use, then &amp;lt;code&amp;gt;module load&amp;lt;/code&amp;gt; to enable it.&lt;br /&gt;
&lt;br /&gt;
== Adding a custom module directory to your environment ==&lt;br /&gt;
To allow the &amp;lt;code&amp;gt;module&amp;lt;/code&amp;gt; program to find the custom module directory, the location of that directory has to be added to &amp;lt;code&amp;gt;MODULEPATH&amp;lt;/code&amp;gt; variable. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
export MODULEPATH=$MODULEPATH:/shared/apps/WUR/ABGC/modulefiles&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This can be made permanent by adding this line of code to the &amp;lt;code&amp;gt;.bash_profile&amp;lt;/code&amp;gt; file in the root of your home folder. To then load the modified &amp;lt;code&amp;gt;MODULEPATH&amp;lt;/code&amp;gt; variable you have to load  &amp;lt;code&amp;gt;.bash_profile&amp;lt;/code&amp;gt; again:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
source .bash_profile&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This needs to be done only for terminals that are already open. Next time you login, &amp;lt;code&amp;gt;.bash_profile&amp;lt;/code&amp;gt; will be loaded automatically.&lt;br /&gt;
&lt;br /&gt;
You can check if the modules are found.&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
module avail&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
This should give output that includes something similar to this:&lt;br /&gt;
&lt;br /&gt;
  ---------------------------------------- /shared/modulefiles/ -----------------------------------------&lt;br /&gt;
  ALLPATHS-LG/48961      bwa/0.7.5a             jellyfish/2.1.1        RepeatMasker/4.0.3&lt;br /&gt;
  augustus/2.7           cegma/2.4              MAFFT/7.130            RepeatModeler/1.0.7&lt;br /&gt;
  bedtools/2.18.0        cufflinks/2.1.1        maker/2.28             samtools/0.1.12a&lt;br /&gt;
  BLAST+/2.2.28          exonerate/2.2.0-x86_64 muscle/3.8.31          samtools/0.1.19&lt;br /&gt;
  blat/v35               geneid/1.4.4           plink/1.07             snap/2013-11-29&lt;br /&gt;
  bowtie/2-2.2.1         genewise/2.2.3-rc7     provean/1.1.3          SOAPdenovo2/r240&lt;br /&gt;
  bwa/0.5.9              hmmer/3.1b1            RAxML/8.0.0            tophat/2.0.11&lt;br /&gt;
&lt;br /&gt;
== See also ==&lt;br /&gt;
* [[installation_by_users | Installing domain specific software: installation by users]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
* [[ABGC_modules | modules specific for ABGC ]]&lt;br /&gt;
&lt;br /&gt;
== External links ==&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Node_usage_graph&amp;diff=2223</id>
		<title>Node usage graph</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Node_usage_graph&amp;diff=2223"/>
		<updated>2023-06-16T08:55:36Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;There is a graphing tool that uses elements directly from sacct to display information about the current cluster usage, node_usage_graph (located at in the anunna module ).&lt;br /&gt;
&lt;br /&gt;
Example:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[user@login0 ~]# module load anunna&lt;br /&gt;
[user@login0 ~]# usage_graph&lt;br /&gt;
node:   |0%                                                                             100%|&lt;br /&gt;
fat001: DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
        DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
fat002: CCCCCCCCC                                                                            &lt;br /&gt;
        MMMMMmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm                                       &lt;br /&gt;
node001:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node002:cccccccccc                                                                           &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMmmmmmmmmmm                                           &lt;br /&gt;
node003:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC           &lt;br /&gt;
        MM                                                                                   &lt;br /&gt;
node004:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                      &lt;br /&gt;
        M                                                                                    &lt;br /&gt;
node005:CCCCCCCCCC                                                                           &lt;br /&gt;
                                                                                             &lt;br /&gt;
node006:CCCCCCCCCC                                                                           &lt;br /&gt;
                                                                                             &lt;br /&gt;
node007:CCCCCCCCCC                                                                           &lt;br /&gt;
                                                                                             &lt;br /&gt;
node008:CCCCCCCCCCccccc                                                                      &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMM                                                                &lt;br /&gt;
node009:cccccccccc                                                                           &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMM                                           &lt;br /&gt;
node010:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node011:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node012:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                &lt;br /&gt;
        M                                                                                    &lt;br /&gt;
node013:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node014:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node015:CCCCC                                                                                &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm &lt;br /&gt;
node016:CCCCCCCCCCCCCCCCCCCCC                                                                &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm &lt;br /&gt;
node017:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node018:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node019:CCCCC                                                                                &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm &lt;br /&gt;
node020:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node021:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node022:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node023:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node024:CCCCCCCCCCCCCCC                                                                      &lt;br /&gt;
                                                                                             &lt;br /&gt;
node025:CCCCCCCCCCCCCCCCCCCCC                                                                &lt;br /&gt;
                                                                                             &lt;br /&gt;
node026:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                                &lt;br /&gt;
                                                                                             &lt;br /&gt;
node027:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                                &lt;br /&gt;
                                                                                             &lt;br /&gt;
node028:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC&lt;br /&gt;
        MMM                                                                                  &lt;br /&gt;
node029:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node030:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node031:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node032:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node033:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node034:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node035:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node036:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node037:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node038:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node039:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node040:DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
        DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
node041:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCcccccc                                                &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMM                                                               &lt;br /&gt;
node042:RRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR&lt;br /&gt;
        RRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR&lt;br /&gt;
node049:DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
        DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
node050:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                      &lt;br /&gt;
        M                                                                                    &lt;br /&gt;
node051:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node052:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                                &lt;br /&gt;
        MMMMMMmmmmmmmmmmmmmmm                                                                &lt;br /&gt;
node053:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                                &lt;br /&gt;
        M                                                                                    &lt;br /&gt;
node054:DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
        DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This gives an overview of the current per-node resource usage. There are four types of letter:&lt;br /&gt;
* M: Memory reserved and in use&lt;br /&gt;
* m: Memory reserved and not in use&lt;br /&gt;
* C: CPU reserved and in use&lt;br /&gt;
* c: CPU reserved and not in use&lt;br /&gt;
* D: Drained node (not available for submission for some adminstrative reason&lt;br /&gt;
* R: Reserved node&lt;br /&gt;
&lt;br /&gt;
It cannot however give you an indication of how much the queue is right now for any node. for that, squeue is a better resource.&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Node_usage_graph&amp;diff=2222</id>
		<title>Node usage graph</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Node_usage_graph&amp;diff=2222"/>
		<updated>2023-06-16T08:55:20Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;There is a graphing tool that uses elements directly from sacct to display information about the current cluster usage, node_usage_graph (located at /cm/shared/apps/accounting/node_usage_graph ).&lt;br /&gt;
&lt;br /&gt;
Example:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[user@login0 ~]# module load anunna&lt;br /&gt;
[user@login0 ~]# usage_graph&lt;br /&gt;
node:   |0%                                                                             100%|&lt;br /&gt;
fat001: DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
        DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
fat002: CCCCCCCCC                                                                            &lt;br /&gt;
        MMMMMmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm                                       &lt;br /&gt;
node001:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node002:cccccccccc                                                                           &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMmmmmmmmmmm                                           &lt;br /&gt;
node003:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC           &lt;br /&gt;
        MM                                                                                   &lt;br /&gt;
node004:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                      &lt;br /&gt;
        M                                                                                    &lt;br /&gt;
node005:CCCCCCCCCC                                                                           &lt;br /&gt;
                                                                                             &lt;br /&gt;
node006:CCCCCCCCCC                                                                           &lt;br /&gt;
                                                                                             &lt;br /&gt;
node007:CCCCCCCCCC                                                                           &lt;br /&gt;
                                                                                             &lt;br /&gt;
node008:CCCCCCCCCCccccc                                                                      &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMM                                                                &lt;br /&gt;
node009:cccccccccc                                                                           &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMM                                           &lt;br /&gt;
node010:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node011:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node012:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                &lt;br /&gt;
        M                                                                                    &lt;br /&gt;
node013:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node014:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node015:CCCCC                                                                                &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm &lt;br /&gt;
node016:CCCCCCCCCCCCCCCCCCCCC                                                                &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm &lt;br /&gt;
node017:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node018:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node019:CCCCC                                                                                &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm &lt;br /&gt;
node020:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node021:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node022:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node023:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node024:CCCCCCCCCCCCCCC                                                                      &lt;br /&gt;
                                                                                             &lt;br /&gt;
node025:CCCCCCCCCCCCCCCCCCCCC                                                                &lt;br /&gt;
                                                                                             &lt;br /&gt;
node026:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                                &lt;br /&gt;
                                                                                             &lt;br /&gt;
node027:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                                &lt;br /&gt;
                                                                                             &lt;br /&gt;
node028:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC&lt;br /&gt;
        MMM                                                                                  &lt;br /&gt;
node029:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node030:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node031:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node032:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node033:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node034:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node035:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node036:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node037:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node038:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node039:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node040:DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
        DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
node041:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCcccccc                                                &lt;br /&gt;
        MMMMMMMMMMMMMMMMMMMMMM                                                               &lt;br /&gt;
node042:RRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR&lt;br /&gt;
        RRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR&lt;br /&gt;
node049:DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
        DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
node050:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                      &lt;br /&gt;
        M                                                                                    &lt;br /&gt;
node051:                                                                                     &lt;br /&gt;
                                                                                             &lt;br /&gt;
node052:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                                &lt;br /&gt;
        MMMMMMmmmmmmmmmmmmmmm                                                                &lt;br /&gt;
node053:CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC                                                &lt;br /&gt;
        M                                                                                    &lt;br /&gt;
node054:DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
        DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This gives an overview of the current per-node resource usage. There are four types of letter:&lt;br /&gt;
* M: Memory reserved and in use&lt;br /&gt;
* m: Memory reserved and not in use&lt;br /&gt;
* C: CPU reserved and in use&lt;br /&gt;
* c: CPU reserved and not in use&lt;br /&gt;
* D: Drained node (not available for submission for some adminstrative reason&lt;br /&gt;
* R: Reserved node&lt;br /&gt;
&lt;br /&gt;
It cannot however give you an indication of how much the queue is right now for any node. for that, squeue is a better resource.&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
	<entry>
		<id>https://wiki.anunna.wur.nl/index.php?title=Main_Page&amp;diff=2183</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://wiki.anunna.wur.nl/index.php?title=Main_Page&amp;diff=2183"/>
		<updated>2023-01-27T14:40:32Z</updated>

		<summary type="html">&lt;p&gt;Dawes0011: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Anunna is a [http://en.wikipedia.org/wiki/High-performance_computing High Performance Computer] (HPC) infrastructure hosted by [http://www.wageningenur.nl/nl/activiteit/Opening-High-Performance-Computing-cluster-HPC.htm Wageningen University &amp;amp; Research Centre]. It is open for use for all WUR research groups as well as other organizations, including companies, that have collaborative projects with WUR. &lt;br /&gt;
&lt;br /&gt;
= Using Anunna =&lt;br /&gt;
* [[Tariffs | Costs associated with resource usage]]&lt;br /&gt;
&lt;br /&gt;
== Gaining access to Anunna==&lt;br /&gt;
Access to the cluster and file transfer are traditionally done via [http://en.wikipedia.org/wiki/Secure_Shell SSH and SFTP].&lt;br /&gt;
* [[log_in_to_B4F_cluster | Logging into cluster using ssh]]&lt;br /&gt;
* [[file_transfer | File transfer options]]&lt;br /&gt;
* [[Services | Alternative access methods, and extra features and services on Anunna]]&lt;br /&gt;
* [[Filesystems | Data storage methods on Anunna]]&lt;br /&gt;
&lt;br /&gt;
== Access Policy ==&lt;br /&gt;
[[Access_Policy | Main Article: Access Policy]]&lt;br /&gt;
&lt;br /&gt;
Access needs to be granted actively (by creation of an account on the cluster by FB-IT). Use of resources is limited by the scheduler. Depending on availability of queues (&#039;partitions&#039;) granted to a user, priority to the system&#039;s resources is regulated. Note that the use of Anunna is not free of charge. List price of CPU time and storage, and possible discounts on that list price for your organisation, can be retrieved from Shared Research Facilities or FB-IT.&lt;br /&gt;
&lt;br /&gt;
= Events =&lt;br /&gt;
&lt;br /&gt;
* [[Courses]] that have happened and are happening&lt;br /&gt;
* [[Downtime]] that will affect all users&lt;br /&gt;
* [[Meetings]] that may affect the policies of Anunna&lt;br /&gt;
&lt;br /&gt;
= Other Software =&lt;br /&gt;
&lt;br /&gt;
== Cluster Management Software and Scheduler ==&lt;br /&gt;
Anunna uses Bright Cluster Manager software for overall cluster management, and Slurm as job scheduler.&lt;br /&gt;
* [[BCM_on_B4F_cluster | Monitor cluster status with BCM]]&lt;br /&gt;
* [[Using_Slurm | Submit jobs with Slurm]]&lt;br /&gt;
* [[node_usage_graph | Be aware of how much work the cluster is under right now with &#039;node_usage_graph&#039;]]&lt;br /&gt;
* [[SLURM_Compare | Rosetta Stone of Workload Managers]]&lt;br /&gt;
&lt;br /&gt;
== Installation of software by users ==&lt;br /&gt;
&lt;br /&gt;
* [[Domain_specific_software_on_B4Fcluster_installation_by_users | Installing domain specific software: installation by users]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
* [[Virtual_environment_Python_3.4_or_higher | Setting up and using a virtual environment for Python3.4 or higher ]]&lt;br /&gt;
* [[Installing WRF and WPS]]&lt;br /&gt;
* [[Running scripts on a fixed timeschedule (cron)]]&lt;br /&gt;
&lt;br /&gt;
== Installed software ==&lt;br /&gt;
&lt;br /&gt;
* [[Globally_installed_software | Globally installed software]]&lt;br /&gt;
* [[ABGC_modules | ABGC specific modules]]&lt;br /&gt;
&lt;br /&gt;
= Useful Notes = &lt;br /&gt;
&lt;br /&gt;
== Being in control of Environment parameters ==&lt;br /&gt;
&lt;br /&gt;
* [[Using_environment_modules | Using environment modules]]&lt;br /&gt;
* [[Setting local variables]]&lt;br /&gt;
* [[Setting_TMPDIR | Set a custom temporary directory location]]&lt;br /&gt;
* [[Installing_R_packages_locally | Installing R packages locally]]&lt;br /&gt;
* [[Setting_up_Python_virtualenv | Setting up and using a virtual environment for Python3 ]]&lt;br /&gt;
&lt;br /&gt;
== Controlling costs ==&lt;br /&gt;
&lt;br /&gt;
* [[SACCT | using SACCT to see your costs]]&lt;br /&gt;
* [[get_my_bill | using the &amp;quot;get_my_bill&amp;quot; script to estimate costs]]&lt;br /&gt;
&lt;br /&gt;
== Management ==&lt;br /&gt;
Product Owner of Anunna is Alexander van Ittersum (Wageningen UR,FB-IT, C&amp;amp;PS). [[User:dawes001 | Gwen Dawes (Wageningen UR, FB-IT, C&amp;amp;PS)]] and [[User:haars001 | Jan van Haarst (Wageningen UR,FB-IT, C&amp;amp;PS)]] are responsible for [[Maintenance_and_Management | Maintenance and Management]] of the cluster.&lt;br /&gt;
&lt;br /&gt;
* [[Roadmap | Ambitions regarding innovation, support and administration of Anunna ]]&lt;br /&gt;
&lt;br /&gt;
= Miscellaneous =&lt;br /&gt;
* [[Mailinglist | Electronic mail discussion lists]]&lt;br /&gt;
* [[History_of_the_Cluster | Historical information on the startup of Anunna]]&lt;br /&gt;
* [[Bioinformatics_tips_tricks_workflows | Bioinformatics tips, tricks, and workflows]]&lt;br /&gt;
* [[Parallel_R_code_on_SLURM | Running parallel R code on SLURM]]&lt;br /&gt;
* [[Convert_between_MediaWiki_and_other_formats | Convert between MediaWiki format and other formats]]&lt;br /&gt;
* [[Manual GitLab | GitLab: Create projects and add scripts]]&lt;br /&gt;
* [[Monitoring_executions | Monitoring job execution]]&lt;br /&gt;
* [[Shared_folders | Working with shared folders in the Lustre file system]]&lt;br /&gt;
&lt;br /&gt;
= See also =&lt;br /&gt;
* [[Maintenance_and_Management | Maintenance and Management]]&lt;br /&gt;
* [[BCData | BCData]]&lt;br /&gt;
* [[Mailinglist | Electronic mail discussion lists]]&lt;br /&gt;
* [[About_ABGC | About ABGC]]&lt;br /&gt;
* [[Computer_cluster | High Performance Computing @ABGC]]&lt;br /&gt;
* [[Lustre_PFS_layout | Lustre Parallel File System layout]]&lt;br /&gt;
&lt;br /&gt;
= External links =&lt;br /&gt;
{| width=&amp;quot;90%&amp;quot;&lt;br /&gt;
|- valign=&amp;quot;top&amp;quot;&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [https://www.wur.nl/en/Value-Creation-Cooperation/Facilities/Wageningen-Shared-Research-Facilities/Our-facilities/Show/High-Performance-Computing-Cluster-HPC-Anunna.htm SRF offers a HPC facilty]&lt;br /&gt;
| width=&amp;quot;30%&amp;quot; |&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Scientific_Linux Scientific Linux]&lt;br /&gt;
* [http://en.wikipedia.org/wiki/Help:Cheatsheet Help with editing Wiki pages]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Dawes0011</name></author>
	</entry>
</feed>