You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Welcome the **Center for High Performance Computing (CHPC)'s Student Cluster Competition (SCC)** - Team Selection Round. This round requires each team to build a **prototype multi-node compute cluster** within the National Integrated Cyber Infrastructure Systems (NICIS) **virtual compute cloud** (described below).
5
5
6
-
The goal of this tutorial is to introduce you to the competition platform and familiarise you with some Linux and systems administration concepts. This competition provides you with a fixed set of virtual resources, that you will use to initialize a set a set of virtual machines instances based on your choice _or flavor_ of **
6
+
The goal of this document is to introduce you to the competition platform and familiarise you with some Linux and systems administration concepts. This competition provides you with a fixed set of virtual resources, that you will use to initialize a set a set of virtual machines instances based on your choice _or flavor_ of **
7
7
8
8
# Table of Contents
9
9
10
10
<!-- markdown-toc start - Don't edit this section. Run M-x markdown-toc-refresh-toc -->
1.[Links to Livestreams and Lecture Recordings](#links-to-livestreams-and-lecture-recordings)
28
-
1.[Day 1 - Welcome, Introduction and Getting Started](#day-1---welcome-introduction-and-getting-started)
29
-
1.[Day 2 - HPC Hardware, HPC Networking and Systems Administration](#day-2---hpc-hardware-hpc-networking-and-systems-administration)
30
-
1.[Day 3 - Benchmarking, Compilation and Parallel Computing](#day-3---benchmarking-compilation-and-parallel-computing)
31
-
1.[Day 4 - HPC Administration and Application Visualization](#day-4---hpc-administration-and-application-visualization)
32
-
1.[Day 5 - Career Guidance](#day-5---career-guidance)
26
+
1.[Day 1 - Welcome, Introduction and Getting Started](#day-1---welcome-introduction-and-getting-started)
27
+
1.[Day 2 - HPC Hardware, HPC Networking and Systems Administration](#day-2---hpc-hardware-hpc-networking-and-systems-administration)
28
+
1.[Day 3 - Benchmarking, Compilation and Parallel Computing](#day-3---benchmarking-compilation-and-parallel-computing)
29
+
1.[Day 4 - HPC Administration and Application Visualization](#day-4---hpc-administration-and-application-visualization)
30
+
1.[Day 5 - Career Guidance](#day-5---career-guidance)
33
31
1.[Tutorial Glossary and Section Overview](#tutorial-glossary-and-section-overview)
34
-
1.[Tutorial 1](#tutorial-1)
35
-
1.[Tutorial 2](#tutorial-2)
36
-
1.[Tutorial 3](#tutorial-3)
37
-
1.[Tutorial 4](#tutorial-4)
32
+
1.[Tutorial 1](#tutorial-1)
33
+
1.[Tutorial 2](#tutorial-2)
34
+
1.[Tutorial 3](#tutorial-3)
35
+
1.[Tutorial 4](#tutorial-4)
38
36
1.[Contributing to the Project](#contributing-to-the-project)
39
-
1.[Steps to follow when editing existing content](#steps-to-follow-when-editing-existing-content)
40
-
1.[Syntax and Style](#syntax-and-style)
37
+
1.[Steps to follow when editing existing content](#steps-to-follow-when-editing-existing-content)
38
+
1.[Syntax and Style](#syntax-and-style)
41
39
1.[Collaborating with your Team and Storing you Progress on GitHub](#collaborating-with-your-team-and-storing-you-progress-on-github)
42
-
1.[Forking the Tutorials into Your Own Team's Private GitHub Repository](#forking-the-tutorials-into-your-own-teams-private-github-repository)
43
-
1.[Editing the Git Markdown Files to Track Your Team's Progress](#editing-the-git-markdown-files-to-track-your-teams-progress)
40
+
1.[Forking the Tutorials into Your Own Team's Private GitHub Repository](#forking-the-tutorials-into-your-own-teams-private-github-repository)
41
+
1.[Editing the Git Markdown Files to Track Your Team's Progress](#editing-the-git-markdown-files-to-track-your-teams-progress)
44
42
45
43
<!-- markdown-toc end -->
46
44
@@ -204,15 +202,14 @@ Tutorial 1 deals with introducing concepts to users and getting them started wit
204
202
205
203
## Tutorial 2
206
204
207
-
Tutorial 2 deals with reverse proxy access for internal websites, central authentication and shared file systems.
205
+
Tutorial 2 deals with understaning the roles of the head and compute nodes, adding a compute node to create your cluster, configuring linux services such as the firewall, time server.
208
206
1.[Checklist](tutorial2/README.md#checklist)
209
207
1.[Spinning Up a Compute Node in OpenStack](tutorial2/README.md#spinning-up-a-compute-node-in-openstack)
@@ -357,7 +355,7 @@ You are strongly encouraged to contribute and improve the project by [Opening an
357
355
In order to effectively manage the various workflows and stages of development, testing and deployment, the project is comprised of three primary branches:
358
356
*`main`: *Stable* and production-ready deployment branch of the project.
359
357
*`stag`: *Staging* branch which mirrors production and is used for integration testing of new features.
360
-
*`dev`: *Development* branch fore incorporating new features and bug fixes.
358
+
*`dev`: *Development* branch for incorporating new features and bug fixes.
361
359
362
360
Editing the content directly, will require the use of Git. Using a terminal application or [Git for Windows PowerShell](https://git-scm.com/book/en/v2/Appendix-A:-Git-in-Other-Environments-Git-in-PowerShell) or [Git for MobaXTerm](https://www.geeksforgeeks.org/how-to-install-git-on-mobaxterm/).
Copy file name to clipboardExpand all lines: tutorial4/README.md
+24-15
Original file line number
Diff line number
Diff line change
@@ -34,9 +34,6 @@
34
34
35
35
# Checklist
36
36
37
-
Tutorial 4 demonstrates environment module manipulation and the compilation and optimisation of HPC benchmark software. This introduces the reader to the concepts of environment management and workspace sanity, as well as compilation of software on Linux.
38
-
39
-
40
37
This tutorial demonstrates _cluster monitoring_ and _workload scheduling_. These two components are critical to a typical HPC environment. Monitoring is a widely used component in system administration (including enterprise datacentres and corporate networks). Monitoring allows administrators to be aware of what is happening on any system that is being monitored and is useful to proactively identify where any potential issues may be. A workload scheduler ensures that users' jobs are handled properly to fairly balance all scheduled jobs with the resources available at any time.
41
38
42
39
In this tutorial you will:
@@ -194,11 +191,11 @@ The Slurm Workload Manager (formerly known as Simple Linux Utility for Resource
194
191
195
192
1. Make sure the clocks, i.e. chrony daemons, are synchronized across the cluster.
196
193
197
-
2. Generate a SLURM and MUNGE user on all of your nodes:
194
+
2. Generate a **SLURM** and **MUNGE** user on all of your nodes:
198
195
199
-
- **If you have FreeIPA authentication working**
200
-
- Create the users using the FreeIPA web interface. **Do NOT add them to the sysadmin group**.
201
-
- **If you do NOT have FreeIPA authentication working**
196
+
- **If you have Ansible User Module working**
197
+
- Create the users as shown in tutorial 2 **Do NOT add them to the sysadmin group**.
198
+
- **If you do NOT have your Ansible User Module working**
202
199
- `useradd slurm`
203
200
- Ensure that users and groups (UIDs and GIDs) are synchronized across the cluster. Read up on the appropriate [/etc/shadow](https://linuxize.com/post/etc-shadow-file/) and [/etc/password](https://www.cyberciti.biz/faq/understanding-etcpasswd-file-format/) files.
204
201
@@ -213,10 +210,11 @@ The Slurm Workload Manager (formerly known as Simple Linux Utility for Resource
213
210
[...@headnode ~]$ sudo dnf install epel-release
214
211
```
215
212
216
-
Then we can install MUNGE, pulling the development source code from the `powertools` repository:
213
+
Then we can install MUNGE, pulling the development source code from the `crb` "CodeReady Builder" repository:
0 commit comments