Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.
Table of Contents

...

  1. Connect to the ASU Cisco AnyConnect VPN

  2. Login with your ASURITE & password

  3. Choose a connection method (terminal / web portal)

  4. Transfer needed files

  5. Run an interactive session or create an SBATCH script

  6. Access mamba for Python environments, and other public software modules, and /wiki/spaces/RC/pages/1754857495

Important Terms

  • HPC: Short for “High Performance Computing” it refers to a group (or a cluster) of interconnected computers designed for parallelism across many computers at once. Publicly these are often called “supercomputers”.

  • Node: A single machine in a supercomputer. This will be either a physical machine or a virtual machine. 

  • Login Node: A node intended as a launching point to compute nodes. Login nodes have minimal resources and should not be used for any application that consumes a lot of CPU or memory. This is also known as a “head node”.

Warning

Using the login nodes for computational work will result in temporary penalties on the account, e.g., do NOT install Python packages or connect to vscode on a login node.

  • Compute Node: Nodes intended for heavy computing. This is where all heavy processing should be done.

  • Job: Work assigned to be done on a compute node. Any time a compute node is assigned a job is created.

  • Memory (RAM): Short for “Random-Access Memory“. This is used for the amount of memory that each calculation or computation requires in order to execute and complete successfully. The term “memory“ is not used for disk space. This is another main component that defines a node.

  • CPU: Short for “Central Processing Unit”, also called a core. This is one of the main components that defines a computing device, such as a node.

  • GPU: Short for “Graphic Processing Unit”. This is a specialized piece of hardware that can enable and accelerate certain computational research.

  • Scheduler: The application on our end that manages and assigns (allocates) compute resources for jobs. The scheduler used on the ASU Supercomputers is called Slurm.

  • Fairshare: Jobs will cost the user’s fairshare points to run, and the lower it is, the longer the job queuing time will be. Please spend it wisely. Here is more info about it.

...

Using Python on supercomputers is a little bit different than on workstations or local computers. Please use the system-provided mamba instead of conda or pip. Please follow our guide closely for the best practices with Python on the ASU supercomputers: Working with Python

...

  • If your job is failing, a Job ID helps us significantly as we can pull detailed information about the job.

  • If you are new to Linux, or need a refresher, Research Computing has created a guide at The Linux Shell on the Sol Supercomputer. For a great reference on building proficiency with command-line tools, we provide the following MIT link from CSAIL.

...