The Saion System
Saion is the main accelerator-based cluster at OIST. We have a number of GPU-based and high core-count partitions useful for deep learning and other specialised tasks. The "test-" and "intel" partitions are available for everyone with an account; apply to activate your account here. The regular partitions are restricted. Please apply for access here.
You log in to Saion with the "ssh" command. Please see our connection guide for details on setting it up for your own computer. To log in to Saion, run:
$ ssh email@example.com
This lands you in a login node. Here you can launch jobs, move data and prepare your work. Do not do any computation on the login nodes. To actually run a job you use Slurm, through the "srun" and "sbatch" commands. For instance, to start a 1-hour long interactive job on a test-amd node, run:
$ srun -p test-amd -t 0-1 --mem=10G --pty bash -l
A batch script looks like this:
#!/bin/bash -l SBATCH -p <partition> SBATCH -t 0-1 SBATCH --mem=10G ...
All you need to know about using Slurm and about running jobs on the clusters is available on this page. You can find Saion partition-specific information (how to specify the GPU for instance) below.
Saion has a number of partitions. "test-gpu" and "intel" are open and available for anyone, and are meant for test jobs and short computations. See Open Saion Partitions.
The "gpu", "powernv" and "kannel" partitions are restricted systems for general computation. "amd" and "prio-gpu" are dedicated for specific recurring data analysis tasks. See Restricted Saion Partitions
- "test-gpu" - 6 nodes with 4xP100 GPUs, 36 cores and 512GB memory. Available to all Saion users.
"intel" - 4 nodes with 40 cores and 128GB memory per user. Fast floating-point calculations. Gang scheduling, available to all Saion users.
- "gpu" - 8 nodes with 4xP100, and 8 nodes with 4xV100. Nodes are Intel w. 36 cores, 512GB mem. Good for any GPU-accelerated applications.
- "powernv" - 8 nodes with 4×GPUs. Nodes use dual IBM Power 8 or Power 9 CPUs with 512GB memory. A high-bandwidth system for more advanced users.
"prio-gpu" - 1 node with 4xP100 GPUs, 36 cores and 512GB memory. For specific recurring data processing jobs.
"kofugaku" - an 8-node ARM A64fx cluster. Explore the feasibility of running your code on the Fugaku supercomputer, or test an ARM-based HPC system.
Work is the compute storage on Saion. It's available as "/work". It is a 200TB SSD (solid state) storage system, and thus very, very fast. The allocation is 10TB per unit, and this allocation is not expandable. This system is meant only for data needed while you're running jobs; use Bucket for all long-term storage. Please see the section "workflow" below for tips on how to handle this storage.
On the old Saion system you could mount /work as a remote drive (using
\\tamashi.oist.jp). With this new storage system it is no longer possible. Instead you can use HPacquire to transfer data to Saion, and Bucket to transfer data both ways.
Bucket is mounted on the login nodes and read-only on compute nodes as "
/bucket", just like on Deigo. You can use the same data sets in the same way: Read your data directly from Bucket, use Work for temporary files and ouptut, then copy results back to Bucket and clean up Work at the end of the computation. See the Bucket page for more.
Deigo Flash is the
/flash file system on Deigo, accessible read-only as
/deigo_flash on Saion. You can use this to run a job on Deigo CPU nodes, then another on Saion GPU with the data from the Deigo job as input without having to copy anything.
HPacquire is fast storage for data import. The HPacquire system is mounted as /hpaqcuire on Saion. It is a fast, temporary storage area for transferring data from scientific instruments and outside sources into our HPC systems. You can mount hpaquire on your personal laptop or workstation; data collection system or other external data source, then copy the data to it.
On Saion, /hpacquire is available directly from the compute nodes, so you can use the data without having to copy it over to /work first. But HPacquire is read-only on the cluster side; you need to use /work to store any results, then copy it back to Bucket at the end. See the HPacquire manual for more on this system.
Home is the same as on Deigo. Please avoid storing any research data or perform computations from here. The system is quite slow and has a small 50G capacity.
Network, OS and administration
All components are connected using 100GB infiniband EDR. This is a fast, low-latency network that should give you file access times similar to a local solid-state drive.
Saion has two login nodes, and two administration nodes (for redundancy). The operating system is CentOS 7.6 throughout.
The Work storage system gives each unit 10TB of storage space. This is fast but the storage is limited, so you need to use the same workflow as on Deigo.
Keep your data on Bucket, where it is safe and backed-up. Read data directly from Bucket during your computations.
- Use Work as temporary storage during your computation: log files, temporary files, output files and so on. You can create a subdirectory and keep all your temporary data there.
When you're done, copy any files you want to keep back to Bucket. Finally clean up Work to free up space. If you created a temporary directory you can just delete the whole directory at once. Never keep data permanently on Work.
Start your job from your home directory. The Slurm output files get written to the starting directory. If you start from Bucket they can't be written to, and the job will fail. If you start from /flash you need to remember to clean them up separately.
Alternatively, explicitly specify the absolute path to your Slurm output files with the "-o" parameter. We suggest you use a folder in your home for this.
Create a temporary directory
You can use the "mktemp" command to create files or directories that are guaranteed to be unique. You don't want to accidentally mix files for different jobs with each other. We save the directory name in a shell variable:
$ tmpdir=$(mktemp -d /work/MyUnit/tempdir.XXXXX)
$ cd $tmpdir
Now we're in a temporary directory on Work. All output will end up here, and we can clean it up afterwards by deleting the directory.
copy back (stage out) your results
Bucket is read-only, so you can't use "cp" to copy results back directly during a job. Instead, use "scp" to copy results back from a compute node or in a job script:
scp /work/myUnit/myresult.dat oist-username@saion:/bucket/myUnit/
or (this is technically faster):
$ ssh oist-username@saion cp /work/myUnit/myresult.dat /bucket/myUnit/
Then remove all the original data on work:
$ rm -r $tmpdir
Now you've reclaimed the space on work and it's ready to use for the next job and your other unit members.
A full example
Here is an example of a complete Slurm script that creates a temporary directory, then :
#!/bin/bash -l #SBATCH -p short #SBATCH -t 0-1 #SBATCH --mem=20G # create a temporary directory for this job and save the name tempdir=$(mktemp -d /work/MyUnit/myprog.XXXXXX) # enter the temporary directory cd $tempdir # Start 'myprog' with input from bucket, # and output to our temporary directory myprog /bucket/MyUnit/mydata.dat -o output.dat # copy our result back to Bucket. We use "scp" to copy the data # back as bucket isn't writable directly from the compute nodes. scp output.dat saion:/bucket/MyUnit/ # Clean up by removing our temporary directory rm -r $tempdir
Section: The open Saion partitions.
Section: The restricted Saion partitions.