HKUST ITSO AI Chatbot

Important reminder

Please do not disclose personal data such as your HKUST account number, staff/student ID or name in the chatbot. Information provided will be retained to enhance system performance.

By using the HKUST ITSO Chat service, you confirm that you have read, understood, and agreed to the Disclaimer


Log in to access additional information for your user group in addition to the publicly accessible content.

Send Icon
HPC4 Cluster

The HPC4 is the fourth-generation high-performance computing cluster implemented and maintained by ITSO.  It is officially rolled out in mid-Oct 2024 after a pilot testing period.  Hosted in the new High Performance Infrastructural Center (HPCIC), the HPC4 equipment is primarily funded by the University, while also accepting contributions from faculty members. The HPC4 platform is designed to support scientific computations with Intel-based and AMD-based CPU nodes, and it also features Nvidia-based GPU machines for tasks that do not require hardware specific for AI centric workloads like HKUST SuperPOD.

Available To
HKUST Researchers (Account applications are available to HKUST faculty members as Principal Investigators of their research projects)
Service Fee

HPC4 usage is charged based on a charging model.  For details, please visit Charging/Pricing Model of HPC4 

Service Hours

7x24

Quick Links

        

What's New

 

More announcements and training...

 

HPC4 Highlights

The HPCIC adopts Liquid Immersion Cooling Technology, which offers the following advantages:

• Allows higher density of computing resources resulting in more hardware in less physical space.

• Offers high energy efficiency for sustainability

• Operating cost reduction from energy saving in cooling of machines

 

The system environment of HPC4 is based on Rocky Linux 9. Secure computing approach would be adopted, with regular operating system upgrades and security patches to be applied to maintain a safe environment for research workloads.

 

User Documentation

New to our HPC environment? We provide a comprehensive HPC Handbook.

This documentation serves as a central knowledge base for AI and HPC users, offering detailed guides on:

  • Getting Started: Account setup and connecting to the cluster.
  • Job Scheduling: How to submit and manage jobs using SLURM.
  • Software Environment: Guides for Python, R, MATLAB, Java, and module management.
  • Operating System: Navigating the Linux environment.

 

HPC4 Hardware Specification

The HPC4 cluster consists of multiple types of node, each with distinct hardware configurations to cater to a variety of computational needs. The specifications for each type of nodes are detailed below.

CPU Nodes
Processor Nodes CPU Cores / Threads (per Node) System Memory (per Node)
Intel Xeon Platinum 8592+ 10 128 Cores / 128 Threads 512GB DDR5-5600 ECC
AMD EPYC 9754 76 + 1* 256 Cores / 256 Threads 768GB DDR5-4800 ECC
AMD EPYC 9754 18* + 12^ 256 Cores / 256 Threads 1.5TB DDR5-4800 ECC

GPU Nodes
GPU Model Host CPU (per Node) Nodes GPUs (per Node) GPU Memory (per GPU) System Memory (per Node)
NVIDIA A30 2× Intel Xeon Gold 6448Y 15 4 24GB HBM2 ECC 512GB DDR5-4800 ECC
NVIDIA L20 2× Intel Xeon Gold 6548Y+ 5 4 48GB GDDR6 ECC 512GB DDR5-4800 ECC
NVIDIA RTX 4090D 2× Intel Xeon Gold 6448Y 10* + 2 6 24GB GDDR6X 512GB DDR5-4800 ECC
NVIDIA RTX 5880 Ada 2× Intel Xeon Gold 6448Y 8 6 48GB GDDR6 ECC 512GB DDR5-4800 ECC

 

* Contributed servers. Not yet available to public use.

^ To be available soon.

Shared Resources
  • Network: All nodes are interconnected with a 200Gbps RoCE (RDMA over Converged Ethernet) internode fabric.
  • Cluster Storage: A 5.4PB Dell PowerScale storage system is available, shared with other clusters.

 

Performance Comparison

To give an idea of the performance of CPUs and GPUs in HPC4 with others, please check here.

 

HPC4 Account Application

All HKUST faculty members are eligible to apply for a HPC4 account.  To apply, please complete the HPC4 Account Application Form.  Students who wish to utilize HPC4 should consult their supervisors to support their applications by completing the above application form.

 

Getting Started

How to login to the cluster

Click here to view the instructions on how to get access to the HKUST HPC4 cluster

Use of SLURM Job Scheduling System

The Simple Linux Utility for Resource Management (SLURM) is the resource management and job scheduling system of the cluster. All jobs in the cluster must be run with the SLURM.

Click here to learn how to submit your first SLURM job

Click here to view details of using SLURM

Partition and Resource Quota

Click here to view more information on partition and resource quota.

Storage Types

Click here to view more information on different storage types.

 

HPC4 Software Environment

The software environment promotes a Do-It-Yourself installation approach using Spack and Apptainer, providing the flexibility to customize their software environment.

Modules

Lmod is used to manage installations for most application software. With the modules system, the shell environment can be set up to give access to applications and make running and compiling software easier.

Click here for details of the module system.

Spack - User-managed software installation manager

Spack is a package manager that helps building and managing multiple versions and configurations of software.

Click here for details of Spack.

Use of Apptainer (Singularity)

Apptainer (formerly known as Singularity) container lets user run applications in a Linux environment of their choice. Apptainer literally enables BYOE (Bring-Your-Own-Environment) computing in the multi-tenant and shared HPC cluster.

Click here to view details of Apptainer (Singularity)

 

HPC4 Policies: Charging & Contribution

Charging Model

Charging for HPC4 services ensures efficient resource allocation and sustainability. It encourages accountability and fair usage among researchers.

Click here for details on the charging for use of HPC4.

Contribution Model

Subject to HPCIC resource availability, the HPC4 adopts community cluster model similar to HPC3 and accepts hardware contribution from faculty members on regular basis.

Details of the HPC4 contribution model are available here.

 

Usage Tips

Specific Use Cases

Support

General Enquiries cchelp@ust.hk
Suggestions & Complaints cclisten@ust.hk
Serviceline +852-2358-6200