From 83de2251b0c7572974a2fc38329076f2e463f2b7 Mon Sep 17 00:00:00 2001 From: hyandt Date: Wed, 14 Aug 2024 14:13:28 -0600 Subject: [PATCH] kestrel release note updats --- .../Systems/Kestrel/kestrel_release_notes.md | 30 +++++++++++++++++-- docs/Documentation/Systems/Kestrel/running.md | 2 +- 2 files changed, 28 insertions(+), 4 deletions(-) diff --git a/docs/Documentation/Systems/Kestrel/kestrel_release_notes.md b/docs/Documentation/Systems/Kestrel/kestrel_release_notes.md index 2b09f380e..d011da111 100644 --- a/docs/Documentation/Systems/Kestrel/kestrel_release_notes.md +++ b/docs/Documentation/Systems/Kestrel/kestrel_release_notes.md @@ -2,7 +2,31 @@ *We will update this page with Kestrel release notes after major Kestrel upgrades.* -## July 29 - July 30 +## August 14, 2024 + +Jobs running on `debug` GPU nodes are now limited to half a GPU node's resources across one or two nodes. This is equivalent to 64 CPUs, 2 GPUs, and 180G of RAM on one node. `--exclusive` can no longer be used for GPU debug jobs. + +## August 9, 2024 + +As of 08/09/2024 we have released new modules for VASP on Kestrel CPUs: + +``` +------------ /nopt/nrel/apps/cpu_stack/modules/default/application ------------- + #new modules: + vasp/5.4.4+tpc vasp/6.3.2_openMP+tpc vasp/6.4.2_openMP+tpc + vasp/5.4.4_base vasp/6.3.2_openMP vasp/6.4.2_openMP + + #legacy modules will be removed during next system time: + vasp/5.4.4 vasp/6.3.2 vasp/6.4.2 (D) +``` + + What’s new: + + * New modules have been rebuilt with the latest Cray Programming Environment (cpe23), updated compilers, and math libraries. + * OpenMP capability has been added to VASP 6 builds. + * Modules that include third-party codes (e.g., libXC, libBEEF, VTST tools, and VASPsol) are now denoted with +tpc. Use `module show vasp/` to see details of a specific version. + +## July 29 - July 30, 2024 1. Two [GPU login nodes](../Kestrel/index.md) were added. Use the GPU login nodes for compiling software to run on GPU nodes and for submitting GPU jobs. 1. GPU compute nodes were made available for general use and additional GPU partitions were added. See [Running on Kestrel](../Kestrel/running.md) for additional information and recommendations. @@ -51,7 +75,7 @@ Intel-oneapi-compilers. * The 2024 version is now added. -## April 12 - April 17 +## April 12 - April 17, 2024 1. The size of the [shared node partition](./running.md#shared-node-partition) was doubled from 32 nodes to 64 nodes. @@ -62,7 +86,7 @@ Intel-oneapi-compilers. 4. `/kfs2/pdatasets` was renamed to `/kfs2/datasets` and a symlink `/datasets` was added. -## Jan. 29 - Feb. 14 Upgrades +## Jan. 29 - Feb. 14, 2024 Upgrades 1. We have experienced that most previously built software runs without modification (this includes NREL provided modules) and performs at the same level. diff --git a/docs/Documentation/Systems/Kestrel/running.md b/docs/Documentation/Systems/Kestrel/running.md index e90f2a3b2..ca73d7694 100644 --- a/docs/Documentation/Systems/Kestrel/running.md +++ b/docs/Documentation/Systems/Kestrel/running.md @@ -39,7 +39,7 @@ The following table summarizes the partitions on Kestrel: | Partition Name | Description | Limits | Placement Condition | | -------------- | ------------- | ------ | ------------------- | -| ```debug``` | Nodes dedicated to developing and
troubleshooting jobs. Debug nodes with each of the non-standard
hardware configurations are available.
The node-type distribution is:
- 2 bigmem nodes
- 2 nodes with 1.7 TB NVMe
- 4 standard nodes
- 2 GPU nodes (shared)
**10 total nodes** | 1 job with a max of 2 nodes per user.
2 GPUs per user.
01:00:00 max walltime. | ```-p debug```
or
```--partition=debug``` | +| ```debug``` | Nodes dedicated to developing and
troubleshooting jobs. Debug nodes with each of the non-standard
hardware configurations are available.
The node-type distribution is:
- 2 bigmem nodes
- 2 nodes with 1.7 TB NVMe
- 4 standard nodes
- 2 GPU nodes (shared)
**10 total nodes** | - 1 job with a max of 2 nodes per user.
- 2 GPUs per user.
- 1/2 GPU node resources per user (Across 1-2 nodes).
- 01:00:00 max walltime. | ```-p debug```
or
```--partition=debug``` | |```short``` | Nodes that prefer jobs with walltimes
<= 4 hours. | 2016 nodes total.
No limit per user. | ```--time <= 4:00:00```
```--mem <= 246064```
```--tmp <= 1700000 (256 nodes)```| | ```standard``` | Nodes that prefer jobs with walltimes
<= 2 days. | 2106 nodes total.
1050 nodes per user. | ```--mem <= 246064```
```--tmp <= 1700000```| | ```long``` | Nodes that prefer jobs with walltimes > 2 days.
*Maximum walltime of any job is 10 days*| 525 nodes total.
262 nodes per user.| ```--time <= 10-00```
```--mem <= 246064```
```--tmp <= 1700000 (256 nodes)```|