You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: CHANGELOG.md
+15-16
Original file line number
Diff line number
Diff line change
@@ -17,15 +17,20 @@ This file is used to list changes made in each version of the AWS ParallelCluste
17
17
18
18
**ENHANCEMENTS**
19
19
- Add support for RHEL8.
20
-
- Enforce the DCV Authenticator Server to use at least `TLS-1.2` protocol when creating the SSL Socket.
21
-
- Track head node memory and root volume disk utilization using the `mem_used_percent` and `disk_used_percent` metrics collected through the CloudWatch Agent.
22
-
- Add log rotation support for ParallelCluster managed logs.
23
20
- Add support for customizing the cluster Slurm configuration via the ParallelCluster configuration YAML file.
21
+
- Build Slurm with support for LUA.
24
22
- Add health check manager and GPU health check, which can be activated through cluster configuration.
25
23
Health check manager execution is triggered by a Slurm prolog script. GPU check verifies healthiness of a node by executing NVIDIA DCGM L2 diagnostic.
26
-
- Build Slurm with support for LUA.
24
+
- Add log rotation support for ParallelCluster managed logs.
25
+
- Track head node memory and root volume disk utilization using the `mem_used_percent` and `disk_used_percent` metrics collected through the CloudWatch Agent.
26
+
- Enforce the DCV Authenticator Server to use at least `TLS-1.2` protocol when creating the SSL Socket.
27
27
28
28
**CHANGES**
29
+
- Upgrade Slurm to version 23.02.1.
30
+
- Upgrade munge to version 0.5.15.
31
+
- Set Slurm default `TreeWidth` to 30.
32
+
- Set Slurm prolog and epilog configurations to target a directory, `/opt/slurm/etc/scripts/prolog.d/` and `/opt/slurm/etc/scripts/epilog.d/` respectively.
33
+
- Set Slurm `BatchStartTimeout` to 3 minutes so to allow max 3 minutes Prolog execution during compute node registration.
29
34
- Upgrade EFA installer to `1.22.1`
30
35
- Dkms : `2.8.3-2`
31
36
- Efa-driver: `efa-2.1.1g`
@@ -34,20 +39,14 @@ This file is used to list changes made in each version of the AWS ParallelCluste
34
39
- Libfabric-aws: `libfabric-aws-1.17.1-1`
35
40
- Rdma-core: `rdma-core-43.0-1`
36
41
- Open MPI: `openmpi40-aws-4.1.5-1`
37
-
- Upgrade Lustre client version to 2.12 on Amazon Linux 2. Lustre client 2.12 has been installed on Ubuntu 20.04, 18.04 and CentOS >= 7.7. Upgrade Lustre client version to 2.10.8 on CentOS 7.6.
38
-
- Upgrade aws-cfn-bootstrap to version 2.0-24.
39
-
- Set Slurm prolog and epilog configurations to target a directory, /opt/slurm/etc/scripts/prolog.d/ and /opt/slurm/etc/scripts/epilog.d/ respectively.
40
-
- Set Slurm BatchStartTimeout to 3 minutes so to allow max 3 minutes Prolog execution during compute node registration.
41
-
- Upgrade Slurm to version 23.02.1.
42
-
- Upgrade munge to version 0.5.15.
43
-
- Set Slurm default `TreeWidth` to 30.
44
-
45
-
**BUG FIXES**
46
-
- Fix IP association on instances with multiple network cards.
47
-
- Fix replacement of StoragePass in slurm_parallelcluster_slurmdbd.conf when a queue parameter update is performed and the Slurm accounting configurations are not updated.
42
+
- Upgrade Lustre client version to 2.12 on Amazon Linux 2 (same version available on Ubuntu 20.04, 18.04 and CentOS >= 7.7).
43
+
- Upgrade Lustre client version to 2.10.8 on CentOS 7.6.
44
+
- Upgrade `aws-cfn-bootstrap` to version 2.0-24.
48
45
49
46
**BUG FIXES**
50
-
- Fix issue causing cfn-hup daemon to fail when it gets restarted.
47
+
- Fix an issue that was causing misalignment of compute nodes IP on instances with multiple network interfaces.
48
+
- Fix replacement of `StoragePass` in `slurm_parallelcluster_slurmdbd.conf` when a queue parameter update is performed and the Slurm accounting configurations are not updated.
49
+
- Fix issue causing `cfn-hup` daemon to fail when it gets restarted.
0 commit comments