Aws-parallelcluster

Latest version: v3.9.1

Safety actively analyzes 621803 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 1 of 15

3.10.0

------

**BUG FIXES**
- Fix DRA configuration to make `AutoExportPolicy` and `AutoImportPolicy` optional.
- Consider Compute fleet clean-up completed during cluster deletion when instances are either in shutting-down or terminated state.
This is to avoid cluster deletion failure for instance types with longer termination cycles.
- Allow cloudwatch dashboard to be enabled and alarms to be disabled in the `Monitoring` section of the cluster config.

3.9.1

------

**BUG FIXES**
- Fix an issue that may cause data loss when a shared storage is unmounted as part of a cluster update
while it is in use and with no backup strategy in place.

3.9.0

------

**ENHANCEMENTS**
- Permit to update the external shared storage of type Efs, FsxLustre, FsxOntap, FsxOpenZfs and FileCache
without replacing compute and login fleet.
- Permit to update `MinCount`, `MaxCount`, `Queue` and `ComputeResource` configuration parameters without the need to
stop the compute fleet. It's now possible to update them by setting `Scheduling/SlurmSettings/QueueUpdateStrategy`
to TERMINATE. ParallelCluster will terminate only the nodes removed during a resize of the cluster capacity
performed through a cluster update.
- Add support for RHEL9.
- Add support for Rocky Linux 9 as `CustomAmi` created through `build-image` process. No public official ParallelCluster Rocky9 Linux AMI is made available at this time.
- Remove `CommunicationParameters` from the Custom Slurm Settings deny list.
- Add the configuration parameter `DeploymentSettings/DefaultUserHome` to allow users to move the default user's home directory to `/local/home` instead of `/home` (default).
- Add configuration parameter `DeploymentSettings/DisableSudoAccessForDefaultUser` to disable sudo access of default user in supported OSes.

**CHANGES**
- Upgrade Slurm to 23.11.4 (from 23.02.7).
- Upgrade Pmix to 4.2.9 (from 4.2.6).
- Add support for Python 3.11, 3.12 in pcluster CLI and aws-parallelcluster-batch-cli.
- Build network interfaces using network card index from `NetworkCardIndex` list of EC2 DescribeInstances response,
instead of looping over `MaximumNetworkCards` range.
- Fail cluster creation when using instance types P3, G3, P2 and G2 because their GPU architecture is not compatible with Open Source Nvidia Drivers (OpenRM) introduced as part of 3.8.0 release.
- Upgrade the default FSx Lustre server version managed by ParallelCluster to 2.15.
- Upgrade NVIDIA driver to version 535.154.05.
- Upgrade EFA installer to `1.30.0`.
- Efa-driver: `efa-2.6.0-1`
- Efa-config: `efa-config-1.15-1`
- Efa-profile: `efa-profile-1.6-1`
- Libfabric-aws: `libfabric-aws-1.19.0`
- Rdma-core: `rdma-core-46.0-1`
- Open MPI: `openmpi40-aws-4.1.6-2` and `openmpi50-aws-5.0.0-11`
- Upgrade NICE DCV to version `2023.1-16388`.
- server: `2023.1.16388-1`
- xdcv: `2023.1.565-1`
- gl: `2023.1.1047-1`
- web_viewer: `2023.1.16388-1`
- Upgrade ARM PL to version 23.10.
- Upgrade third-party cookbook dependencies:
- nfs-5.1.2 (from nfs-5.0.0)


**BUG FIXES**
- Refactor IAM policies defined in CloudFormation template `parallelclutser-policies.yaml` to prevent ParallelCluster API deployment failure
caused by policies exceeding IAM limits.

3.8.0

------

**ENHANCEMENTS**
- Add support for EC2 Capacity Blocks for ML.
- Add support for Rocky Linux 8 as `CustomAmi` created through `build-image` process. No public official ParallelCluster Rocky8 Linux AMI is made available at this time.
- Add `Scheduling/ScalingStrategy` parameter to control the cluster scaling strategy to use when launching EC2 instances for Slurm compute nodes.
Possible values are `all-or-nothing`, `greedy-all-or-nothing`, `best-effort`, with `all-or-nothing` being the default.
- Add `HeadNode/SharedStorageType` parameter to use EFS storage instead of NFS exports from the head node root volume
for intra-cluster shared file system resources: ParallelCluster, Intel, Slurm, and `/home` data. This enhancement reduces the load on the head node networking.
- Allow for mounting `home` as an EFS or FSx external shared storage via the `SharedStorage` section of the config file.
- Add new parameter `SlurmSettings/MungeKeySecretArn` to permit to use an external user-defined MUNGE key from AWS Secrets Manager.
- Add `Monitoring/Alarms/Enabled` parameter to toggle Amazon CloudWatch Alarms for the cluster.
- Add head node alarms to monitor EC2 health checks, CPU utilization and the overall status of the head node, and add them to the CloudWatch Dashboard created with the cluster.
- Add support for Data Repository Associations when using `PERSISTENT_2` as `DeploymentType` for a managed FSx for Lustre.
- Add `Scheduling/SlurmSettings/Database/DatabaseName` parameter to allow users to specify a custom name for the database on the database server to be used for Slurm accounting.
- Make `InstanceType` an optional configuration parameter when configuring `CapacityReservationTarget/CapacityReservationId` in the compute resource.
- Add possibility to specify a prefix for IAM roles and policies created by ParallelCluster API.
- Add possibility to specify a permissions boundary to be applied for IAM roles and policies created by ParallelCluster API.
- Add support for il-central-1 region.

**CHANGES**
- Upgrade Slurm to 23.02.7 (from 23.02.6).
- Upgrade NVIDIA driver to version 535.129.03.
- Upgrade CUDA Toolkit to version 12.2.2.
- Use Open Source NVIDIA GPU drivers (OpenRM) as NVIDIA kernel module for Linux instead of NVIDIA closed source module.
- This change removes support for P3, G3, P2 and G2 instances with GPU architecture not supported by OpenRM. The Open Source Nvidia only works on platforms that have the GSP (GPU System Processor).
- Remove support of `all_or_nothing_batch` configuration parameter in the Slurm resume program, in favor of the new `Scheduling/ScalingStrategy` cluster configuration.
- Changed cluster alarms naming convention to '[cluster-name]-[component-name]-[metric]'.
- Change default EBS volume types in ADC regions from `gp2` to `gp3`, for both the root and additional volumes.
- The optional permissions boundary for the ParallelCluster API is now applied to every IAM role created by the API infrastructure.
- Upgrade EFA installer to `1.29.1`.
- Efa-driver: `efa-2.6.0-1`
- Efa-config: `efa-config-1.15-1`
- Efa-profile: `efa-profile-1.5-1`
- Libfabric-aws: `libfabric-aws-1.19.0-1`
- Rdma-core: `rdma-core-46.0-1`
- Open MPI: `openmpi40-aws-4.1.6-1`
- Upgrade GDRCopy to version 2.4 in all supported OSes, except for Centos 7 where version 2.3.1 is used.
- Upgrade `aws-cfn-bootstrap` to version 2.0-28.
- Add support for Python 3.10 in aws-parallelcluster-batch-cli.

**BUG FIXES**
- Fix inconsistent scaling configuration after cluster update rollback when modifying the list of instance types declared in the Compute Resources.
- Fix users SSH keys generation when switching users without root privilege in clusters integrated with an external LDAP server through cluster configuration files.
- Fix disabling Slurm power save mode when setting `ScaledownIdletime = -1`.
- Fix hard-coded path to Slurm installation dir in `update_slurm_database_password.sh` script for Slurm Accounting.

3.7.2

------
**CHANGES**
- Upgrade Slurm to 23.02.6.

3.7.1

------

**CHANGES**
- Upgrade Slurm to 23.02.5 (from 23.02.4).
- Upgrade Pmix to 4.2.6 (from 3.2.3).
- Upgrade libjwt to 1.15.3 (from 1.12.0).
- Upgrade EFA installer to `1.26.1`, fixing RDMA writedata issue in P5.
- Efa-driver: `efa-2.5.0-1`
- Efa-config: `efa-config-1.15-1`
- Efa-profile: `efa-profile-1.5-1`
- Libfabric-aws: `libfabric-aws-1.18.2-1`
- Rdma-core: `rdma-core-46.0-1`
- Open MPI: `openmpi40-aws-4.1.5-4`

Page 1 of 15

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.