copyright | lastupdated | keywords | subcollection | ||
---|---|---|---|---|---|
|
2024-11-22 |
hpc-spectrum-symphony |
{:shortdesc: .shortdesc} {:codeblock: .codeblock} {:screen: .screen} {:external: target="_blank" .external} {:pre: .pre} {:tip: .tip} {:note: .note} {:important: .important} {:faq: data-hd-content-type='faq'}
{: #spectrum-symphony-faqs}
{: #terraform-location} {: faq}
The Terraform-based templates can be found in this public GitHub repository{: external}.
What Spectrum Symphony and Spectrum Scale versions are used in cluster nodes deployed with this offering?
{: #my-faq-packages} {: faq}
Cluster nodes that are deployed with this offering include {{site.data.keyword.symphony_full_notm}} 7.3.2 Advanced Edition. See the following for a summary of the features associated with each edition: {{site.data.keyword.symphony_full_notm}} editions{: external}.
If the cluster uses {{site.data.keyword.scale_short}} storage, the storage nodes include {{site.data.keyword.scale_full_notm}} 5.2.1.1 software. For more information, see the {{site.data.keyword.scale_full_notm}}{: external} product documentation.
{: #locations-vpc-resources} {: faq}
Available regions and zones for deploying VPC resources and mapping them to city locations and data centers can be found in Locations for resource deployment.
{: #permissions-cluster-offering} {: faq}
Instructions for setting the appropriate permissions for {{site.data.keyword.cloud_notm}} services that are used by the offering to create a cluster can be found in Granting user permissions for VPC resources, Managing user access for Schematics, and Assigning access to Secrets Manager.
{: #ssh-among-nodes} {: faq}
All the nodes in the HPC cluster have the same public key that you register at your cluster creation. You can use ssh-agent forwarding, which is a common technique to access remote nodes that have the same public key. It automates to securely forward private keys to remote nodes. Forwarded keys are deleted immediately after a session is closed.
To securely forward private keys to remote nodes, you need to do ssh-add
and ssh -A
.
[your local PC]~$ ssh-add {id_rsa for symphony cluster}
[your local PC]~# ssh -A -J root@jumpbox_fip root@management_private_ip
...
[root@management]~# ssh -A worker_private_ip
{: codeblock}
For Mac OS X, you can persist ssh-add
by adding the following configuration to .ssh/config
:
Host *
UseKeychain yes
AddKeysToAgent yes
{: codeblock}
You can even remove -A
by adding "ForwardAgent yes" to .ssh/config
.
{: #worker-nodes} {: faq}
Before deploying a cluster, it is important to ensure that the VPC resource quota settings are appropriate for the size of the cluster that you would like to create (see Quotas and service limits).
The maximum number of worker nodes that are supported for the deployment value worker_node_max_count
is 500 (see Deployment values). The worker_node_min_count
variable specifies the number of worker nodes that are provisioned at the time that the cluster is created, which will exist throughout the life of the cluster. The delta between those two variables specifies the maximum number of worker nodes that can either be created or destroyed by the Symphony Host Factory auto-scaling feature.
The {{site.data.keyword.symphony_short}} offering supports both bare metal worker nodes and {{site.data.keyword.scale_short}} storage nodes. The following combinations of values are supported:
- If
worker_node_type
is set asbaremetal
, a maximum of 16 bare metal nodes are supported. - If
spectrum_scale_enabled
is set totrue
andstorage_type
is set aspersistent
, a maximum of 10 bare metal nodes are supported.
For more information, see Deployment values.
When creating or deleting a cluster with many worker nodes, you might encounter VPC resource provisioning or deletion failures. In those cases, running the {{site.data.keyword.bpshort}} apply
or destroy
operation again might result in the remaining resources being successfully provisioned or deleted. If you continue to see errors, see Getting help and support.
Why are there two different resource group parameters that can be specified in the IBM Cloud catalog tile?
{: #resource-group-parameters} {: faq}
The first resource group parameter entry in the Configure your workspace section in the {{site.data.keyword.Bluemix_notm}} catalog applies to the resource group where the {{site.data.keyword.bpshort}} workspace is provisioned on your {{site.data.keyword.Bluemix_notm}} account. The value for this parameter can be different than the one used for the second entry in the Parameters with default values section in the catalog. The second entry applies to the resource group where VPC resources are provisioned. As specified in the description for this second resource_group
parameter, only the default resource group is supported for use of the Symphony Host Factory auto-scaling feature.
Can I use the Spectrum Symphony Host Factory feature for auto scaling on any cluster deployed with this offering?
{: #host-factory-auto-scaling} {: faq}
No, the use of Host Factory to provision and delete compute nodes is not supported in the following cases:
- Provisioning and deleting compute nodes on dedicated hosts. Only static compute nodes can be deployed on dedicated hosts.
- When using {{site.data.keyword.scale_short}} for shared storage in the cluster.
{: #custom-image-name-mappings} {: faq}
The mappings can be found in the image-map.tf
file and the scale-image-map.tf
file in this public GitHub repository{: external}.
{: #egosh-command-not-working} {: faq}
This is expected behavior. Even after the {{site.data.keyword.bpshort}} web console shows that the cluster is successfully provisioned, there are still some tasks that run in the background for several minutes. Allow a few minutes (typically 2 minutes is sufficient) after the cluster gets provisioned for egosh
to be available.
Why does cluster creation by using dedicated hosts fail sometimes with the error status of dedicated host is failed
?
{: #cluster-creation-dedicated-hosts-fails} {: faq}
In some regions, dedicated hosts have a limitation on the number of virtual server instances that can be placed on them at one time. You can try to provision the cluster with a smaller number of virtual server instances to overcome this.
Why does Spectrum Scale not allow use of the default value of 0.0.0.0/0 for security group creation?
{: #security-group-default-value} {: faq}
For security reasons, {{site.data.keyword.scale_short}} does not allow you to provide a default value that would allow network traffice from any external device. Instead, you can provide the address of your user system (for example, by using https://ipv4.icanhazip.com/) or a range of multiple IP addresses.
{: #multiple-key-pair-support} {: faq}
Yes, the {{site.data.keyword.symphony_short}} offering supports multiple single key pairs that can be provided for access to all of the nodes that are part of the cluster. In addition, {{site.data.keyword.symphony_short}} has a feature where each node of the cluster can be accessed through passwordless SSH.
{: #available-storage-types} {: faq}
In the {{site.data.keyword.symphony_short}} offering, you can use {{site.data.keyword.scale_short}} scratch storage or persistent storage. A scratch storage configuration uses virtual server instances with instance storage. A persistent storage configuration uses bare metal servers with locally attached NVMe storage.
{: #supported-operating-system-worker-nodes} {: faq}
The solution supports custom images based on RHEL 8.10 for virtual server instance worker nodes, and it supports the use of the stock RHEL 8.10 VPC images for bare metal worker nodes. At this time, custom images are not supported for use with VPC bare metal servers.
{: #custom-resolver} {: faq}
Yes, the solution supports the use of a custom resolver that is already associated to a VPC. If a VPC already has a custom resolver, the automation uses of it and the DNS service and associates the new DNS domain that is created from the solution for hostname resolution.
{: #DNS-zone-names} {: faq}
No, adding the same permitted network (for example, VPC) to two DNS zones of the same name is not allowed as mentioned here. Therefore, when you select values for vpc_scale_storage_dns_domain
and vpc_worker_dns_domain
, ensure that they are unique and that there are no DNS zones that use either of those names that are already associated with the VPC that you might have specified in vpc_name
.
What file storage for {{site.data.keyword.cloud_notm}} Virtual Private Cloud (VPC) profiles are supported for the {{site.data.keyword.symphony_full_notm}} cluster shared storage?
{: #file-storages-for-vpc-profiles} {: faq}
{{site.data.keyword.filestorage_vpc_full_notm}} is a zonal file storage offering that provides NFS-based file storage services. You create file share mounts from a subnet in an availability zone within a region. You can also share them with multiple virtual server instances within the same zone within a vpc. {{site.data.keyword.symphony_full_notm}} supports the use of dp2 profiles.
Can you specify the total IOPS (input or output operations per second) for a file share when deploying an {{site.data.keyword.symphony_full_notm}} cluster?
{: #iops-cluster} {: faq}
Yes, when you deploy an {{site.data.keyword.symphony_full_notm}} cluster, you can choose the required IOPS value appropriate for your file share size.
How to share data, packages, or applications with {{site.data.keyword.symphony_full_notm}} compute nodes?
{: #shares} {: faq}
{{site.data.keyword.filestorage_vpc_full_notm}} with two file shares (/mnt/vpcstorage/tools
or /mnt/vpcstorage/data
), and up to five file shares, is provisioned to be accessible by both Spectrum Symphony management and compute nodes. To copy to a file share, SSH to the Spectrum Symphony management node and use your file copy of choice (such as scp, rsync, or IBM Aspera) to the appropriate file share.
{: #nodes} {: faq}
File sharing is implemented on Symphony nodes as follows:
-
RHEL Symphony Nodes
- IBM Cloud® VPC File Storage is used to support file sharing.
- Files are mounted on all RHEL nodes in the cluster according to the specifications defined in the
custom_file_shares
variable.
-
Windows Worker Nodes
- File sharing is implemented using integration with Cloud Object Storage (COS).
For more information, refer to the IBM HPC Spectrum Symphony Deployment Values documentation.