Using with High-performance Storage

Requirements

  • Required conditions for creating a Managed GPU cluster (Active service, metal cloud quota, SSH key, internal subnet LB, etc.).

  • Ensure that the File Storage – High Performance service is enabled and has been allocated a quota within the tenant.

  • To use a Mount Point belonging to the Metal cloud network, navigate to the File Storage – High Performance tab to create a new Mount Point following the instructions here

Enable the File Storage – High Performance

Integrate with a new Managed GPU cluster:

Step 1: On the FPT Cloud Portal menu, select AI Infrastructure → Managed GPU Cluster → Create a Managed GPU Cluster.

  • Select the correct network of the Metal Cloud server as the worker node in the GPU cluster, and the mountpoints of File Storage – High Performance will be displayed depending on this network.

Step 2: Once you have the MountPoint in the desired metal cloud network, enable File Storage – High Performance and select the desired MountPoint.

Note: If the tenant has not activated the File Storage – High Performance service, the following message will appear. You must submit a request to activate the service before performing integration on the Managed GPU Cluster.

Step 4: Review all High Performance Storage Integration information and proceed to create the Managed GPU Cluster.

Integrate with an existing Managed GPU cluster

Step 1: On the FPT Portal menu, select AI Infrastructure → Managed GPU Cluster → select an existing cluster to integrate File Storage – High Performance

Note: The Managed GPU cluster integrated with File Storage – High Performance must be in the Succeeded (Running) state before integration can be performed.

Step 2: In the Essential Properties tab → High Performance Storage Integration, click Enable High Performance Storage → select MountPoint from the list, then click the "Confirm" button.

The High Performance Storage integration process will take a few minutes, and the Cluster status will change to Processing until the integration is successful. The Cluster will continue to operate normally during the integration.

Remove File Storage – High Performance integration

  • Only remove the File Storage – High Performance integration when the cluster status is Succeeded (Running). Before removing the integration, delete all PVCs in the cluster using the selected mountpoint. Canceling the integration does not automatically delete data written by Kubernetes in the MountPoint directory.

  • Step 1: On the FPT Portal menu, select AI Infrastructure → Managed GPU Cluster → select the cluster that has integrated File Storage – High Performance

  • Step 2: In the Essential Properties tab → High Performance Storage Integration → disintegrate → Confirm

Modify Mount Point

At any given time, only one MountPoint can be used on a Managed GPU cluster. To change the MountPoint used in the cluster, you must first unmount the old MountPoint (Section 2.3) and then mount the new MountPoint for the cluster (Section 2.2).

Using the Mount Point in Cluster

Managed GPU cluster: After successful integration, the cluster will have a storageclass available to create Persistence Volumes (PVs) located in the directory assigned to the MountPoint path. The name of the storageclass is the name of the integrated MountPoint QoS Policy.

For example, if the MountPoint path is /k8s-cluster1, PVs created by CSI in Kubernetes will have paths such as /k8s-cluster1/PV1, /k8s-cluster1/PV2, etc.

  • Create a PersistentVolumeClaim (PVC) using the system's existing storageclass for the integrated MountPoint. Since the storageclass's VOLUMEBINDINGMODE is WaitForFirstConsumer, a Pod must use this PVC for CSI to create the PV and bind it to the PVC.

  • Note: Do not modify the cluster's default storageclass configuration. If the user changes that configuration, it will automatically roll back to the system's original configuration.

  • Example manifest of a PVC:

apiVersion: v1 

kind: PersistentVolumeClaim 

metadata: 

  name: csi-pvc-dynamic-1 

  namespace: default 

spec: 

  accessModes: 

    - ReadWriteMany 

  resources: 

    requests: 

      storage: 15Gi 

  storageClassName: k8s-tester 

  volumeMode: Filesystem
  • To resize the PVC capacity, directly edit the PVC resource in the spec.resources.requests.storage field.

  • Note: Capacity cannot be reduced (only increased). If PVC is being used by a Pod, the system will automatically resize the capacity of the mountPath in the Pod (resize volume online).

Last updated