Providers Profiles#
Setting Up Cloud Providers#
A provider is identified by its platform name and region and primary zone. To integrate a provider, you need to create a ProviderProfile resource in skycluster-system namespace. The following examples show how to configure the a ProfileProvider for the major cloud providers: AWS, GCP, and Azure.
apiVersion: core.skycluster.io/v1alpha1
kind: ProviderProfile
metadata:
name: aws-us-east-1
namespace: skycluster-system
spec:
platform: aws # Platform can be aws, azure, gcp
region: us-east-1 # Region identifier
regionAlias: us-east
continent: north-america
enabled: true
zones:
- name: us-east-1a # Zone identifier
locationName: us-east-1a
defaultZone: true # Only one zone can be default
enabled: true
type: cloud # Optional
- name: us-east-1b # Zone identifier
locationName: us-east-1b
defaultZone: false
enabled: true
type: cloud # Optional
apiVersion: core.skycluster.io/v1alpha1
kind: ProviderProfile
metadata:
name: gcp-us-east1
namespace: skycluster-system
spec:
platform: gcp # Platform can be aws, azure, gcp
region: us-east1 # Region identifier
regionAlias: us-east
continent: north-america
enabled: true
zones:
- name: us-east1-b # Zone identifier
locationName: us-east1-b
defaultZone: true # Only one zone can be default
enabled: true
type: cloud # Optional
- name: us-east1-c # Zone identifier
locationName: us-east1-c
defaultZone: false
enabled: true
type: cloud # Optional
Warning
Azure support is limited and some features may not work.
apiVersion: core.skycluster.io/v1alpha1
kind: ProviderProfile
metadata:
name: azure-eastus
namespace: skycluster-system
spec:
platform: azure # Platform can be aws, azure, gcp
region: eastus # Region identifier
regionAlias: eastus
continent: north-america
enabled: true
zones:
- name: "1" # Zone identifier
locationName: eastus-1
defaultZone: true # Only one zone can be default
enabled: true
type: cloud # Optional
- name: "2" # Zone identifier
locationName: eastus-2
defaultZone: false
enabled: true
type: cloud # Optional
SkyCluster automatically detects available images and instance types for major cloud providers such as AWS, Azure, and GCP. For OpenStack and baremetal, these services must be configured manually
After creating a ProfileProvider resource, the SkyCluster operator generates a ConfigMap in the skycluster-system namespace containing the available images and instance types for that provider. Verify that the provider profile is ready by checking its status:
kubectl get providerprofile -n skycluster-system
# NAME REGION READY
# aws-us-east-1 us-east-1 Ready
Setting Up On-premises Providers#
OpenStack#
Similar to cloud provider, a edge or private provider is identified by its platform name and region
and primary zone. In addition to ProfileProvider resource, you also need to create dependency
resources to for the provider. The following example shows how to configure the OpenStack
provider for the on-premises savi edge cluster with scinet region with primary zone default.
Provider Profiles#
apiVersion: core.skycluster.io/v1alpha1
kind: ProviderProfile
metadata:
name: savi-scinet-default
namespace: skycluster-system
spec:
platform: openstack
region: scinet
regionAlias: scinet # Optional
continent: north-america # Optional
enabled: true
zones:
- name: default
locationName: default # Optional
defaultZone: true
enabled: true
type: edge
Note
ProviderProfile status becomes Ready when the dependency resources are created for the provider. Check out
the status of the ProviderProfile resource by running the following command. You should see the Ready
status set to False when you create the object for the first time. It will change to True once the
dependency objects are created.
kubectl get providerprofile -n skycluster-system
# NAME REGION READY
# savi-scinet-default scinet False
The operator creates a config map for each provider profile that constains offerings by the provider. You can verify the config map by running the following command:
# Change the label value to match your provider profile name
kubectl get configmap -n skycluster-system \
-l skycluster.io/provider-profile=savi-scinet-default
Dependency Resources#
The following dependency resources are needed to
ensure ProviderProfile becomes ready and can be used:
Images#
You need to configure the available images manually in skycluster-system namespace for on-premises providers.
The following example shows how to configure images including ubuntu-20.04, ubuntu-22.04 and ubuntu-24.04 for the savi-scinet-default provider profile for the scinet region.
apiVersion: core.skycluster.io/v1alpha1
kind: Image
metadata:
name: savi-scinet-default-images
namespace: skycluster-system
spec:
providerRef: savi-scinet-default
# Must match the provider profile name in the Provider resource
images:
- zone: default
nameLabel: ubuntu-20.04
name: ubuntu-20.04 # Local identifier corresponding to the image
- zone: default
nameLabel: ubuntu-22.04
name: ubuntu-22.04 # Local identifier corresponding to the image
- zone: default
nameLabel: ubuntu-24.04
name: ubuntu-24.04 # Local identifier corresponding to the image
Check the status of Image resource by running the following command:
kubectl get images.core.skycluster.io -n skycluster-system
# NAME REGION READY
# savi-scinet-default-images scinet True
Instance Type#
To configure the instance types, you need to create an InstanceType resource in skycluster-system namespace. The following example shows how to introduce available instance types.
apiVersion: core.skycluster.io/v1alpha1
kind: InstanceType
metadata:
name: savi-scinet-default-instance-types
namespace: skycluster-system
spec:
providerRef: savi-scinet-default
# Must match the provider name in the Provider resource
offerings:
- zone: default
zoneOfferings:
- name: m1.small
nameLabel: 1vCPU-2GB
price: "0.0015"
ram: 2GB
vcpus: 1
- name: n1.small
nameLabel: 1vCPU-4GB
vcpus: 1
ram: 4GB
price: "0.02"
Check the status of the InstanceType resource by running the following command:
kubectl get instancetype.core.skycluster.io -n skycluster-system
# NAME REGION READY
# savi-scinet-default-instance-types scinet True
Once all dependency resources are created and ready, the ProviderProfile status changes to Ready. At this point, you can use the provider profile to create resources in your cluster. To verify its status, run:
kubectl get providerprofile -n skycluster-system
# NAME REGION READY
# savi-scinet-default scinet True
On-premises Custom Providers#
On-premises providers can be registered similarly to cloud providers but you must also create DeviceNode resources in skycluster-system namespace to supply on-premises configuration such as the gateway node address, required keys, and edge-device details.
Provider Profiles#
apiVersion: core.skycluster.io/v1alpha1
kind: ProviderProfile
metadata:
name: savi-toronto-default
namespace: skycluster-system
spec:
platform: baremetal
region: toronto
regionAlias: toronto # Optional
continent: north-america # Optional
enabled: true
zones:
- name: default
locationName: BahenBuilding # Optional
defaultZone: true
enabled: true
type: edge
The baremetal type requires specifying gateway connection and access details, along with worker node data and their capabilities. You must configure this by creating DeviceNode resources:
- Device Nodes
The DeviceNode API represents an edge device that can run workloads. A DeviceNode resource can be defined as either a gateway or a worker node by setting its type field. By default the gateway node does not run any workload. The DeviceNode resource are created in the skycluster-system namespace.
apiVersion: core.skycluster.io/v1alpha1
kind: DeviceNode
metadata:
name: savi-toronto-gw
namespace: skycluster-system
spec:
providerRef: savi-toronto-default
deviceSpec:
type: gateway
zone: default
publicIp: x.y.z.w
# by default the subnet mask is /24 for this network
privateIp: 10.23.100.22
auth:
privateKeySecretRef:
# secret containing the private SSH key, see below
name: savi-toronto-edge-ssh-key
key: privateKey
username: ubuntu
apiVersion: core.skycluster.io/v1alpha1
kind: DeviceNode
metadata:
name: savi-toronto-edge-jetson-nano1
namespace: skycluster-system
spec:
providerRef: savi-toronto-edge
deviceSpec:
type: worker
zone: default
privateIp: 10.23.100.200
# must be reachable from the gateway node
auth:
username: ubuntu
# same private key as the gateway node will be used
configs:
name: Jetson Nano
cpus: 1
ram: 4GB
gpu:
count: 472
unit: GFLOPS # GFLOPS | TFLOPS | TOPS | GPU
enabled: true
manufacturer: NVIDIA
memory: Maxwell
model: JetsonNano
storage: 100GB
price: "0"
NVIDIA Device Preparation#
SkyCluster supports DeviceNode resources equipped with NVIDIA GPUs by installing a local K3s cluster and joining the devices as worker nodes. To enable GPU workloads, the devices must have the NVIDIA Container Runtime installed. Since installation steps vary depending on the device specifications (e.g., Jetson, desktop GPUs), you need to follow the official NVIDIA instructions to set up the toolkit. This toolkit allows containers to access the GPU resources on the Jetson device, enabling GPU-accelerated applications to run smoothly.
After setting up the toolkit, ensure that your containerized applications are configured to utilize the GPU resources effectively. You may need to specify the appropriate runtime and environment variables in your container configurations to leverage the GPU capabilities.