Gpu Workloads /
Deploy GPUs on Northflank's managed cloud
You can run GPU workloads on Northflank's managed cloud.
This allows you to use GPU acceleration for your applications and services and only pay for the resources consumed. To deploy GPU workloads on Northflank you must create a GPU-enabled project.
You can deploy combined and deployment services and jobs with GPU access. Northflank GPUs do not currently support timeslicing.
Check the pricing page to find out more about GPU pricing.
Deploy a GPU-enabled project
To deploy GPU workloads on Northflank's managed cloud, you'll first need to create a new project in a GPU-enabled region.
Select your desired region from the list under GPU, and create your project as normal. Any services or jobs deployed in this project will have a GPU option available in resources.
Different regions may have different availability of specific GPU models.
Deploy a GPU workload on Northflank
To deploy workloads with GPU access, create a new service or job as normal, or navigate to the resources page of an existing service or job.
Select an available GPU from the drop-down list as well as the number of GPUs to deploy with. Each instance of a service will have access to the number of GPU models selected.
GPU Count | Instances deployed | GPUs per instance | Total GPUs in the service |
---|---|---|---|
1 | 1 | 1 | 1 |
4 | 1 | 4 | 4 |
8 | 2 | 8 | 16 |
You may need to configure your workload to make use of GPU hardware, or multiple GPUs.
Next steps
Configure applications to use GPUs
You can directly deploy or build your applications with Docker images that are optimised for your desired GPU model and AI/ML libraries.
Build with GPU-optimised images
You can directly deploy or build your applications with Docker images that are optimised for your desired GPU model and AI/ML libraries.
Right-size resources for GPU workloads
Scale CPU, memory, and ephemeral storage to handle GPU workloads.
Persist models and data
You can directly deploy or build your applications with Docker images that are optimised for your desired GPU model and AI/ML libraries.