Scale /
Autoscale deployments
You can configure autoscaling to automatically respond to increases in activity on your deployments, and ensure you have the resources needed to process your traffic even at peak times.
Horizontal autoscaling allows your services to automatically scale up the number of instances for your deployment. The Northflank load-balancer will evenly distribute incoming traffic between different instances of a deployment.
Configure horizontal autoscaling
To enable autoscaling, expand advanced resource options in your deployment's resource page, or in the resources section when creating your deployment. This is available in combined and deployment services.
Select enable horizontal autoscaling to turn on autoscaling for the deployment and view the configuration options:
- Minimum instances: set the minimum number of instances the deployment will scale down to. This should be set to the number of instances that can comfortably handle your deployment's normal level of activity.
- Maximum instances: set the maximum instances the deployment can scale up to, to limit your spend
- Scale on CPU threshold: enable and enter the value (% of CPU in use) to trigger scaling when your instances pass the threshold
- Scale on memory threshold: enable and enter the value (% of memory in use) to trigger scaling when your instances pass the threshold
- Scale on RPS (requests per second): enable and enter the value (number of requests per second) to trigger scaling when your instances pass the threshold
You can trigger autoscaling on either CPU or memory usage, RPS, or all. You can monitor the metrics for your containers in the Northflank application, or by using the Northflank API to decide upon the best thresholds to set for your services. If your usage spikes quite quickly, and/or your application takes some time to initialise, you may want to set lower thresholds so that new containers are available before the existing ones reach capacity.
Horizontal autoscaling behaviour
Your deployment will be checked every 15 seconds to calculate whether to trigger autoscaling. The average usage across all your instances is taken for each metric. If you are using more than one metric your deployment will be scaled to the highest number of required instances from any metric.
The required number of instances is calculated using this formula: requiredInstances = ceil[currentInstances * ( currentMetricValue / desiredMetricValue )]
.
When scaling up, autoscaling will not be triggered again until the new containers are in the running state, which prevents the creation of too many new instances.
Downscaling compares all the checks at 15-second intervals from a moving 5-minute window to ensure capacity is not reduced too quickly based on short-term drops in activity. For example, if a deployment is scaled to 5 instances and checks in the 5-minute window calculate that the deployment can be scaled down to either 1, 3, or 4 instances, then the deployment will be scaled down to 4 instances. If at any point a check returns a higher number of required instances, the deployment will immediately be scaled up.
Next steps
Monitor containers
Monitor the health and resource usage of deployments, and view detailed logs and metrics for individual container.
View metrics
View detailed, real-time metrics from builds, deployments, and more.
View logs
View detailed, real-time logs from builds, deployments, and more.
Configure log sinks
Integrate third-party logging and observability services with Northflank.