The CharmScaler is an autoscaler for Juju applications. Based on Elastisys'
autoscaling engine, it rightsizes your application deployments using
sophisticated auto-scaling algorithms to ensure that the application runs
cost-efficiently and is responsive at all times, even in the face of sudden
load spikes. At times of high anticipated load your charm is reinforced with
additional units -- units that are automatically decomissioned as the
pressure on your application goes down.
The Elastisys CharmScaler is an autoscaler for Juju applications. It
automatically scales your charm by adding units at times of high load and by
removing units at times of low load.
The initial edition of the CharmScaler features a simplified version of
Elastisys' autoscaling engine (described below),
and with limited scaling metric support. Work is underway on a more
fully-featured CharmScaler, but no release date has been set yet.
The initial CharmScaler edition scales the number of units of your applications
based on the observed CPU usage. These CPU metrics are collected from your
application by a telegraf agent, which
pushes the metrics into an
InfluxDB backend, from
where they are consumed by the CharmScaler.
The CharmScaler is available both free-of-charge and as a subscription service.
The free version comes with a size restriction which currently limits the size
of the scaled application to four units. Subscription users will see no such
size restrictions. For more details refer to the Subscription
If you are eager to try out the CharmScaler, head directly to the
Quickstart section. If you want to learn more about the
Elastisys autoscaler, read on ...
Introducing the Elastisys Autoscaler
User experience is king. You want to offer your users a smooth ride. From a
performance perspective, this translates into providing them with a responsive
service. As response times increase you will see more and more users leaving,
perhaps for competing services.
An application can be tuned in many ways, but one critical aspect is to make
sure that it runs on sufficient hardware, capable of bearing the weight that is
placed on your system. However, resource planning is notoriously hard and
involves a lot of guesswork. A fixed "peak-dimensioned" infrastructure is
certain to have you overspending most of the time and, what's worse, you can
never be sure that it actually will be able to handle the next load surge.
Ideally, you want to run with just the right amount of resources at all times.
It is plain to see that such a process involves a lot of planning and manual
Elastisys automates this process with a sophisticated autoscaler. The Elastisys
autoscaler uses proactive scaling algorithms based on state-of-the-art
research, which, predictively offers just in time capacity. That is, it can
provision servers in advance so that the right amount of capacity is available
when it is needed, not when you realize that it's needed (by then your
application may already be suffering). Research has shown that there is no
single scaling algorithm to rule them all. Different workload patterns require
different algorithms. The Elastisys autoscaler is armed with a growing
collection of such algorithms.
The Elastisys autoscaler already supports a
wide range of clouds and platforms.
With the addition of the Juju CharmScaler, which can scale any Juju application
Charm, integration with your application has never been easier. Whether it’s a
Wordpress site, a Hadoop cluster, a Kubernetes cluster, or even OpenStack
compute nodes, or your own custom-made application charm, hooking it up to be
scaled by the Elastisys autoscaler is really easy.
Read more about Elastisys' cloud automation platform at
The free edition places a constraint on the size of the scaled application to
four units. To remove this restriction you need to become a paying subscription
user. Juju is currently in beta, and does not yet support commercial charms.
Once Juju is officially released, the CharmScaler will be available as a
subscription service. Until then, you can contact us and we will help you set
up a temporary subscription arrangement.
For upgrading to a premium subscription, for a customized solution, or for
general questions or feature requests, feel free to contact Elastisys at
If you can't wait to get started, the following minimal example (relying on
configuration defaults) will let you start scaling your charm right away. For a
description of the CharmScaler and further details on its configuration, refer
to the sections below.
Minimal config.yaml example
charmscaler: juju_api_endpoint: "[API address]:17070" juju_model_uuid: "[uuid]" juju_username: "[username]" juju_password: "[password]"
Deploy and relate the charms
juju deploy cs:~elastisys/charmscaler --config=config.yaml juju deploy cs:~chris.macnaughton/influxdb juju deploy telegraf juju deploy [charm] juju relate charmscaler:db-api influxdb:api juju relate telegraf:influxdb-api influxdb:api juju relate telegraf:juju-info [charm]:juju-info juju relate charmscaler:juju-info [charm]:juju-info
How the CharmScaler operates
The image above illustrates the flow of the CharmScaler when scaling a
Wordpress application. Scaling decisions executed by the CharmScaler are
dependent on a load metric. In this case it looks at the CPU usage of machines
where Wordpress instances are deployed.
Metrics are collected by the Telegraf agent which is deployed as a subordinate
charm attached to the Wordpress application. This means that whenever the
Wordpress application is scaled out, another Telegraf collector will be
deployed as well and automatically start pushing new metrics to InfluxDB.
The CharmScaler will ask InfluxDB for new metric datapoints at every poll
interval (configured using the
metric_poll_interval option). From these load
metrics the CharmScaler decides how many units are needed by your application.
In the case of Wordpress it is necessary to distribute the load on all of the
units using a load balancer. If you haven't already, checkout the Juju
documentation page on
The CharmScaler's configuration is comprised of three main parts:
The CharmScaler manages the number of units of the scaled charm via the Juju
controller. To be able to do that it needs to authenticate with the controller.
Controller authentication credentials are passed to the CharmScaler through
options prefixed with
Note that in a foreseeable future, passing this kind of credentials to the
CharmScaler may no longer be necessary. Instead of requiring you to manually
type in the authentication details one could envision Juju giving the charm
access through relations or something similar.
The CharmScaler has a number of config options that control the autoscaler's
behavior. Those options are prefixed with either
metric_ options control the way metrics are fetched and processed while the
scaling_ options control when and how the charm units are scaled.
The scaling algorithm available in this edition of the CharmScaler is a
rule-based one that looks at CPU usage. At each iteration (configured using the
scaling_interval option) the following rules are considered by the autoscaler
before making a scaling decision:
scaling_cooldown- Has enough time passed since the last scale-event
(scale in or out) occured?
scaling_cpu_[max/min]- Is the CPU usage above/below the set limit?
scaling_period_[up/down]scale- Has the CPU usage been above/below
scaling_cpu_[max/min]for a long enough period of time?
If all three rules above are satisifed either a scale-out or a scale-in occurs
and the scaled charm will automatically add or remove a unit.
Note that configuring the scaling algorithm is a balancing act -- one always
needs to balance the need to scale "quickly enough" against the need to avoid
"jumpy behavior". Too frequent scale-ups/scale-downs could have a negative
impact on overall performance/system stability.
The default behavior adds a new unit when the average CPU usage (over all charm
units) has exceeded 80% for at least one minute. If you want to make the
CharmScaler quicker to respond to changes, you can, for example, lower the
threshold to 60% and the evaluation period to 30 seconds:
juju config charmscaler scaling_cpu_max=60 juju config charmscaler scaling_period_upscale=30
Similarly, the default behavior removes a new unit when the average CPU usage
has been under 20% (
scaling_cpu_min) for at least two minutes
scaling_period_downscale). Typically, it is preferable to allow the
application to be overprovisioned for some time to prevent situations where we
are too quick to scale down, only to realize that the load dip was only
temporary and that we need to scale back up again. We can, for instance, set
the evaluation period preceding scale-downs a bit longer (five minutes) via:
juju config charmscaler scaling_period_downscale=300
Finally, changing the amount of time required between two scaling decisions can
be done via:
juju config charmscaler scaling_cooldown=300
This parameter should, however, be kept long enough to give scaling decisions a
chance to take effect, before a new scaling decision is triggered.
Lastly, the options with the
alert_ prefix are used to enable CharmScaler
alerts (these are turned off by default).
Alerts are used to notify the outside world (such as the charm owner) of
noteable scaling events or error conditions. Alerts are, for example, sent
ERROR) if there are problems to reach the Juju
controller. Alerts are also sent (with severity-level
INFO) when a scaling
decision has been made.
This edition of the CharmScaler includes email alerts which are configured by
entering the SMTP server details which the autoscaler is supposed to send the
alert email messages to.
When deploying on LXD provider
Due to missing support for the Docker LXC profile in Juju you need to apply it
The Docker layer currently
does not support installing Docker Compose in a virtual environment. Until
this is supported
InfluxDB cannot be co-located with the CharmScaler charm because of dependency
CPU threshold where the load is considered low enough to scale down the number of units.
Number of seconds that the CPU usage needs to be lower than the threshold before scaling down.
Maximum amount of units to keep in pool
Toggle installation from ubuntu archive vs the docker PPA
URL to use for HTTP_PROXY to be used by Docker. Only useful in closed environments where a proxy is the only option for routing to the registry to pull images
Juju model UUID
Extra options to pass to the docker daemon. e.g. --insecure-registry
Seconds between each scaling decision
How often the charmscaler should sync against the Juju model.
Juju account password
The minimum age (in seconds) of requested data points. When requesting recent aggregate metric data points, there is always a risk of seeing partial/incomplete results before metric values from all sources have been registered. The value to set for this field depends on the reporting frequency of monitoring agents, but as a general rule-of-thumb, this value can be set to be about 1.5 times the length of the reporting-interval for monitoring agents.
Juju controller API endpoint
Comma-separated list of destinations (either domain names or IP addresses) that should be directly accessed, by opposition of going through the proxy defined above.
URL to use for HTTPS_PROXY to be used by Docker. Only useful in closed environments where a proxy is the only option for routing to the registry to pull images
Toggle e-mail alerts on/off
Username to auth with the SMTP server
Enable GRUB cgroup overrides cgroup_enable=memory swapaccount=1. WARNING changing this option will reboot the host - use with caution on production services
E-mail address that alert mails should be sent from
A comma-separated list of nagios servicegroups. If left empty, the nagios_context will be used as the servicegroup
Seconds between polls for new metric values
URL to the Charmpool component. By default both the autoscaler and the pool is run in the same Docker network and will reach eachother by their local hostnames.
Use SSL when connecting to SMTP host
Juju account username
Time (in seconds) before making another scaling decision from the time of the last up- or downscale. This is useful to prevent extra resizes due to slow teardowns or, in perticular, upstarts.
Port which the Autoscaler API should be served on.
The name of the service - mainly shows up in the alert e-mails Also useful to distinguish between multiple CharmScaler charms
Used by the nrpe subordinate charms. A string that will be prepended to instance name to set the host name in nagios. So for instance the hostname would be something like: juju-myservice-0 If you're running multiple environments with the same services in them this allows you to differentiate between them.
Password to auth with the SMTP server
CPU usage threshold at which the number of units should be scaled up.
Space separated list of e-mail addresses that should recieve alerts
Minimum amount of units to keep in pool
Alert levels that should trigger alert mails to be sent out
- INFO NOTICE WARN ERROR FATAL
Number of seconds that the CPU usage needs to be higher than the threshold before scaling up.