• By OpenStack Charmers
  • Cloud
Channel Version Revision Published Runs on
latest/stable 530 530 09 Feb 2022
Ubuntu 21.10 Ubuntu 21.04 Ubuntu 20.10 Ubuntu 20.04 Ubuntu 18.04 Ubuntu 16.04
latest/edge 564 564 05 Sep 2022
Ubuntu 22.10 Ubuntu 22.04
yoga/stable 569 569 12 Sep 2022
Ubuntu 22.04 Ubuntu 20.04
xena/stable 560 560 05 Aug 2022
Ubuntu 20.04
wallaby/edge 565 565 06 Sep 2022
Ubuntu 20.04
victoria/edge 567 567 09 Sep 2022
Ubuntu 20.04
ussuri/edge 568 568 09 Sep 2022
Ubuntu 20.04 Ubuntu 18.04
train/edge 82bc1e2 544 25 Feb 2022
Ubuntu 20.04
stein/edge 82bc1e2 544 25 Feb 2022
Ubuntu 20.04
rocky/edge 82bc1e2 544 25 Feb 2022
Ubuntu 20.04
queens/edge 82bc1e2 545 25 Feb 2022
Ubuntu 20.04
juju deploy cinder
Show information


21.10 21.04 20.10 20.04 18.04 16.04


The cinder charm deploys Cinder, the Block Storage (volume) service for OpenStack. The charm works alongside other Juju-deployed OpenStack services.



This section covers common and/or important configuration options. See file config.yaml for the full list of options, along with their descriptions and default values. See the Juju documentation for details on configuring applications.


Specifies the storage source. Setting this option to 'None' will allow for storage to be managed by separate charms. See sections Ceph-backed storage, LVM-backed storage, and NetApp-backed storage.

Important: The practice of setting the block-device option to a local

block device is deprecated. Doing so enacts the charm's built-in support for LVM storage. This feature will soon be removed from the charm, along with the option's default value of 'sdb'.


States the software sources. A common value is an OpenStack UCA release (e.g. 'cloud:bionic-ussuri' or 'cloud:focal-wallaby'). See Ubuntu Cloud Archive. The underlying host's existing apt sources will be used if this option is not specified (this behaviour can be explicitly chosen by using the value of 'distro').


This section includes two different deployment scenarios, each of which requires these applications to be present: keystone, nova-cloud-controller, nova-compute, rabbitmq-server, and a cloud database.

The database application is determined by the series. Prior to focal percona-cluster is used, otherwise it is mysql-innodb-cluster. In the example deployment below mysql-innodb-cluster has been chosen.

Ceph-backed storage

Cinder can be backed by Ceph, which provides volumes with scalability and redundancy.

Note: Ceph is the recommended storage method for production Cinder


These instructions assume a pre-existing Ceph cluster.

File cinder.yaml contains the following:

        block-device: None

Option block-device must be set to 'None' to disable the local block device.

Here, Cinder is deployed to a container on machine '1' and related to the Ceph cluster via the cinder-ceph subordinate charm:

juju deploy --to lxd:1 --config cinder.yaml cinder
juju deploy cinder-ceph
juju add-relation cinder-ceph:storage-backend cinder:storage-backend
juju add-relation cinder-ceph:ceph ceph-mon:client
juju add-relation cinder-ceph:ceph-access nova-compute:ceph-access

Proceed with a group of commands common to both scenarios:

juju add-relation cinder:identity-service keystone:identity-service
juju add-relation cinder:cinder-volume-service nova-cloud-controller:cinder-volume-service
juju add-relation cinder:amqp rabbitmq-server:amqp

juju deploy mysql-router cinder-mysql-router
juju add-relation cinder-mysql-router:db-router mysql-innodb-cluster:db-router
juju add-relation cinder-mysql-router:shared-db cinder:shared-db

LVM-backed storage

Cinder can be backed by storage local to the cinder unit, where local block devices are used as LVM physical volumes, and volumes are offered via iSCSI. This functionality is provided by the cinder-lvm subordinate charm.

Note: Built-in support for LVM in the cinder charm is deprecated.

NetApp-backed storage

Cinder can be backed by a NetApp appliance local to the cinder unit, where volumes are offered via iSCSI or NFS. This functionality is provided by the cinder-netapp subordinate charm.

High availability

When more than one unit is deployed with the hacluster application the charm will bring up an HA active/active cluster.

There are two mutually exclusive high availability options: using virtual IP(s) or DNS. In both cases the hacluster subordinate charm is used to provide the Corosync and Pacemaker backend HA functionality.

See OpenStack high availability in the OpenStack Charms Deployment Guide for details.

Network spaces

This charm supports the use of Juju network spaces (Juju v.2.0). This feature optionally allows specific types of the application's network traffic to be bound to subnets that the underlying hardware is connected to.

Note: Spaces must be configured in the backing cloud prior to deployment.

API endpoints can be bound to distinct network spaces supporting the network separation of public, internal, and admin endpoints.

Access to the underlying MySQL instance can also be bound to a specific space using the shared-db relation.

For example, providing that spaces 'public-space', 'internal-space', and 'admin-space' exist, the deploy command above could look like this:

juju deploy --config cinder.yaml cinder \
   --bind "public=public-space internal=internal-space admin=admin-space shared-db=internal-space"

Alternatively, configuration can be provided as part of a bundle:

      charm: cs:cinder
      num_units: 1
        public: public-space
        internal: internal-space
        admin: admin-space
        shared-db: internal-space

Note: Existing cinder units configured with the os-admin-network,

os-internal-network, or os-public-network options will continue to honour them. Furthermore, these options override any space bindings, if set.


This section covers Juju actions supported by the charm. Actions allow specific operations to be performed on a per-unit basis. To display action descriptions run juju actions --schema cinder. If the charm is not deployed then see file actions.yaml.

  • openstack-upgrade
  • pause
  • remove-services
  • rename-volume-host
  • resume
  • security-checklist
  • volume-host-add-driver

Policy overrides

Policy overrides is an advanced feature that allows an operator to override the default policy of an OpenStack service. The policies that the service supports, the defaults it implements in its code, and the defaults that a charm may include should all be clearly understood before proceeding.

Caution: It is possible to break the system (for tenants and other

services) if policies are incorrectly applied to the service.

Policy statements are placed in a YAML file. This file (or files) is then (ZIP) compressed into a single file and used as an application resource. The override is then enabled via a Boolean charm option.

Here are the essential commands (filenames are arbitrary):

zip override-file.yaml
juju attach-resource cinder
juju config cinder use-policyd-override=true

See Policy overrides in the OpenStack Charms Deployment Guide for a thorough treatment of this feature.


The OpenStack Charms project maintains two documentation guides:


Please report bugs on Launchpad.