Juju Charm - Swift proxy
Go to file
Alex Kavanagh 2470580d5f Update charm icon
Change-Id: Ifb44da6f1e68d329a140df9491807530dc79b118
Closes-Bug: #1686967
2017-08-02 17:28:33 +01:00
actions Implement swauth 2016-09-23 09:12:43 -04:00
charmhelpers Update tests to use keystoneauth1 2017-07-24 15:49:53 -07:00
hooks Network space aware address for cluster relation 2017-05-05 10:44:30 -07:00
lib Fix alphanumeric comparisons for openstack and ubuntu releases 2017-04-27 12:13:59 +01:00
scripts - removed health checks 2013-04-05 17:34:01 -04:00
templates Fix Keystone v3 auth for swift-proxy 2016-12-08 07:17:26 +01:00
tests Update tests to use keystoneauth1 2017-07-24 15:49:53 -07:00
unit_tests Network space aware address for cluster relation 2017-05-05 10:44:30 -07:00
.coveragerc Move swift helpers out of hooks, into lib for re-use by more than just hooks 2015-09-07 14:44:04 +01:00
.gitignore Resync charm-helpers 2016-03-02 12:11:23 +00:00
.gitreview Add gitreview prior to migration to openstack 2016-02-24 21:53:36 +00:00
.project Fixup missing files 2013-02-27 16:18:34 +00:00
.pydevproject Fixup missing files 2013-02-27 16:18:34 +00:00
.testr.conf Add tox support, tidy unit tests 2015-11-03 13:52:37 +00:00
LICENSE Re-license charm as Apache-2.0 2016-07-01 18:12:11 +01:00
Makefile Use bundletester for amulet test execution 2016-07-19 11:10:36 +01:00
README.Swift_ring_management Refactor of ring management code 2016-03-30 10:40:43 +01:00
README.md Replace obsolete vanity openstack.org URLs 2017-03-08 19:32:58 +01:00
actions.yaml Implement swauth 2016-09-23 09:12:43 -04:00
charm-helpers-hooks.yaml Add support for application version 2016-09-20 12:45:17 +01:00
charm-helpers-tests.yaml Fix alphanumeric comparisons for openstack and ubuntu releases 2017-04-27 12:13:59 +01:00
config.yaml Cleanup config.yaml 2017-07-05 17:57:40 +01:00
copyright Re-license charm as Apache-2.0 2016-07-01 18:12:11 +01:00
hardening.yaml Add hardening support 2016-03-24 11:36:26 +00:00
icon.svg Update charm icon 2017-08-02 17:28:33 +01:00
metadata.yaml Enable Ocata Amulet Tests 2017-03-09 14:35:25 -08:00
requirements.txt Fix pbr requirement 2016-04-13 10:24:38 +00:00
revision Add config options for proxy-server.conf template. node-timeout and recoverable-node-timeout. For large swift/glance uploads, higher defaults are needed to prevent uploads from failing on real servers which are slower than VMs. 2014-06-03 20:38:25 -05:00
setup.cfg Move swift helpers out of hooks, into lib for re-use by more than just hooks 2015-09-07 14:44:04 +01:00
test-requirements.txt Enable Zesty-Ocata Amulet Tests 2017-04-28 07:50:58 -07:00
tox.ini Enable Zesty-Ocata Amulet Tests 2017-04-28 07:50:58 -07:00

README.md

Overview

This charm provides the swift-proxy component of the OpenStack Swift object storage system. It can be deployed as part of its own stand-alone storage cluster or it can be integrated with the other OpenStack components, assuming those are also managed by Juju. For Swift to function, you'll also need to deploy additional swift-storage nodes using the cs:precise/swift-storage charm.

For more information about Swift and its architecture, visit the official project website

This charm was developed to support deploying multiple version of Swift on Ubuntu Precise 12.04, as they relate to the release series of OpenStack. That is, OpenStack Essex corresponds to Swift 1.4.8 while OpenStack Folsom shipped 1.7.4. This charm can be used to deploy either (and future) versions of Swift onto an Ubuntu Precise 12.04 or Trusty 14.04 making use of the Ubuntu Cloud Archive when needed.

Usage

Currently, Swift may be deployed in two ways. In either case, additional storage nodes are required. The configuration option that dictates how to deploy this charm is the 'zone-assignment' setting. This section describes how to select the appropriate zone assignment policy, as well as a few other configuration settings of interest. Many of the configuration settings can be left as default.

Zone Assignment

This setting determines how the charm assigns new storage nodes to storage zones.

The default, 'manual' option is suggested for production as it allows administrators to carefully architect the storage cluster. It requires each swift-storage service to be deployed with an explicit storage zone configured in its deployment settings. Upon relation to a swift-proxy, the storage node will request membership to its configured zone and be assigned by the swift-proxy charm accordingly. Using the cs:precise/swift-storage charm with this charm, a deployment would look something like:

$ cat >swift.cfg <<END
    swift-proxy:
        zone-assignment: manual
        replicas: 3
    swift-storage-zone1:
        zone: 1
        block-device: /etc/swift/storage.img|2G
    swift-storage-zone2:
        zone: 2
        block-device: /etc/swift/storage.img|2G
    swift-storage-zone3:
        zone: 3
        block-device: /etc/swift/storage.img|2G
END
$ juju deploy --config=swift.cfg swift-proxy
$ juju deploy --config=swift.cfg swift-storage swift-storage-zone1
$ juju deploy --config=swift.cfg swift-storage swift-storage-zone2
$ juju deploy --config=swift.cfg swift-storage swift-storage-zone3
$ juju add-relation swift-proxy swift-storage-zone1
$ juju add-relation swift-proxy swift-storage-zone2
$ juju add-relation swift-proxy swift-storage-zone3

This will result in a configured storage cluster of 3 zones, each with one node. To expand capacity of the storage system, nodes can be added to specific zones in the ring.

$ juju add-unit swift-storage-zone1
$ juju add-unit -n5 swift-storage-zone3    # Adds 5 units to zone3

This charm will not balance the storage ring until there are enough storage zones to meet its minimum replica requirement, in this case 3.

The other option for zone assignment is 'auto'. In this mode, swift-proxy gets a relation to a single swift-storage service unit. Each machine unit assigned to that service unit will be distributed evenly across zones.

$ cat >swift.cfg <<END
swift-proxy:
    zone-assignment: auto
    replicas: 3
swift-storage:
    zone: 1
    block-device: /etc/swift/storage.img|2G
END
$ juju deploy --config=swift.cfg swift-proxy
$ juju deploy --config=swift.cfg swift-storage
$ juju add-relation swift-proxy swift-storage
# The swift-storage/0 unit ends up the first node in zone 1
$ juju add-unit swift-storage
# swift-storage/1 ends up the first node in zone 2.
$ juju add-unit swift-storage
# swift-storage/2 is the first in zone 3, replica requirement is satisfied
# the ring is balanced.

Extending the ring in the case is just a matter of adding more units to the single service unit. New units will be distributed across the existing zones.

$ juju add-unit swift-storage
# swift-storage/3 is assigned to zone 1.
$ juju add-unit swift-storage
# swift-storage/4 is assigned to zone 2.
etc.

Installation repository.

The 'openstack-origin' setting allows Swift to be installed from installation repositories and can be used to setup access to the Ubuntu Cloud Archive to support installing Swift versions more recent than what is shipped with Ubuntu 12.04 (1.4.8). For more information, see config.yaml.

Authentication.

By default, the charm will be deployed using the tempauth auth system. This is a simple and not-recommended auth system that functions without any external dependencies. See Swift documentation for details.

The charm may also be configured to use Keystone, either manually (via config) or automatically via a relation to an existing Keystone service using the cs:precise/keystone charm. The latter is preferred, however, if a Keystone service is desired but it is not managed by Juju, the configuration for the auth token middleware can be set manually via the charm's config. A relation to a Keystone server via the identity-service interface will configure swift-proxy with the appropriate credentials to make use of Keystone and is required for any integration with other OpenStack components.

Glance

Swift may be used to as a storage backend for the Glance image service. To do so, simply add a relation between swift-proxy and an existing Glance service deployed using the cs:precise/glance charm.

HA/Clustering

There are two mutually exclusive high availability options: using virtual IP(s) or DNS. In both cases, a relationship to hacluster is required which provides the corosync back end HA functionality.

To use virtual IP(s) the clustered nodes must be on the same subnet such that the VIP is a valid IP on the subnet for one of the node's interfaces and each node has an interface in said subnet. The VIP becomes a highly-available API endpoint.

At a minimum, the config option 'vip' must be set in order to use virtual IP HA. If multiple networks are being used, a VIP should be provided for each network, separated by spaces. Optionally, vip_iface or vip_cidr may be specified.

To use DNS high availability there are several prerequisites. However, DNS HA does not require the clustered nodes to be on the same subnet. Currently the DNS HA feature is only available for MAAS 2.0 or greater environments. MAAS 2.0 requires Juju 2.0 or greater. The clustered nodes must have static or "reserved" IP addresses registered in MAAS. The DNS hostname(s) must be pre-registered in MAAS before use with DNS HA.

At a minimum, the config option 'dns-ha' must be set to true and at least one of 'os-public-hostname', 'os-internal-hostname' or 'os-internal-hostname' must be set in order to use DNS HA. One or more of the above hostnames may be set.

The charm will throw an exception in the following circumstances: If neither 'vip' nor 'dns-ha' is set and the charm is related to hacluster If both 'vip' and 'dns-ha' are set as they are mutually exclusive If 'dns-ha' is set and none of the os-{admin,internal,public}-hostname(s) are set

Network Space support

This charm supports the use of Juju Network Spaces, allowing the charm to be bound to network space configurations managed directly by Juju. This is only supported with Juju 2.0 and above.

API endpoints can be bound to distinct network spaces supporting the network separation of public, internal and admin endpoints.

To use this feature, use the --bind option when deploying the charm:

juju deploy swift-proxy --bind "public=public-space internal=internal-space admin=admin-space"

alternatively these can also be provided as part of a juju native bundle configuration:

swift-proxy:
  charm: cs:xenial/swift-proxy
  num_units: 1
  bindings:
    public: public-space
    admin: admin-space
    internal: internal-space

NOTE: Spaces must be configured in the underlying provider prior to attempting to use them.

NOTE: Existing deployments using os-*-network configuration options will continue to function; these options are preferred over any network space binding provided if set.