This patchset modifies the add-disk action so that it now
can optionally start a Crimson OSD daemon.
Change-Id: I59bf4e41f1f56c6bda2352b5613289ff73113342
Depends-On: If58bde4d5445ed5de420abc007db6bf8b8e43269
This new action allows users to either purge an OSD, or remove it,
opening up the possibility of recycling the previous OSD id. In
addition, this action will clean up any bcache devices that were
created in previous steps.
Change-Id: If3566031ba3f02dac0bc86938dcf9e85a66a66f0
Depends-On: Ib959e81833eb2094d02c7bdd507b1c8b7fbcd3db
func-test-pr: https://github.com/openstack-charmers/zaza-openstack-tests/pull/683
As part of the task to improve disk replacement, the 'add-disk' action
needs some changes. This includes:
- Creating 'bcache' devices to accelerate disk access.
- Creating caching partitions of a specified size.
- Recycling previously deactivated OSD ids.
Change-Id: Id5027f30d51c23d2be4c34f82867d65a50b35137
Depends-On: I43d0a0bc11664c37532c0117711affc93c9d1ad1
func-test-pr: https://github.com/openstack-charmers/zaza-openstack-tests/pull/675
The 'get-availability-zone' action will get information about an
availability zone that will contain information about the CRUSH
structure. Specifically 'rack' and 'row'.
Closes-Bug: #1911006
Change-Id: I99ebbef5f23d6efe3c848b089c7f2b0d26ad0077
The `osds` parameter has been copied from function start/stop to preserve
the same functionality. By default, osd-in/osd-out needs list of IDs as
an argument or it will not do anything (previously, it applied the change
to all the osds). It's possible to take in/out *all* with provided `osds`
parameter as `all`.
Closes-Bug: #1910150
Change-Id: I0275f015e2d0bbbb661d2b7dea59c320ba6c021c
Review of README.
Corrected doc URLs in actions.yaml.
The trailing spaces on these lines are deliberate
(forces a carriage return):
260
265
291
307
332
337
Change-Id: Ia61edbfcbf27bf9bc6b35a71793df39c7cb46907
The actions pause and resume actually take all osds on a unit out of the
cluster. This is incredibly misleading.
This change renames to osd-out and osd-in to better describe what the
actions actually do.
Change-Id: I76793999f5d3382563eff308a5d7c4db18d065a0
Closes-Bug: #1793507
Output of `juju list-action` is at time of this writing
formatted in such a way that we should keep description
as terse as possible and refer to documentation elsewhere.
Change-Id: Ib8e7a4804e696199803b9ac386da7bf02aafd465
Add `non-pristine` key to `list-disks` action.
No longer attempt to do initializtion of `osd-journal` devices.
Make py27 test noop
Flip pep8 test to py3
Partial-Bug: #1698154
Change-Id: I0ca574fa7f0683b4e8a693b9f62fbf6b39689789
Depends-On: I90a866aa138d18e4242783c42d4c7c587f696d7d
This action includes configuration for disk(s) to
zap, as well as an additional required flag for
the administrator to acknowledge pending data loss
Change-Id: I3106e2f10cf132a628aad025f73161b04215598e
Related-Bug: #1698154
At present this action does not work. While looking to repair the
functionality I found a number of issues with the current
implementation.
For now I suggest we remove this functionality, and at some point we may
consider replacing it with a `remove-disk` action.
Sync in relevant changes from charms.ceph
Depends-On: Id61b87927c43d807aacc93cf05ec8f88d91b7a39
Change-Id: Ic71d304ff65a05ab7249f4dd07adc45429a323e9
The blacklist actions allow for adding and removing devices
to a unit-local list of devices to be skipped during osd
initialization. This list will be used to override the
application level, and thereby deployment wide, 'osd-devices'
configuration option on a individual unit basis.
The pre-existing list-disk action is extended to return
list of blacklisted devices under the 'blacklist' key.
Change-Id: I28a3c5d6076fb496dead3fe3387d9bbbbe9ec083
Closes-Bug: #1730267
This action is fairly simple in that it returns
a list of unmounted disks
This also includes a git-sync to pull in charms.ceph
changes.
Change-Id: I0daa514958799cf5899375335e8e9e684df27704
Closes-Bug: 1645481
This changeset provides pause and resume actions to the ceph charm.
The pause action issues a 'ceph osd out <local_id>' for each of the
ceph osd ids that are on the unit. The action does not stop the
ceph osd processes.
Note that if the pause-health action is NOT used on the ceph-mon
charm then the cluster will start trying to rebalance the PGs accross
the remaining OSDs. If the cluster might reach its 'full ratio' then
this will be a breaking action. The charm does NOT check for this
eventuality.
The resume action issues a 'ceph osd in <local_id>' for each of the
local ceph osd process on the unit.
The charm 'remembers' that a pause action was issued, and if
successful, it shows a 'maintenance' workload status as a reminder.
Change-Id: I9f53c9c6c4bb737670ffcd542acec0b320cc7f6a
This patch adds an action to replace a hard drive for an particular
osd server. The user executing the action will give the OSD number
and also the device name of the replacement drive. The rest is
taken care of by the action. The action will attempt to go through
all the osd removal steps for the failed drive. It will force
unmount the drive and if that fails it will lazy unmount the drive.
This force and then lazy pattern comes from experience with dead
hard drives not behaving nicely with umount.
Change-Id: I914cd484280ac3f9b9f1fad8b35ee53e92438a0a