Tools and automation to achieve Disaster Recovery of OpenStack Cloud Platforms
Go to file
xuanyandong 1f3892bb08 Switch to Ussuri jobs
Change-Id: I8ee14a5b0e50aaad9333d0bea85a924a230839e4
2019-10-25 15:52:00 +08:00
config-generator freezer-dr big bang 2016-05-09 09:55:31 +00:00
doc Update .gitignore 2018-11-12 06:20:22 -05:00
etc Set default notifier endpoint 2018-10-15 20:28:09 +08:00
freezer_dr Merge "Set default notifier endpoint" 2018-11-19 09:22:28 +00:00
tests Adding pep8, pylint, coverage, sphinx testing 2016-05-09 15:00:10 +00:00
.coveragerc Adding pep8, pylint, coverage, sphinx testing 2016-05-09 15:00:10 +00:00
.gitignore Update .gitignore 2018-11-12 06:20:22 -05:00
.gitreview OpenDev Migration Patch 2019-04-19 19:30:00 +00:00
.pylintrc update pylint 2018-08-15 13:22:31 +09:00
.stestr.conf Switch to stestr 2018-07-11 10:59:33 +07:00
.zuul.yaml Switch to Ussuri jobs 2019-10-25 15:52:00 +08:00
CREDITS.rst Adding CREDITS.rst 2016-02-12 16:28:58 +00:00
HACKING.rst Sync Sphinx requirement 2019-08-03 11:53:24 +02:00
LICENSE Add LICENSE file 2017-01-17 13:18:06 +07:00
README.rst Sync Sphinx requirement 2019-08-03 11:53:24 +02:00
bindep.txt Add bindep to fix py37 tests 2019-08-03 11:38:14 +02:00
lower-constraints.txt Update tox.ini and fix pep8 errors 2018-11-07 06:14:17 -05:00
requirements.txt Update tox.ini and fix pep8 errors 2018-11-07 06:14:17 -05:00
setup.cfg Drop the py35 testing, add the py37 testing 2019-05-24 21:53:50 -04:00
setup.py Update pbr version 2018-11-07 01:37:20 -05:00
test-requirements.txt Update hacking 2019-08-03 11:55:12 +02:00
tox.ini Add Python 3 Train unit tests 2019-06-24 09:25:17 -04:00

README.rst

Team and repository tags

image

Freezer Disaster Recovery

freezer-dr, OpenStack Compute node High Available provides compute node high availability for OpenStack. Simply freezer-dr monitors all compute nodes running in a cloud deployment and if there is any failure in one of the compute nodes freezer-dr will fence this compute node then freezer-dr will try to evacuate all running instances on this compute node, finally freezer-dr will notify all users who have workload/instances running on this compute node as well as will notify the cloud administrators.

freezer-dr has a pluggable architecture so it can be used with:

  1. Any monitoring system to monitor the compute nodes (currently we support only native OpenStack services status)
  2. Any fencing driver (currently supports IPMI, libvirt, ...)
  3. Any evacuation driver (currently supports evacuate api call, may be migrate ??)
  4. Any notification system (currently supports email based notifications, ...)

just by adding a simple plugin and adjust the configuration file to use this plugin or in future a combination of plugins if required

freezer-dr should run in the control plane, however the architecture supports different scenarios. For running freezer-dr under high availability mode, it should run with active passive mode.

How it works

Starting freezer-dr:

  1. freezer-dr Monitoring manager is going to load the required monitoring driver according to the configuration
  2. freezer-dr will query the monitoring system to check if it considers any compute nodes to be down ?
  3. if no, freezer-dr will exit displaying No failed nodes
  4. if yes, freezer-dr will call the fencing manager to fence the failed compute node
  5. Fencing manager will load the correct fencer according to the configuration
  6. once the compute node is fenced and is powered off now we will start the evacuation process
  7. freezer-dr will load the correct evacuation driver
  8. freezer-dr will evacuate all instances to another computes
  9. Once the evacuation process completed, freezer-dr will call the notification manager
  10. The notification manager will load the correct driver based on the configurations
  11. freezer-dr will start the notification process ...