--- # Copyright 2014, Rackspace US, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. {% if ansible_facts['os_family'] | lower == 'redhat' %} deployment_environment_variables: LIBSYSTEMD_VERSION: {{ systemd_version.stdout_lines[0].split('-')[1] }} {% endif %} ## General options debug: True ## Installation method for OpenStack services install_method: "{{ bootstrap_host_install_method }}" ## Tempest settings tempest_public_subnet_cidr: "172.29.248.0/22" tempest_public_subnet_allocation_pools: "172.29.249.110-172.29.249.200" {% if _neutron_plugin_driver == 'ml2.ovn' %} tempest_private_net_provider_type: geneve {% else %} tempest_private_net_provider_type: vxlan {% endif %} {% if _neutron_plugin_driver == 'ml2.lxb' %} tempest_public_net_physical_name: flat {% else %} tempest_public_net_physical_name: vlan {% endif %} # systemd-journald limitations openstack_hosts_journald_config: RateLimitIntervalSec: 60 RateLimitBurst: 600 SystemMaxUse: 4G RuntimeMaxUse: 2G SystemMaxFileSize: 100M RuntimeMaxFileSize: 100M ## Galera settings galera_monitoring_allowed_source: "0.0.0.0/0" # TODO(noonedeadpunk): This should be enabled, once we will re-work SSL part #galera_use_ssl: "{{ ('infra' in bootstrap_host_scenarios_expanded) }}" galera_innodb_log_buffer_size: 4M galera_wsrep_provider_options: - { option: "gcache.size", value: "4M" } galera_my_cnf_overrides: mysqld: read_buffer_size: '64K' innodb_buffer_pool_size: '16M' thread_stack: '192K' thread_cache_size: '8' tmp_table_size: '8M' sort_buffer_size: '8M' max_allowed_packet: '8M' ### Set workers for all services to optimise memory usage ## Repo repo_nginx_threads: 2 ## Keystone keystone_httpd_mpm_start_servers: 2 keystone_httpd_mpm_min_spare_threads: 5 keystone_httpd_mpm_max_spare_threads: 10 keystone_httpd_mpm_thread_limit: 15 keystone_httpd_mpm_thread_child: 15 keystone_wsgi_threads: 2 keystone_wsgi_processes: 2 ## Barbican barbican_wsgi_processes: 1 barbican_wsgi_threads: 1 ## Blazar blazar_wsgi_processes: 1 blazar_wsgi_threads: 1 ## Cinder cinder_wsgi_processes: 1 cinder_wsgi_threads: 1 cinder_wsgi_buffer_size: 16384 cinder_osapi_volume_workers_max: 1 ## cloudkitty cloudkitty_wsgi_processes: 1 cloudkitty_wsgi_threads: 1 ## Glance glance_api_threads: 1 glance_api_workers: 1 glance_wsgi_threads: 1 glance_wsgi_processes: 1 ## Placement placement_wsgi_threads: 1 placement_wsgi_processes: 1 placement_wsgi_buffer_size: 16384 ## Manila manila_wsgi_processes: 1 manila_wsgi_threads: 1 manila_osapi_share_workers: 2 manila_wsgi_buffer_size: 65535 ## mistral mistral_wsgi_processes: 1 mistral_wsgi_threads: 1 ## Nova nova_reserved_host_memory_mb: 256 nova_wsgi_threads: 1 nova_wsgi_processes: 1 nova_wsgi_buffer_size: 16384 nova_api_threads: 1 nova_osapi_compute_workers: 1 nova_conductor_workers: 1 nova_metadata_workers: 1 nova_scheduler_workers: 1 ## Neutron neutron_rpc_workers: 1 neutron_metadata_workers: 1 neutron_api_workers: 1 neutron_api_threads: 2 neutron_num_sync_threads: 1 neutron_wsgi_processes: 1 neutron_plugin_type: "{{ _neutron_plugin_driver }}" {% if _neutron_plugin_driver != 'ml2.ovn' %} neutron_ml2_drivers_type: "flat,vlan,vxlan" neutron_plugin_base: - router - metering {% endif %} {% if 'neutron' in bootstrap_host_scenarios %} tempest_test_includelist_neutron: - "neutron_tempest_plugin.api.test_networks*" - "tempest.scenario.test_network_basic_ops.TestNetworkBasicOps.test_network_basic_ops" tempest_tempest_conf_overrides_neutron: network-feature-enabled: api_extensions: agent,allowed-address-pairs,binding,dhcp_agent_scheduler,ext-gw-mode,external-net,extra_dhcp_opt,extra_dhcp_optagent,extraroute,l3_agent_scheduler,metering,provider,quotas,router,security-group,service-type,subnet_allocation {% endif %} ## Octavia octavia_wsgi_threads: 1 octavia_wsgi_processes: 1 octavia_wsgi_buffer_size: 16384 octavia_management_net_subnet_cidr: 172.29.232.0/22 octavia_management_net_subnet_allocation_pools: "172.29.232.50-172.29.235.254" ## Heat heat_api_workers: 1 heat_api_threads: 1 heat_wsgi_threads: 1 heat_wsgi_processes: 1 heat_wsgi_buffer_size: 16384 ## Horizon horizon_wsgi_processes: 1 horizon_wsgi_threads: 1 ## Ceilometer ceilometer_notification_workers: 1 ## AODH aodh_wsgi_threads: 1 aodh_wsgi_processes: 1 ## Gnocchi gnocchi_wsgi_threads: 1 gnocchi_wsgi_processes: 1 gnocchi_metricd_workers: 1 ## Swift swift_account_server_replicator_workers: 1 swift_server_replicator_workers: 1 swift_object_replicator_workers: 1 swift_account_server_workers: 1 swift_container_server_workers: 1 swift_object_server_workers: 1 swift_proxy_server_workers_not_capped: 1 swift_proxy_server_workers_capped: 1 swift_proxy_server_workers: 1 ## Ironic ironic_wsgi_threads: 1 ironic_wsgi_processes: 1 ## Ironic Inspector ironic_inspector_wsgi_threads: 1 ironic_inspector_wsgi_processes: 1 ## Trove trove_service_net_setup: true trove_api_workers: 1 trove_conductor_workers: 1 trove_wsgi_threads: 1 trove_wsgi_processes: 1 ## Magnum magnum_wsgi_processes: 1 magnum_conductor_workers: 1 {% if 'metal' in bootstrap_host_scenarios %} venv_wheel_build_enable: false {% endif %} ## Sahara sahara_api_workers: 1 sahara_wsgi_threads: 1 sahara_wsgi_processes: 1 sahara_wsgi_buffer_size: 16384 ## Zun zun_api_threads: 1 zun_wsgi_threads: 1 zun_wsgi_processes: 1 ## Senlin senlin_api_threads: 1 senlin_wsgi_threads: 1 senlin_wsgi_processes: 1 # NOTE: hpcloud-b4's eth0 uses 10.0.3.0/24, which overlaps with the # lxc_net_address default # TODO: We'll need to implement a mechanism to determine valid lxc_net_address # value which will not overlap with an IP already assigned to the host. lxc_net_address: 10.255.255.1 lxc_net_netmask: 255.255.255.0 lxc_net_dhcp_range: 10.255.255.2,10.255.255.253 lxc_net_mtu: {{ hostvars[inventory_hostname]['ansible_' ~ bootstrap_host_public_interface]['mtu'] | default(1500) }} {% if cache_timeout is defined %} ## Package cache timeout cache_timeout: {{ cache_timeout }} {% endif %} lxc_container_backing_store: {{ _lxc_container_backing_store }} {% if _lxc_container_backing_store == 'zfs' %} lxc_container_zfs_root_name: "osa-test-pool/lxc" {% endif %} # bind mount the zuul repos into the containers lxc_container_bind_mounts: - host_directory: "/home/zuul/src" container_directory: "/openstack/src" - host_directory: "/opt/cache/files" container_directory: "/opt/cache/files" ## Always setup tempest, the resources for it, then execute tests tempest_install: yes tempest_run: yes rally_install: yes # Do a gateway ping test once the tempest role creates it tempest_network_ping_gateway: yes {% if nodepool_dir.stat.exists %} # Copy /etc/pip.conf into containers to get mirrors for wheels # and due to extra-index-url bugs in Ubuntu, we workaround it # by ignoring the config file during PIP upgrade time venv_pip_upgrade_noconf: true lxc_container_cache_files_from_host: - /etc/pip.conf # Disable chronyd in OpenStack CI security_rhel7_enable_chrony: no # The location where images are downloaded in openstack-infra tempest_image_dir: "/opt/cache/files" {% endif %} # Set all the distros to the same value: a "quiet" print # of kernel log messages. openstack_user_kernel_options: - key: 'kernel.printk' value: '4 1 7 4' openstack_hosts_package_state: latest ## Octavia {% if 'octavia' in bootstrap_host_scenarios_expanded %} # Enable Octavia V2 API/standalone octavia_v2: True # Disable Octavia V1 API octavia_v1: False octavia_management_net_subnet_cidr: '172.29.232.0/22' tempest_run_concurrency: 0 {% if 'metal' in bootstrap_host_scenarios %} # TODO(mnaser): The Octavia role relies on gathering IPs of hosts in the # LBaaS network and using those in the health manager pool # IPs. We don't store those IPs when running metal so we # have to override it manually. We should remove this and # fix the role (or the inventory tool) eventually. octavia_hm_hosts: 172.29.232.100 # br-lbaas IP {% endif %} {% endif %} {% if 'proxy' in bootstrap_host_scenarios_expanded %} # For testing with the 'proxy' scenario configure deployment environment # to point to the local squid # Playbooks will set a runtime proxy to the AIO host squid deployment_environment_variables: http_proxy: http://172.29.236.100:3128/ https_proxy: http://172.29.236.100:3128/ no_proxy: "localhost,127.0.0.1,172.29.236.100,172.29.236.101,{{ bootstrap_host_public_address | default(ansible_facts['default_ipv4']['address']) }}" # Remove eth0 from all container so there is no default route and everything # must go via the http proxy lxc_container_networks: {} {% endif %} {% if 'ceph' not in bootstrap_host_scenarios_expanded and 'nfs' not in bootstrap_host_scenarios_expanded %} cinder_backends: lvm: volume_group: cinder-volumes volume_driver: cinder.volume.drivers.lvm.LVMVolumeDriver volume_backend_name: LVM_iSCSI iscsi_ip_address: "172.29.236.100" lvm_type: "thin" extra_volume_types: - low-iops - high-iops - ultra-high-iops {% if 'cinder' in bootstrap_host_scenarios %} tempest_test_includelist_cinder: - tempest.scenario.test_minimum_basic.TestMinimumBasicScenario.test_minimum_basic_scenario {% endif %} {% endif %} {% if 'quorum' in bootstrap_host_scenarios_expanded %} oslomsg_rabbit_quorum_queues: True {% endif %}