fuel-plugin-nsxv/doc/test_plan/test_suite_destructive.rst

10 KiB
Raw Blame History

Destructive

TC-101: Check abilities to bind port on NSXv to VM, disable and enable this port.

ID

nsxv_ability_to_bind_port

Description :

Verifies that system could manipulate with port.

Complexity

core

Requre to automate

Yes

Steps :

Log in to Horizon Dashboard.
Navigate to Project ->  Compute -> Instances

Launch instance VM_1 with image TestVM-VMDK and flavor m1.tiny.

Launch instance VM_2  with image TestVM-VMDK and flavor m1.tiny.

Verify that VMs  should communicate between each other. Send icmp ping from VM _1 to VM_2  and vice versa.
Disable NSXv_port of VM_1.
Verify that VMs  should not communicate between each other. Send icmp ping from VM _2 to VM_1  and vice versa.

Enable NSXv_port of VM_1.

Verify that VMs  should communicate between each other. Send icmp ping from VM _1 to VM_2  and vice versa.
Pings should get a response

TC-102: Verify that vmclusters should migrate after shutdown controller.

ID

nsxv_shutdown_controller

Description :

Verify that vmclusters should migrate after shutdown controller.

Complexity

core

Requre to automate

No

Steps :

Create a new environment using the Fuel UI Wizard:
add name of env and select release version with OS
as hypervisor type: select vcenter check box and QEMU/KVM radio button
network setup : Neutron with tunnel segmentation.
storage backends: default
additional services: all by default

In Settings tab:
enable NSXv plugin

Add nodes:
3 controllers

Setup Fuel interfaces on slaves:
eth0 - admin(PXE)
eth1 - public
eth2 - management
eth3 - VM(Fixed) ID:103
eth4  storage

Networks tab:
Public network: start '172.16.0.2' end '172.16.0.126'
CIDR '172.16.0.0/24'
Gateway 172.16.0.1
Floating ip range start '172.16.0.130' end '172.16.0.254'
Storage: CIDR '192.168.1.0/24'
Vlan tag is not set
Managment: CIDR '192.168.0.0/24'
Vlan tag is not set
Neutron L2 configuration by default
Neutron L3 configuration by default
Click button 'save settings'
Click button 'verify networks'
Fill vcenter credentials:
Availability zone: vcenter
vCenter host: '172.16.0.254'
vCenter username: <login>
vCenter password: <password>

Add 2 vSphere Clusters:
vSphere Cluster: Cluster1
Service name: vmcluster1
Datastore regex:.*
vSphere Cluster: Cluster2
Service name: vmcluster2
Datastore regex: .*

Deploy Cluster

Run OSTF

Shutdown controller with  vmclusters.

Check that vmclusters should migrate to another controller. Vmclusters should migrate to another controller.

TC-103: Deploy cluster with plugin, addition and deletion of nodes.

ID

nsxv_add_delete_nodes

Description :

Verify that system functionality is ok after redeploy.

Complexity

advanced

Requre to automate

No

Steps :

Create a new environment using the Fuel UI Wizard:
add name of env and select release version with OS
as hypervisor type: select vcenter check box and QEMU/KVM radio button
network setup : Neutron with tunnel segmentation.
storage backends: default
additional services: all by default

In Settings tab:
enable NSXv plugin
select Vmware vcenter esxi datastore for images (glance)

Add nodes:
3 controllers
2 compute-vmwares
1 cinder-vmdk

Interfaces on slaves should be setup this way in Fuel interface:
eth0 - admin(PXE)
eth1 - public
eth2 - management
eth3 - VM(Fixed) ID:103
eth4  storage

Networks tab:
Public network: start '172.16.0.2' end '172.16.0.126'
CIDR '172.16.0.0/24'
Gateway 172.16.0.1
Floating ip range start '172.16.0.130' end '172.16.0.254'
Storage: CIDR '192.168.1.0/24'
Vlan tag is not set
Management: CIDR '192.168.0.0/24'
Vlan tag is not set
Neutron L2 configuration by default
Neutron L3 configuration by default

Verify networks

Fill vcenter credentials:
Availability zone: vcenter
vCenter host: '172.16.0.254'
vCenter username: <login>
vCenter password: <password>

Add 2 vSphere Clusters:
vSphere Cluster: Cluster1
 Service name: vmcluster1
Datastore regex:.*
vSphere Cluster: Cluster2
Service name: vmcluster2
Datastore regex: .*

Run OSTF

Remove node with cinder-vmdk role.

Add node with cinder role

Redeploy cluster.

Run OSTF

Remove node with compute-vmware role
Add node with cinder-vmware  role

Redeploy cluster.

Run OSTFCluster should be deployed and all OSTF test cases should be passed.

TC-104: Deploy cluster with plugin and deletion one node with controller role.

ID

nsxv_add_delete_controller

Description :

Verifies that system functionality is ok when controller has been removed.

Complexity

advanced

Requre to automate

No

Steps :

Create a new environment using the Fuel UI Wizard:
add name of env and select release version with OS
as hypervisor type: select vcenter check box and QEMU/KVM radio button
network setup : Neutron with tunnel segmentation.
storage backends: default
additional services: all by default

In Settings tab:
enable NSXv plugin
select Vmware vcenter esxi datastore for images (glance)

Add nodes:
4 controller
1 compute-vmware
1 cinder-vmdk

Interfaces on slaves should be setup this way in Fuel interface:
eth0 - admin(PXE)
eth1 - public
eth2 - management
eth3 - VM(Fixed) ID:103
eth4  storage

Networks tab:
Public network: start '172.16.0.2' end '172.16.0.126'
CIDR '172.16.0.0/24'
Gateway 172.16.0.1
Floating ip range start '172.16.0.130' end '172.16.0.254'
Storage: CIDR '192.168.1.0/24'
Vlan tag is not set
Management: CIDR '192.168.0.0/24'
Vlan tag is not set
Neutron L2 configuration by default
Neutron L3 configuration by default

Verify networks
Fill vcenter credentials:
Availability zone: vcenter
vCenter host: '172.16.0.254'
vCenter username: <login>
vCenter password: <password>

Add 2 vSphere Clusters:
vSphere Cluster: Cluster1
 Service name: vmcluster1
Datastore regex:.*
vSphere Cluster: Cluster2
Service name: vmcluster2
Datastore regex: .*

Run OSTF
Remove node with controller role.

Redeploy cluster

Run OSTF
Add controller
Redeploy cluster

Run OSTFCluster should be deployed and all OSTF test cases should be passed.

TC-105: Verify that it is not possible to uninstall of Fuel NSXv plugin with deployed environment.

ID

nsxv_plugin

Description :

It is not possible to remove plugin while at least one environment exists.

Complexity

core

Requre to automate

Yes

Steps :

Copy plugin to to the Fuel master node using scp.
Install plugin
fuel plugins --install plugin-name-1.0-0.0.1-0.noarch.rpm

Ensure that plugin is installed successfully using cli, run command 'fuel plugins'.
Connect to the Fuel web UI.

Create a new environment using the Fuel UI Wizard:
add name of env and select release version with OS
as hypervisor type: select vcenter check box and Qemu radio button
 network setup : Neutron with tunnel segmentation
 storage backends: default
additional services: all by default

Click on the Settings tab.

In Settings tab:
enable NSXv plugin

Add nodes:
1 controller

Interfaces on slaves should be setup this way in Fuel interface:
eth0 - admin(PXE)
eth1 - public
eth2 - management
eth3 - VM(Fixed) ID:103
eth4  storage

Networks tab:
Public network: start '172.16.0.2' end '172.16.0.126'
CIDR '172.16.0.0/24'
Gateway 172.16.0.1
Floating ip range start '172.16.0.130' end '172.16.0.254'
Storage: CIDR '192.168.1.0/24'
Vlan tag is not set-Management: CIDR '192.168.0.0/24'
Vlan tag is not set
Neutron L2 configuration by default
Neutron L3 configuration by default

Verify networks.

Fill vcenter credentials:
Availability zone: vcenter
vCenter host: '172.16.0.254'
vCenter username: <login>
vCenter password: <password>

Add 2 vSphere Clusters:
vSphere Cluster: Cluster1
Service name: vmcluster1
Datastore regex:.*

Deploy cluster
Run OSTF
Try to delete plugin via cli Remove plugin from master node  fuel plugins --remove plugin-name==1.0.0
Alert: "400 Client Error: Bad Request (Can't delete plugin which is enabled for some environment.)" should be displayed.

TC-106: Check cluster functionality after reboot vcenter.

ID

nsxv_plugin

Description :

Verifies that system functionality is ok when vcenter has been rebooted.

Complexity

core

Requre to automate

Yes

Steps :

Create a new environment using the Fuel UI Wizard:
add name of env and select release version with OS
as hypervisor type: select vcenter check box and QEMU/KVM radio button
network setup : Neutron with tunnel segmentation.
storage backends: default
additional services: all by default
In Settings tab:
enable NSXv plugin
select Vmware vcenter esxi datastore for images (glance)

Add nodes:
3 controller
1 computer
1 cinder-vmware
1 cinder

Interfaces on slaves should be setup this way in Fuel interface:
eth0 - admin(PXE)
eth1 - public
eth2 - management
eth3 - VM(Fixed) ID:103
eth4  storage

Networks tab:
Public network: start '172.16.0.2' end '172.16.0.126'
CIDR '172.16.0.0/24'
Gateway 172.16.0.1
Floating ip range start '172.16.0.130' end '172.16.0.254'
Storage: CIDR '192.168.1.0/24'
Vlan tag is not set
Management: CIDR '192.168.0.0/24'
Vlan tag is not set
Neutron L2 configuration by default
Neutron L3 configuration by default

Verify networks

Fill vcenter credentials:
Availability zone: vcenter
vCenter host: '172.16.0.254'
vCenter username: <login>
vCenter password: <password>

Add 2 vSphere Clusters:
vSphere Cluster: Cluster1
 Service name: vmcluster1
Datastore regex:.*
vSphere Cluster: Cluster2
Service name: vmcluster2
Datastore regex: .*

Run OSTF

Launch instance VM_1 with image TestVM-VMDK and flavor m1.tiny.

Launch instance VM_2  with image TestVM-VMDK and flavor m1.tiny.

Check connection between VMs, send ping from VM_1 to VM_2 and vice verse.
Reboot vcenter
vmrun -T ws-shared -h https://localhost:443/sdk -u vmware -p VMware01 reset "[standard] vcenter/vcenter.vmx"

Check that controller lost connection with vCenter

Wait for vCenter

Ensure that all instances from vCenter displayed in dashboard.

Ensure connectivity between vcenter1's and vcenter2's VM.
Run OSTF
Cluster should be deployed and all OSTF test cases should be passed. Ping should get response.