Test execution results:
[^vmwaremapnew2.xls]
| | vmware datacenter cloudstack zone mapping model | | | | |
| | DVS switch | | | | |
| | Hypervisor: ESXi 4.1 ESXi 5.0 | | | | |
| | | | | | |
| | \\ | | | | |
| | | | | | |
| Test Id | Test case description | Steps | Expected Result | Priority | VMWARE |
| | | | | | |
| | | | | | |
| | Zone configuration - Advance | | | | |
| 1 | Vcenter CPP 4.2 \\
VC1 - DC1 \- C1 - H1 zone1 - C1 \- H1 \\
H2 H2 \\
PS1 \\
PS2 \\
C2 - H3 C2 \- H3 \\
PS3 \\
DC2 - C3 - H4 zone2 - C3 \- H4 \\
PS4 \\
ZPS5 \\
\\
ZPS6 | MS - advance zone 1 Vcenter 2 DC with multiple clusters | zone creation successful | P1 | PASS |
| 2 | Vcenter CPP 4.2 \\
VC1 - DC1 \- C1 - H1 zone1 - C1 \- H1 \\
H2 H2 \\
PS1 \\
PS2 \\
C2 - H3 C2 \- H3 \\
PS3 \\
VC2 - DC3 - C4 - H4 zone2 - C4 - H4 \\
PS4 \\
ZPS5 \\
ZPS6 \\
clusters in different subnet | MS - advance zone multiple Vcenters 2 DC multiple clusters \\
different subnet | zone creation successful | P1 | PASS |
| | | | | | |
| | Zone DC | | | | |
| | AddVMwareDC | 1. Vcenter create DC cluster hosts \\
2. MS Add the Vcenter DC to zone | | P1 | PASS |
| | RemoveVMware DC | MS Remove Vcenter DC from zone | \[http://10.223.195.52:8080/client/api?command=removeVmwareDc \\
\] &zoneid=37751cd3-7fe5-4164-be39-d26a0807a469&_=1377404578209 \\
{ "removevmwaredcresponse" : {"uuidList":\[\],"errorcode":530,\\
"cserrorcode":9999,"errortext":"Zone has one or more clusters. \\
Can't remove VMware datacenter to zone which already has clusters."} |
}
\\
\| P1 \| PASS \|
| | ListVMwareDC | MS List Vmware DCs in zone | \[http://10.223.195.52:8080/client/api?command=listVmwareDcs \\
\] &zoneid=37751cd3-7fe5-4164-be39-d26a0807a469&_=1377403694960 \\
{ "listvmwaredcsresponse" : { "count":1 ,"VMwareDC" : \[ {"id":"fbf6e9e5-dc99-4de2-93a1-dc23c9b4af2d",\\
"zoneid":1,"name":"d1","vcenter":"10.223.52.60"} |
\] }
}
\\
\| P1 \| PASS \|
| | AddVMwarecluster | MS Add Vmware cluster to Vcenter DC in zone | \[http://10.223.195.52:8080/client/api?command=addCluster \\
\] &zoneId=37751cd3-7fe5-4164-be39-d26a0807a469&hypervisor=VMware&clustertype=ExternalManaged \\
&podId=299f284a-7bc7-482b-9278-0b92a8f790c9&username=Administrator \\
&password=Password2&publicvswitchtype=vmwaredvs&guestvswitchtype=vmwaredvs \\
&url=http%3A%2F%2F10.223.52.60%2Fd1%2Fc2&clustername=10.223 | P1 | PASS |
| | RemoveVMwarecluster | MS Remove Vmware cluster from Vcenter DC in zone | \[http://10.223.195.52:8080/client/api?command=deleteCluster \\
\] &id=0925469d-deac-47f4-8844-2029a1f001b7 \\
{ "deleteclusterresponse" : { "success" : "true"} |
}
\\
\| P1 \| PASS \|
| | disablecluster | disable vmware cluster | \[http://10.223.195.52:8080/client/api?command=updateCluster \\
\] &id=0925469d-deac-47f4-8844-2029a1f001b7&allocationstate=Disabled \\
{ "updateclusterresponse" : { "cluster" : {"id":"0925469d-deac-47f4-8844-2029a1f001b7",\\
"name":"10.223.52.60/d1/c2","podid":"299f284a-7bc7-482b-9278-0b92a8f790c9",\\
"podname":"z1p1","zoneid":"37751cd3-7fe5-4164-be39-d26a0807a469",\\
"zonename":"z1","hypervisortype":"VMware","clustertype":"ExternalManaged",\\
"allocationstate":"Disabled","managedstate":"Managed","cpuovercommitratio":"1",\\
"memoryovercommitratio":"1"} |
}
} \| P1 \| PASS \|
| | enablecluster | enable vmware cluster | \[http://10.223.195.52:8080/client/api?command=updateCluster \\
\] &id=0925469d-deac-47f4-8844-2029a1f001b7&allocationstate=Enabled \\
{ "updateclusterresponse" : { "cluster" : {"id":"0925469d-deac-47f4-8844-2029a1f001b7",\\
"name":"10.223.52.60/d1/c2","podid":"299f284a-7bc7-482b-9278-0b92a8f790c9",\\
"podname":"z1p1","zoneid":"37751cd3-7fe5-4164-be39-d26a0807a469",\\
"zonename":"z1","hypervisortype":"VMware","clustertype":"ExternalManaged",\\
"allocationstate":"Enabled","managedstate":"Managed","cpuovercommitratio":"1",\\
"memoryovercommitratio":"1"} |
}
}
\\
\| P1 \| PASS \|
| | managecluster | manage vmware cluster | \[http://10.223.195.52:8080/client/api?command=updateCluster \\
\] &id=0925469d-deac-47f4-8844-2029a1f001b7&managedstate=Managed \\
{ "updateclusterresponse" : { "cluster" : {"id":"0925469d-deac-47f4-8844-2029a1f001b7",\\
"name":"10.223.52.60/d1/c2","podid":"299f284a-7bc7-482b-9278-0b92a8f790c9",\\
"podname":"z1p1","zoneid":"37751cd3-7fe5-4164-be39-d26a0807a469",\\
"zonename":"z1","hypervisortype":"VMware","clustertype":"ExternalManaged",\\
"allocationstate":"Enabled","managedstate":"Managed","cpuovercommitratio":"1",\\
"memoryovercommitratio":"1"} |
}
} \| P1 \| PASS \|
| | unmanagecluster | unmanagae vmware cluster | \[http://10.223.195.52:8080/client/api?command=updateCluster \\
\] &id=0925469d-deac-47f4-8844-2029a1f001b7&managedstate=Unmanaged \\
{ "updateclusterresponse" : { "cluster" : {"id":"0925469d-deac-47f4-8844-2029a1f001b7",\\
"name":"10.223.52.60/d1/c2","podid":"299f284a-7bc7-482b-9278-0b92a8f790c9",\\
"podname":"z1p1","zoneid":"37751cd3-7fe5-4164-be39-d26a0807a469",\\
"zonename":"z1","hypervisortype":"VMware","clustertype":"ExternalManaged",\\
"allocationstate":"Enabled","managedstate":"Unmanaged","cpuovercommitratio":"1",\\
"memoryovercommitratio":"1"} |
}
}
\\
\| P1 \| PASS \|
| | | | | | |
| | Network | M | | | |
| | isolated VPC network | 1. In advance zone create isolated VPC network | \[http://10.223.195.52:8080/client/api?command=createVPC&name=vpc1 \\
\] &displaytext=vpc1&zoneid=37751cd3-7fe5-4164-be39-d26a0807a469 \\
&cidr=10.1.1.1%2F16&vpcofferingid=ca01a99f-579f-4c13-9701-e2281868bcfb&_=1377404893237 \\
{ "createvpcresponse" : {"id":"409ad36f-eb5b-4151-b987-d6a87d518e4f",\\
"jobid":"0ce3db96-580e-4433-a4df-afe888505ee5"} |
}
\\
\| P1 \| PASS \|
| | isolated nonVPC network | 1. In advance zone create non VPC network | non VPC network creation successful | P1 | PASS |
| | listvpc | | | | PASS |
| | | | | | |
| | | | | | |
| | | | | | |
| | network states | | | | |
| | Extend IP range of the network | 1. Advance zone Create Shared NW1 scope zone . | Vm deployment should succeed. | P1 | PASS |
| | | 2. Deploy few Vms in this network. | Vm should be assigned address from the extended range. |
| | | 3. Consume all ips in range | All VMs in NW1 unable to access each other | | |
| | | 4. Extend Ip range. | All VMs in NW1 able to reach DHCP server, gateway | |
| | | 5. Deploy Vm in this network. | | | |
| | Restart network | 1. Advance zone Create Shared NW1 scope zone . | Network restart should succeed. | P1 | PASS |
| | | 2. Deploy few Vms in this network. | After network restart: | | |
| | | 3. Restart Network. | All VMs in NW1 unable to access each other | | |
| | | | All VMs in NW1 able to reach DHCP server, gateway | |
| | | | We should be able to deploy new Vms in this network. | |
| | Restart network with cleanup option | 1. Advance zone Create Shared NW1 scope zone . | Network restart should succeed. After network restart: All VMs \\
in NW1 unable to access each other. All VMs in NW1 able to \\
reach DHCP server, gateway We should be able to deploy new \\
Vms in this network.As part of network restart m Router is \\
stopped and started. Following 3 entries for the router should \\
get deleted and re-created in the host:ovs-ofctl dump-flows xenbr01. cookie=0x0, duration=3503.373s, table=0, n_packets=8, n_bytes=2748, priority=100,udp,dl_vlan=998,nw_dst=255.255.255.255, \\
tp_dst=67 actions=strip_vlan,output:182. cookie=0x0, \\
duration=3503.38s, table=0, n_packets=20, n_bytes=1148, priority=200,arp,dl_vlan=998,nw_dst=10.223.161.110 actions=strip_vlan,output:183.cookie=0x0, duration=3503.376s, \\
table=0, n_packets=37, n_bytes=3176, priority=150, dl_vlan=998, dl_dst=06:e0:d8:00:00:1c actions=strip_vlan,output:18 | P1 | PASS |
| | | 2. Deploy few Vms in this network. | | | |
| | | 4. Restart Network with cleanup option. | | | |
| | Delete network with vms in "Running" state | 1. Advance zone Create Shared NW1 scope zone <pVLAN1, sVLAN1>. | Network Deletion should fail. | P1 | PASS |
| | | 2. Deploy few Vms in this network | | | |
| | | 3. Delete Network. | | | |
| | Delete network when there are no Vms associated with it. | 1. Advance zone Create Shared NW1 scope zone . | Network Deletion should succeed. | P1 | |
| | | 2. Deploy few Vms in this network. | | | |
| | | 3. Destroy all the Vms. | | | |
| | | 4. Delete Network after all the Vms are expunged. | | | |
| | Stop all Vms in network and wait for network shutdown | 1. Advance zone Create Shared NW1 scope zone . \\
2. Deploy few Vms in this network. \\
3. Stop all the Vms. \\
4. Wait for network scavenger thread to run. | Network should not be picked up for Shutting down. | P1 | PASS |
| | delete one of the IP range while not in use | In network with multiple IP ranges & no VMs in network, delete one of the IP range | delete one of the IP range while not in use succeed | P1 | PASS |
| | delete one of the IP range while in use in | In network with multiple IP ranges & VMs in IP range of network, delete one of the IP range with VMs. | delete one of the IP range while in use by VMs fail | P1 | PASS |
| | | | | | |
| | | | | | |
| | VM Deployment isolated VPC and nonVPC network | | | | |
| | advance zone VPC network Deploy VM centos with data disk | 1. In advance zone create VPC network \\
2. Deploy VMs with data disk using centos template. Check PF LB rules | 2. VMs in running state. PF LB rules should work | P1 | PASS |
| | advance zone VPC network Deploy VM centos without data disk | 1. In advance zone create VPC network \\
2. Deploy VMs without data disk using centos template. Check PF LB rules | 2. VMs in running state. PF LB rules should work | P1 | PASS |
| | advance zone VPC network Deploy VM windows with data disk | 1. In advance zone create VPC network \\
2. Deploy VMs using windows template. | 2. VMs in running state. PF LB rules should work | P1 | PASS |
| | advance zone nonVPC network Deploy VM centos with data disk | 1. In advance zone create nonVPC network \\
2. Deploy VMs with data disk using centos template. Check PF LB rules | 2. VMs in running state. PF LB rules should work | P1 | PASS |
| | advance zone nonVPC network Deploy VM centos without data disk | 1. In advance zone create nonVPC network \\
2. Deploy VMs without data disk using centos template. Check PF LB rules | 2. VMs in running state. PF LB rules should work | P1 | PASS |
| | | | | | |
| | VM Life cycle | | | | |
| | Stop VM | Stop an existing VM that is in "Running" State. \\
\[we should flesh out the steps here starting with 'deploy a vm'\] | 1. Should Not be able to login to the VM. \\
2. listVM command should return this VM.State of this VM should be "Stopped". \\
3. DB check : VM_INSTANCE table - state should be "Stopped" | P1 | PASS |
| | Start VM | Start an existing VM that is in "Stopped" State. | 1. Should be able to login to the VM. \\
2. listVM command should return this VM.State of this VM should be "Running". \\
3. DB check : VM_INSTANCE table - state should be "Running" | P1 | PASS |
| | Destroy VM | Destroy an existing VM that is in "Running" State. | 1. Should not be able to login to the VM. \\
2. listVM command should return this VM.State of this VM should be "Destroyed". \\
3. DB check : VM_INSTANCE table - state should be "Destroyed" | P1 | PASS |
| | Restore VM | Restore a VM instance that is in "Destroyed" state. | 1. listVM command should return this VM.State of this VM should be "Stopped". \\
2. We should be able to Start this VM successfully. \\
3. DB check : VM_INSTANCE table - state should be "Destroyed" | P1 | PASS |
| | Destroy VM (Expunged) | Destroy an existing VM that is in "Running" State. \\
Wait for the Expunge Interval (expunge.delay) . | 1. listVM command should NOT return this VM any more. \\
2. DB check : 1. VM_INSTANCE table - state should be "Expunging". \\
2. No entries relating to this VM in the NICS table. \\
3. No entries relating to this VM in the VOLUMES table. \\
4. Make sure the volumes get removed from the Primary Storage. | P1 | PASS |
| | Reboot VM | 1.Deploy VM. \\
2.Reboot VM. | 1. Should be able to login to the VM. \\
2. listVM command should return the deployed VM. \\
State of this VM should be "Running". | P1 | PASS |
| | Migrate VM | 1.Deploy VM. | 1. Should be able to login to the VM. | P1 | PASS |
| | | 2.Check to make sure you are able to log in to this VM. | 2. listVM command should return this VM.State of this VM should be "Running" \\
and the host should be the host to which the VM was migrated to. | | |
| | | 3.Migrate vm to another host in the same cluster. | | | |
| | Attach ISO | 1.Deploy VM. | 1.Log in to the Vm. We should be able to see a device attached. | P1 | |
| | | 2.Attach ISO to this VM. | 2. You should be able to mount this device and \\
use the ISO. | | |
| | Detach ISO | 1.Deploy VM. | 1.Log in to the Vm. We should be see device is not attached to the VM anymore. | P1 | |
| | | 2.Attach ISO to this VM. Log in to the Vm and make sure you see \\
a device attached which has the ISO. | | | |
| | | 3. Detach ISO. | | | |
| | Change Service to use less CPU and Memory | 1.Deploy VM using default template, "small" service offering and small data disk offering. | 1.Log in to the Vm .We should see that the CPU and memory Info of this Vm \\
matches the one specified for "Medium" service offering. | P1 | |
| | | 2.Stop VM | 2. Using listVM command verify that this Vm \\
has "Medium" service offering Id. | | |
| | | 3.Change Service of this Vm to use "Medium" service offr | | | |
| | | 4.Start VM | | | |
| | Change Service to use more CPU and Memory | 1.Deploy VM using default template, "Medium' service offering and small data disk offering. | 1.Log in to the Vm .We should see that the CPU and memory Info of this Vm \\
matches the one specified for "Medium" service offering. | P1 | |
| | | 2.Stop VM | 2. Using listVM command verify that this Vm \\
has "small" service offering Id. | | |
| | | 3.Change Service of this Vm to use "Small" service offering. | | |
| | | 4.Start VM | | | |
| | | | | | |
| | SSVM CPVM VR Life cycle | | | | |
| | Stop SSVM CPVM VR | Stop SSVM VR that is in "Running" State. \\
\[we should flesh out the steps here starting with 'deploy a vm' | 1. Should Not be able to login to the VM. \\
2. listVM command should return this VM.State of this VM should be "Stopped". \\
3. DB check : VM_INSTANCE table - state should be "Stopped" | P1 | PASS |
| | start SSVM CPVM VR | Start an existing VM that is in "Stopped" State. | 1. Should be able to login to the VM. \\
2. listVM command should return this VM.State of this VM should be "Running". \\
3. DB check : VM_INSTANCE table - state should be "Running" | P1 | PASS |
| | Destroy SSVM CPVM VR | Destroy an existing VM that is in "Running" State. | 1. Should not be able to login to the VM. \\
2. listVM command should return this VM.State of this VM should be "Destroyed". \\
3. DB check : VM_INSTANCE table - state should be "Destroyed" | P1 | PASS |
| | Restore SSVM CPVM VR | Restore a VM instance that is in "Destroyed" state. | 1. listVM command should return this VM.State of this VM should be "Stopped". \\
2. We should be able to Start this VM successfully. \\
3. DB check : VM_INSTANCE table - state should be "Destroyed" | P1 | |
| | Destroy SSVM CPVM VR (Expunged) | Destroy an existing VM that is in "Running" State. \\
Wait for the Expunge Interval (expunge.delay) . | 1. listVM command should NOT return this VM any more. \\
2. DB check : 1. VM_INSTANCE table - state should be "Expunging". \\
2. No entries relating to this VM in the NICS table. \\
3. No entries relating to this VM in the VOLUMES table. \\
4. Make sure volumes get removed from the Primary Storage. \\
New SSVM CPVM are created | P1 | PASS |
| | Reboot SSVM CPVM VR | 1.Deploy VM. \\
2.Reboot VM. | 1. Should be able to login to the VM. \\
2. listVM command should return the deployed VM. \\
State of this VM should be "Running". | P1 | PASS |
| | Migrate SSVM CPVM VR | 1.Deploy VM. | 1. Should be able to login to the VM. | P1 | PASS |
| | | 2.Check to make sure you are able to log in to this VM. | 2. listVM command should return this VM.State of this VM should be "Running" \\
and the host should be the host to which the VM was migrated to. | | |
| | | 3.Migrate vm to another host in the same cluster. | | | |
| | | | | | |
| | VM live migration | | | | |
| 1 | Vcenter CPP 4.2 \\
VC1 - DC1 \- C1 - H1 zone1 - C1 \- H1 \\
H2 H2 \\
PS1 cluster primary storage \\
PS2 \\
C2 - H3 C2 \- H3 \\
PS3 \\
DC2 - C3 - H4 zone2 - C3 \- H4 \\
PS4 \\
ZPS5 zone primary storage \\
ZPS6 | VM Migrate H1 to H2 vice versa \\
VM Migrate H1 to H3 vice versa \\
VM Migrate H1 to H4 vice versa \\ | successful | P1 | PASS |
| 2 | Vcenter CPP 4.2 \\
VC1 - DC1 \- C1 - H1 zone1 - C1 \- H1 \\
H2 H2 \\
PS1 cluster primary storage \\
PS2 \\
C2 - H3 C2 \- H3 \\
PS3 \\
VC2 - DC3 - C4 - H4 zone2 - C4 - H4 \\
PS4 \\
ZPS5 zone primary storage \\
ZPS6 \\
clusters in different subnet | VM Migrate H1 to H2 vice versa \\
VM Migrate H1 to H3 vice versa \\
VM Migrate H1 to H4 vice versa | successful | P1 | PASS |
| | | | | | |
| | Storage volume | | | | |
| | Attaching Volumes | 1.navigate to Storage-volumes | 1.shows list of volumes | P1 | |
| | | 2.select the detached data disk | 3."Attach Disk" pop-up box will display with list of instances |
| | | 3.click on "Actions" and select "Attach Disk" to attach to a particular instace | 4. disk should be attached to instance successfully and both \\
UI and database(volumes table) should reflect the changes \\
( i.e attached disk details should be updated with attached \\
vm_instance ID and device id ) | | |
| | | 4. select the instance and click "Ok" button to attach the disk to selected \\
instance in the "Attach Disk" pop-up window | | | |
| | Detaching Volumes | 1.navigate to Storage-volumes | data disk should be detached from instance and detached data \\
disk details should be updated properly(i.e Instance Name \\
\-detached and Device ID is null) | P1 | |
| | | 2. selected the data disk which is attached to instance | | | |
| | | 3.click "Actions" and select the "Detach Disk" | | | |
| | Download volumes | 1.navigate to Storage-volumes | 3. download volume will fail with proper error message "Failed \\
\- Invalid state of the volume with ID: . It should be either \\
detached or the VM should be in stopped state" | P1 | |
| | | 2. selected the data disk which is attached to instance | . | | |
| | | \[some fundamental steps here are missing. in each os after attaching disk, \\
you need to format and mount it before writing to it. downloading a blank \\
volume would probably fail | 5. able to download the volume when its not attached to instace |
| | | 3.perform download volume | | | |
| | | 4.select another data disk with is not attahced instance | | | |
| | | 5.perform "download volume " | | | |
| | Delete Volumes | case 1 | case 1: | P1 | |
| | | 1.navigate to Storage-volumes | volume should be deleted successfully and listVolume should not \\
contain the deleted volume details. | | |
| | | 2. selected the data disk | | | |
| | | 3.click on 'Actions" menu and select the "Delete volume" | Case 2: | | |
| | | | "Delete Volume" menu item not shown under "Actions" menu. |
| | | case 2 | (UI should not allow to delete the volume when it is attahced to \\
instance by hiding the menu Item) | | |
| | | 4.select another data disk with is attahced instance | | | |
| | | 5.clcik on "Actions" menu and check for "Delete Volume" | | | |
| | Create Volume | 1. Perform "Add volume" by provinding name, availability zone and disk offering details.(i.e go to Storage-Volumes-) | volume should be created successfully and listVolumes should \\
contain the created volumes. Database should reflects with new \\
created volume details(volumes table) | P1 | PASS |
| | | \[we shoudl execute this against every storage size, sanity check against that but also a few custom sized disk offerings\] |
| | | | | | |
| | Upload a data disk | call upload volume API with following parameters | Upload volume is successful | P1 | |
| | | | | | |
| | | HTTP URL of the data disk, | | | |
| | | Zone ID, | | | |
| | | Name, | | | |
| | | Description, | | | |
| | | Hyper visor | | | |
| | | | | | |
| | Attach a uploaded data disk | call attach volume api with following parameter | Verify Attach Volume API will move (and not Copy) the volume from \\
secondary storage to primary storage and attach to the vm. | P1 | |
| | | | | | |
| | | Volumeid | | | |
| | | VM | | | |
| | | | | | |
| | snapshot | | | | |
| | Create VMSnapshot of Vm in "Stopped" state with root volume with "vm_snapshot_type" set to "Disk" | 1. Deploy Vm. \\
2. Log in to the Vm and create few directories and files. \\
3. Stop the VM. \\
4. Create VMSnapshot for this VM with "vm_snapshot_type" set to "Disk" | Snapshot should get created successfully. \\
It should be stored in the Primary Storage. | P1 | [https://issues.apache.org/jira/browse/CLOUDSTACK-4458|https://issues.apache.org/jira/browse/CLOUDSTACK-4458] |
| | Create VMSnapshot of Vm in "Stopped" state with 1 data disk with "vm_snapshot_type" set to "Disk" | 1. Deploy Vm. Attach a data disk to this VM. \\
2. Log in to the Vm and create few directories and files on the root volume. Create few more directories on the data disk. \\
3. Stop the VM. \\
4. Create VMSnapshot for this VM with "vm_snapshot_type" set to "Disk" | Snapshot should get created successfully. \\
It should be stored in the Primary Storage. | P1 | [https://issues.apache.org/jira/browse/CLOUDSTACK-4458|https://issues.apache.org/jira/browse/CLOUDSTACK-4458] |
| | We should be allowed to initiate Snapshots for multiple Vms in parallel | 1. Deploy few Vms. \\
2. Log in to the each of the Vms and create few directories and files. \\
3. Create VMSnapshot for all VMs at the same time. | Vm Snapshot process for all the Vms should happen in parallel. | P1 | [https://issues.apache.org/jira/browse/CLOUDSTACK-4458|https://issues.apache.org/jira/browse/CLOUDSTACK-4458] |
| | We should be abe to revert multiple Vms to different snapshots in parallel | 1. Deploy few Vms. \\
2. Create Vm snapshots for all the Vms. \\
3. Revert all the above Vms to a snapshot of these Vms in parallel | Vm reverts for all the Vms should happen in parallel. | P1 | |
| | | | | | |
| | 2. Create volume from snapshot | | | | PASS |
| | 3. Create template from snapshot | | | | PASS |
| | | | | | |
| | | | | | |
| | Storage Migration | | | | |
| 1 | Vcenter CPP 4.2 \\
VC1 - DC1 \- C1 - H1 zone1 - C1 \- H1 \\
H2 H2 \\
PS1 cluster primary storage \\
PS2 \\
C2 - H3 C2 \- H3 \\
PS3 \\
DC2 - C3 - H4 zone2 - C3 \- H4 \\
PS4 \\
ZPS5 zone primary storage \\
ZPS6 | storage Migrate PS1 to PS2 vice versa \\
storage Migrate PS1 to PS3 vice versa \\
storage Migrate PS1 to PS4 vice versa \\ | successful | P1 | PASS |
| 2 | Vcenter CPP 4.2 \\
VC1 - DC1 \- C1 - H1 zone1 - C1 \- H1 \\
H2 H2 \\
PS1 cluster primary storage \\
PS2 \\
C2 - H3 C2 \- H3 \\
PS3 \\
VC2 - DC3 - C4 - H4 zone2 - C4 - H4 \\
PS4 \\
ZPS5 zone primary storage \\
ZPS6 \\
clusters in different subnet | storage Migrate PS1 to PS2 vice versa \\
storage Migrate PS1 to PS3 vice versa \\
storage Migrate PS1 to PS4 vice versa \\ | successful | P1 | PASS |
| 15 | Migrate volume when VM is running | Use MigrateVirtualMachine() with VMId , SPId , when VM is in \\
Running state | Should throw appropriate error. Storage migration can happen only if the Vm is in Stopped state. | P1 | PASS |
| | Storage Migration Negative Scenarios | | | | |
| | CASE I : start VM migration. \\
When the Volume migration is still in progress ,restart the management server. | | Vm migration should fail. There should be a roll back done. Vm state should get back to" Stopped" and volumes should get back to "Ready" state. \\
Attempt to start the VM should succeed using the existing (older) storage. | P1 | |
| | CASE II : start VM migration. \\
When the Volume migration is still in progress , shutdown host which is doing the storage migration. | | Vm migration should fail. There should be a roll back done. Vm state should get back to" Stopped" and volumes should get back to "Ready" state \\
Attempt to start the VM should succeed using the existing (older) storage. | P1 | |
| | CASE III : start VM migration. \\
When the Volume migration is still in progress , reboot host which is doing the storage migration. | | Vm migration should fail. There should be a roll back done. Vm state should get back to" Stopped" and volumes should get back to "Ready" state \\
Attempt to start the VM should succeed using the existing (older) storage. | P1 | |
| | CASE IV : migrate vm to a storage pool which belongs to a cluster of different \\
hypervisor type | | The migration should fail with appropriate error message. | P1 | |
| | | | | | |
| | | | | | |
| | Host | | | | |
| | Force reconnect Host() | Force reconnect Host() | | P1 | PASS |
| | UpdateHost() - Edit host tags | UpdateHost() - Edit host tags | | P1 | |
| | deleteHost() - > Put the host in maintenance then delete the host | deleteHost() - > Put the host in maintenance and then delete the host | P1 | |
| | Put the host in maintenance mode | Put the host in maintenance mode | | P1 | [https://issues.apache.org/jira/browse/CLOUDSTACK-4513|https://issues.apache.org/jira/browse/CLOUDSTACK-4513] |
| | Put host in maintenance mode bring host out of maintenance mode | Put host in maintenance mode. Cancel host maintenance mode. | P1 | [https://issues.apache.org/jira/browse/CLOUDSTACK-4513|https://issues.apache.org/jira/browse/CLOUDSTACK-4513] |
| | Power down host | Power down host | | P1 | |
| | Power down host and the power on the host | Power down host and the power on the host | | P1 | |
| | Reboot host | Reboot host | | P1 | |
| | Bring down the network connectivity of the host. | Bring down the network connectivity of the host. | | P1 | |
| | Bringdown network connectivity of host. Bringup nw connectivity , | Bring down network connectivity of host. Bring up network connectivity | P1 | |
| | | | | | |
| | Template | | | | |
| | add template | 1. Add public/private template | 1.database (vm_template table) should be updated with newly created template | P1 | |
| | | 2. check UI/ listTemplates command to see the newly added template. | 2.UI should show the newly added template | P1 | |
| | | | 3.listtemplates API should show the newly added template | P1 | |
| | | | | P1 | |
| | delete template | Delete Template: | 1.listtemplates should not show the delted template | P1 | |
| | | 1.delete a template using UI/ API | 2.database should be updated. | P1 | |
| | | | 3. tempalte should not be displayed in admin UI. | P1 | |
| | | | 4. template should be removed from secondary store (if no VMs were deployed \\
with it) after storage.cleanup.interval | P1 | |
| | Edit template | Edit Template: | 1.data base should be updated with updated values | P1 | |
| | | 1. edit the template attribute | 2.UI should show the updated template attributes | P1 | |
| | | 2. check the updated values are relfecte | 3.listTemplate shoud show the template with updated values | P1 | |
| | Download template | "1. Add public/private template | "1. download template should generate a valid http link to download | P1 | |
| | | 2. select the template and click ""Actions"" | 2. You should be able to download the template with out any issues (no 404, 530, \\
503 HTTP Errors)" | | |
| | | 3. perform ""Download Template"" | | | |
| | Check admin can extract any template | admin should able extract and download the templates | | P1 | |
| | Check all public templates should be visible to the users and admins | listTemplates should display all the public templates for all kind of users | P1 | |
| | check all the system VM templates are not visble to users | Listtemplate should not display the system templates" | | P1 | |
| | | | | P1 | |
| | Copy template to other zone | | Copy template to other zone successful | P1 | |
| | Start a VM from template | | Start a VM from template successful | P1 | |
| | Download template with isolated storage network | 1.Bring up CS in advanced zone | Verify that SSVM downloads the guest OS templates from storage server which \\
is in isolated network. | P1 | |
| | | 2.create zone->pod->cluster>host | Use below sql query to verify the download status of the template. |
| | | 3.Primary and secondary storage on storage server in isolated network \\
(Only management network has the reachability to storage server, \\
public network does not have the reachability) | mysql> select * from template_host_ref; | | |
| | | 4.Disable zone | | | |
| | | 5.Set global settings parameter "secstorage.allowed.internal.sites" to \\
storage server subnet | | | |
| | | 6.Restart management server | | | |
| | | 7.Enable zone | | | |
| | | | | | |
| | ISO | | | | |
| | verify add ISO | 1. Add public/private ISO | 1.database (vm_template table) should be updated with newly created ISO | P1 | |
| | | 2. check UI/ listISO command to see the newly added template. | 2.UI should show the newly added ISO | | |
| | | | 3.listISos API should show the newly added templat | |
| | | | | | |
| | verify delete ISO | 1.delete a template using UI/ API | 1.listIsos should not show the delted template | P1 | |
| | | | 2.database should be updated. | | |
| | | | 3.ISO should not be displayed in admin UI. | | |
| | | | 4. ISO should be removed from secondary store (if no VMs were deployed with it) \\
after storage.cleanup.interval | | |
| | Verify Edit ISO | 1. edit the ISO attribute | 1.data base should be updated with updated values | P1 | |
| | | 2. check the updated values are relfected | 2.UI should show the updated ISO attributes | | |
| | | | 3.listISos shoud show the template with updated value | |
| | | | | | |
| | Download ISO | 1. Add public/private ISO | 1. download template should generate the link to download and able to download \\
the template with out any issues | P1 | |
| | | 2. select the ISO and click "Actions" | :check the ssvm in /var/www/html/userdata | | |
| | | 3. perform "Download Template" | | | |
| | Copy ISO to other zone | 1. perform copy ISO from one zone to another zone | copy ISO should be successful and seconday storage should contain new copied ISO | P1 | |
| | | 2.check the newly copied template in another zone | | | |
| | | | | | |
| | | | | | |
| | primary storage | | | | |
| | Add Primary storage NFS cluster wide | Use createStoragePool with url, name, podid, clusterid | Primary storage should be created and should be in Up state | P1 | PASS |
| | Delete Primary storage NFS | Delete Primary storage NFS | Delete successful | P1 | PASS |
| | | | | P1 | |
| | | | | P1 | |
| | | | | P1 | |
| | Storage maintenance mode with one pool | Initiate maintenance mode for primary Storage pool | VMs and System VMs stopped, Volumes should not be destroyed | P1 | |
| | | (no additional pool present) | | P1 | |
| | Storage maintenance mode with multiple pools | Initiate maintenance mode for Primary storage with additional pools | User VMs should be stopped. System VMs should be restarted in another pool | P1 | |
| | Storage maintenance mode cancel | Cancel maintenance mode (one pool) | User VMs that were stopped should restart and also the system VMs | P1 | |
| | Storage maintenance mode cancel | Cancel maintenance mode (multiple pools) | User VMs that were stopped should restart | P1 | |
| | Enable maintenance with local storage | Enable local storage for System VM and User Vm | User and System VMs should be stopped and restarted when maintenance mode is \\
disabled \(?) | P1 | |
| | | Put Local storage in maintenance mode | | P1 | |
| | Create VMs when Maintenance mode is on | Create User VMs when storage pools are in maintenance mode | Should not be able to create VMs and appropriate message should be thrown | P1 | |
| | Storage Tags | Create Storage tag for storage and use the tag for disk offering | Should deploy disk on the storage which matches the tag | P1 | |
| | | | | | |
| | | | | | |
| | primary storage zone wide | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | secondary storage | | | | |
| | Add NFS Secondary storage | Seed System VM templates | Secondary storage should get added to the zone. | | |
| | | Use addSecondaryStorage with url, host | listHosts should should show up the added sec storage as type "secondarystorage" |
| | | | listSystemVm should show up system vms once they are up and running |
| | | | listTemplates should show default templates downloaded and and in Ready state |
| | | | | | |
| | Change Sec Storage server | Stop cloud services | Sec storage should be back up | | |
| | | Copy files from old to new secondary storage | Should be able to take snapshots and create templates | |
| | | Change IP in DB in host and host_details | | | |
| | | Start services | | | |
| | | Stop and start SSVM | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | PVLAN | | | | |
| | | | | | |
| | Shared Network scope All 1 PVLAN Deploy VM | 1. Advance zone cluster with 2 hosts, Domain D1 domainadmin d1domain, user d1user. Domain D2 user d2user. \\
2. Create Shared NW1 scope All <pVLAN1, sVLAN1>. \\
3. As users from different domains , deploy Vms in this network, | 1. shared NW1 with pVLAN creation succeed. \\
2. All accounts able to create VMs on NW1. \\
3.All VMs in NW1 unable to access each other All VMs in NW1 able to reach DHCP server, gateway | P1 | |
| | Shared NW scope Domain 1 PVLAN Deploy VM | 1. Advance zone cluster with 2 hosts, Domain D1 domainadmin d1domain, user d1user. Domain D2 user d2user. \\
2. Create Shared NW1 scope Domain for D1 <pVLAN1, sVLAN1>. \\
3. As users from different domains , deploy Vms in this network, | 1. shared NW1 with pVLAN creation succeed. \\
2. Only accounts from doamin d1 are allowed to create VMs on NW1. \\
3.All VMs in NW1 unable to access each other All VMs in NW1 able to reach DHCP server, gateway | P1 | |
| | | | | | |
| | UPGRADE | | | | |
| 1 | 2.2.16 \-> campo 4.2. \\
Deployment\- multiple zones, cluster single DC\\
VC DC cluster host CP\\
vc1 a1 c1 h1 zone 1\\
h2\\
\\
after upgrade - VC - add cluster c2 to vc1 a1 : \\
VC DC cluster host CP\\
vc1 a1 c1 h1 zone 1 \\
pod1 c1 h1\\
h2 h2 \\
PS1\\
\\
c2 h4 pod2 c2 h4 \\
PS2\\
ZPS5 ZPS6 | 1. 2.2.6 vmware setup\\
Deployment\- multiple zones, cluster single DC\\
deploy VM, assign public IP to VM\\
ensure internet access\\
\\
2. upgrade campo 4.2 \\
add cluster c2 to Vcenter.\\
add pod2 cluster c2 host h4 to zone1 \\
add cluster PS \\
\\
\\ | New mapping Vcenter & CPP should work \\
1. legacy has single DC \\ | P1 | |
| 2 | Legacy Vcenter Legacy CPP 4.2 \\
VC1 - DC1 \- C1 - H1 zone1 - cluster1 VC1 - DC1 - C1 \- H1 \\
H2 H2 \\
PS1 cluster primary storage \\
PS2 \\
DC2 - C2 - H3 cluster2 VC1 - DC2 - C2 - H3 \\
PS3 \\
DC3 - C3 - H4 cluster3 VC1 - DC3 \- C3 - H4 \\
PS4 \\
ZPS5 zone primary storage \\
ZPS6 \\
2. upgrade campo 4.2 \\
add to current zone: \\
Vcenter1 - DC1 \- cluster1 - Host1 \\
cluster2 - Host2 \\
DC2 \- cluster3 - Host3 \\
Vcenter2 - DC3 \- cluster4 - Host4 \\
clusters in different subnet | Legacy CPP 3.0.7 Patch B\- advance zone 1 Vcenter 2 DC with multiple clusters \\
upgrade campo 4.2 \\
\\ | New mapping Vcenter & CPP NOT allowed | P1 | |
| 3 | Vcenter CPP 4.2 \\
VC1 - DC1 \- C1 - H1 zone1 - cluster1 VC1 - DC1 - C1 \- H1 \\
H2 H2 \\
PS1 \\
PS2 \\
DC2 - C2 - H3 cluster2 VC1 - DC2 - C2 - H3 \\
PS3 \\
VC2 - DC3 - C4 - H4 cluster3 VC2 - DC3 \- C4 - H4 \\
PS4 \\
ZPS5 zone primary storage \\
ZPS6 \\
clusters in different subnet | Legacy CPP 3.0.7 Patch B - advance zone multiple Vcenters 2 DC multiple clusters different subnet \\
upgrade campo 4.2 | New mapping Vcenter & CPP NOT allowed | P1 | |
| | | | | | |
| 4 | Before upgrade Zone with mixed hypervisors \-\- \\
Later add additional clusters VMWARE or any other | Legacy CPP 4.0 advance zone multiple Vcenters 2 DC multiple clusters different subnet \\
upgrade campo 4.2 | | | |
| | | | | | |
| 5 | Before Upgrade Nexus vSwitch | | | | |
| | | | | | |
| 6 | Before Upgrade DVS 2 physical networks \\
after upgrade add dvs Cluster | | | | |
| | | | | | |
| 7 | Before Upgrade DVS 2 physical networks | | | | |
| | after upgrade add Nexus Cluster | | | | |
| | | | | | |
| 8 | Test upgrade to new product (with this feature) when the deployment \\
already has multiple zones, where clusters in each zone has one or more of following, | | | | |
| | 1. Single DC | | | | |
| | 2. Multiple DCs | | | | |
| | 3. Multiple vCenters | | | | |
| | | | | | |
| 9 | After upgrade, test adding new cluster to existing/new zone, where the \\
cluster belongs to a DC/vCenter that is, | | | | |
| | 1. Already part of zone | | | | |
| | 2. Not part of zone | | | | |
| | | | | | |
| | after upgrade Test creation of new zone | | | | | |