[docs] Documentation for BNG PPPoE RFC2544 test cases
[yardstick.git] / docs / testing / user / userguide / opnfv_yardstick_tc093.rst
1 .. This work is licensed under a Creative Commons Attribution 4.0 International
2 .. License.
3 .. http://creativecommons.org/licenses/by/4.0
4 .. (c) OPNFV, Intracom Telecom and others.
5 .. mardim@intracom-telecom.com
6
7 *************************************
8 Yardstick Test Case Description TC093
9 *************************************
10
11 +-----------------------------------------------------------------------------+
12 |SDN Vswitch resilience in non-HA or HA configuration                         |
13 |                                                                             |
14 +--------------+--------------------------------------------------------------+
15 |test case id  | OPNFV_YARDSTICK_TC093: SDN Vswitch resilience in             |
16 |              | non-HA or HA configuration                                   |
17 +--------------+--------------------------------------------------------------+
18 |test purpose  | This test validates that network data plane services are     |
19 |              | resilient in the event of Virtual Switch failure             |
20 |              | in compute nodes. Specifically, the test verifies that       |
21 |              | existing data plane connectivity is not permanently impacted |
22 |              | i.e. all configured network services such as DHCP, ARP, L2,  |
23 |              | L3 Security Groups continue to operate between the existing  |
24 |              | VMs eventually after the Virtual Switches have finished      |
25 |              | rebooting.                                                   |
26 |              |                                                              |
27 |              | The test also validates that new network service operations  |
28 |              | (creating a new VM in the existing L2/L3 network or in a new |
29 |              | network, etc.) are operational after the Virtual Switches    |
30 |              | have recovered from a failure.                               |
31 |              |                                                              |
32 +--------------+--------------------------------------------------------------+
33 |test method   | This testcase first checks if the already configured         |
34 |              | DHCP/ARP/L2/L3/SNAT connectivity is proper. After            |
35 |              | it fails and restarts again the VSwitch services which are   |
36 |              | running on both OpenStack compute nodes, and then checks if  |
37 |              | already configured DHCP/ARP/L2/L3/SNAT connectivity is not   |
38 |              | permanently impacted (even if there are some packet          |
39 |              | loss events) between VMs and the system is able to execute   |
40 |              | new virtual network operations once the Vswitch services     |
41 |              | are restarted and have been fully recovered                  |
42 |              |                                                              |
43 +--------------+--------------------------------------------------------------+
44 |attackers     | In this test case, two attackers called “kill-process” are   |
45 |              | needed. These attackers include three parameters:            |
46 |              |                                                              |
47 |              | 1. fault_type: which is used for finding the attacker's      |
48 |              |    scripts. It should be set to 'kill-process' in this test  |
49 |              |                                                              |
50 |              | 2. process_name: should be set to the name of the Vswitch    |
51 |              |    process                                                   |
52 |              |                                                              |
53 |              | 3. host: which is the name of the compute node where the     |
54 |              |    Vswitch process is running                                |
55 |              |                                                              |
56 |              | e.g. -fault_type: "kill-process"                             |
57 |              |      -process_name: "openvswitch"                            |
58 |              |      -host: node1                                            |
59 |              |                                                              |
60 +--------------+--------------------------------------------------------------+
61 |monitors      | This test case utilizes two monitors of type "ip-status"     |
62 |              | and one monitor of type "process" to track the following     |
63 |              | conditions:                                                  |
64 |              |                                                              |
65 |              | 1. "ping_same_network_l2": monitor ICMP traffic between      |
66 |              |    VMs in the same Neutron network                           |
67 |              |                                                              |
68 |              | 2. "ping_external_snat": monitor ICMP traffic from VMs to    |
69 |              |    an external host on the Internet to verify SNAT           |
70 |              |    functionality.                                            |
71 |              |                                                              |
72 |              | 3. "Vswitch process monitor": a monitor checking the         |
73 |              |    state of the specified Vswitch process. It measures       |
74 |              |    the recovery time of the given process.                   |
75 |              |                                                              |
76 |              | Monitors of type "ip-status" use the "ping" utility to       |
77 |              | verify reachability of a given target IP.                    |
78 |              |                                                              |
79 +--------------+--------------------------------------------------------------+
80 |operations    | In this test case, the following operations are needed:      |
81 |              |  1. "nova-create-instance-in_network": create a VM instance  |
82 |              |     in one of the existing Neutron network.                  |
83 |              |                                                              |
84 +--------------+--------------------------------------------------------------+
85 |metrics       | In this test case, there are two metrics:                    |
86 |              |  1. process_recover_time: which indicates the maximun        |
87 |              |     time (seconds) from the process being killed to          |
88 |              |     recovered                                                |
89 |              |                                                              |
90 |              |  2. outage_time: measures the total time in which            |
91 |              |     monitors were failing in their tasks (e.g. total time of |
92 |              |     Ping failure)                                            |
93 |              |                                                              |
94 +--------------+--------------------------------------------------------------+
95 |test tool     | Developed by the project. Please see folder:                 |
96 |              | "yardstick/benchmark/scenarios/availability/ha_tools"        |
97 |              |                                                              |
98 +--------------+--------------------------------------------------------------+
99 |references    | none                                                         |
100 |              |                                                              |
101 +--------------+--------------------------------------------------------------+
102 |configuration | This test case needs two configuration files:                |
103 |              |  1. test case file: opnfv_yardstick_tc093.yaml               |
104 |              |                                                              |
105 |              |     - Attackers: see above “attackers” description           |
106 |              |     - monitor_time: which is the time (seconds) from         |
107 |              |       starting to stoping the monitors                       |
108 |              |     - Monitors: see above “monitors” discription             |
109 |              |     - SLA: see above “metrics” description                   |
110 |              |                                                              |
111 |              |  2. POD file: pod.yaml The POD configuration should record   |
112 |              |     on pod.yaml first. the “host” item in this test case     |
113 |              |     will use the node name in the pod.yaml.                  |
114 |              |                                                              |
115 +--------------+--------------------------------------------------------------+
116 |test sequence | Description and expected result                              |
117 |              |                                                              |
118 +--------------+--------------------------------------------------------------+
119 |pre-action    |  1. The Vswitches are set up in both compute nodes.          |
120 |              |                                                              |
121 |              |  2. One or more Neutron networks are created with two or     |
122 |              |     more VMs attached to each of the Neutron networks.       |
123 |              |                                                              |
124 |              |  3. The Neutron networks are attached to a Neutron router    |
125 |              |     which is attached to an external network towards the     |
126 |              |     DCGW.                                                    |
127 |              |                                                              |
128 +--------------+--------------------------------------------------------------+
129 |step 1        | Start IP connectivity monitors:                              |
130 |              |  1. Check the L2 connectivity between the VMs in the same    |
131 |              |     Neutron network.                                         |
132 |              |                                                              |
133 |              |  2. Check connectivity from one VM to an external host on    |
134 |              |     the Internet to verify SNAT functionality.               |
135 |              |                                                              |
136 |              | Result: The monitor info will be collected.                  |
137 |              |                                                              |
138 +--------------+--------------------------------------------------------------+
139 |step 2        | Start attackers:                                             |
140 |              | SSH connect to the VIM compute nodes and kill the Vswitch    |
141 |              | processes                                                    |
142 |              |                                                              |
143 |              | Result: the SDN Vswitch services will be shutdown            |
144 |              |                                                              |
145 +--------------+--------------------------------------------------------------+
146 |step 3        | Verify the results of the IP connectivity monitors.          |
147 |              |                                                              |
148 |              | Result: The outage_time metric reported by the monitors      |
149 |              | is not greater than the max_outage_time.                     |
150 |              |                                                              |
151 +--------------+--------------------------------------------------------------+
152 |step 4        | Restart the SDN Vswitch services.                            |
153 |              |                                                              |
154 +--------------+--------------------------------------------------------------+
155 |step 5        | Create a new VM in the existing Neutron network              |
156 |              |                                                              |
157 +--------------+--------------------------------------------------------------+
158 |step 6        | Verify connectivity between VMs as follows:                  |
159 |              |  1. Check the L2 connectivity between the previously         |
160 |              |     existing VM and the newly created VM on the same         |
161 |              |     Neutron network by sending ICMP messages                 |
162 |              |                                                              |
163 +--------------+--------------------------------------------------------------+
164 |step 7        | Stop IP connectivity monitors after a period of time         |
165 |              | specified by “monitor_time”                                  |
166 |              |                                                              |
167 |              | Result: The monitor info will be aggregated                  |
168 |              |                                                              |
169 +--------------+--------------------------------------------------------------+
170 |step 8        | Verify the IP connectivity monitor results                   |
171 |              |                                                              |
172 |              | Result: IP connectivity monitor should not have any packet   |
173 |              | drop failures reported                                       |
174 |              |                                                              |
175 +--------------+--------------------------------------------------------------+
176 |test verdict  | This test fails if the SLAs are not met or if there is a     |
177 |              | test case execution problem. The SLAs are define as follows  |
178 |              | for this test:                                               |
179 |              | * SDN Vswitch recovery                                       |
180 |              |                                                              |
181 |              |   * process_recover_time <= 30 sec                           |
182 |              |                                                              |
183 |              | * no impact on data plane connectivity during SDN            |
184 |              |   Vswitch failure and recovery.                              |
185 |              |                                                              |
186 |              |   * packet_drop == 0                                         |
187 |              |                                                              |
188 +--------------+--------------------------------------------------------------+
189