suite:
- 'posca_stress_traffic'
- 'posca_stress_ping'
+ - 'posca_factor_multistack_storage'
+ - 'posca_factor_multistack_storage_parallel'
+ - 'posca_feature_moon_resources'
+ - 'posca_feature_moon_tenants'
+ - 'posca_feature_vnf_scale_out'
+ - 'posca_factor_soak_throughputs'
jobs:
- 'bottlenecks-{installer}-{suite}-{pod}-daily-{stream}'
# Running test cases through Bottlenecks docker
if [[ $SUITE_NAME == posca_stress_traffic ]]; then
TEST_CASE=posca_factor_system_bandwidth
- testcase_cmd="docker exec bottlenecks-load-master python ${POSCA_SCRIPT}/../run_testsuite.py testcase $TEST_CASE $REPORT"
- echo "BOTTLENECKS INFO: running test case ${TEST_CASE} with report indicator: ${testcase_cmd}"
- ${testcase_cmd} >$redirect
elif [[ $SUITE_NAME == posca_stress_ping ]]; then
TEST_CASE=posca_factor_ping
- testcase_cmd="docker exec bottlenecks-load-master python ${POSCA_SCRIPT}/../run_testsuite.py testcase $TEST_CASE $REPORT"
- echo "BOTTLENECKS INFO: running test case ${TEST_CASE} with report indicator: ${testcase_cmd}"
- ${testcase_cmd} >$redirect
+ else
+ TEST_CASE=$SUITE_NAME
fi
+ testcase_cmd="docker exec bottlenecks-load-master python ${POSCA_SCRIPT}/../run_testsuite.py testcase $TEST_CASE $REPORT"
+ echo "BOTTLENECKS INFO: running test case ${TEST_CASE} with report indicator: ${testcase_cmd}"
+ ${testcase_cmd} >$redirect
fi
build-step-failure-threshold: 'never'
failure-threshold: 'never'
unstable-threshold: 'FAILURE'
+ - project: 'bottlenecks-compass-posca_factor_multistack_storage_parallel-{pod}-daily-{stream}'
+ current-parameters: false
+ predefined-parameters:
+ DEPLOY_SCENARIO={scenario}
+ block: true
+ same-node: true
+ block-thresholds:
+ build-step-failure-threshold: 'never'
+ failure-threshold: 'never'
+ unstable-threshold: 'FAILURE'
+ - project: 'bottlenecks-compass-posca_factor_soak_throughputs-{pod}-daily-{stream}'
+ current-parameters: false
+ predefined-parameters:
+ DEPLOY_SCENARIO={scenario}
+ block: true
+ same-node: true
+ block-thresholds:
+ build-step-failure-threshold: 'never'
+ failure-threshold: 'never'
+ unstable-threshold: 'FAILURE'
- conditional-step:
condition-kind: and
condition-operands:
use-build-blocker: true
blocking-jobs:
- 'daisy-kolla-build-.*'
- - '{installer}-daily-(build|deploy|test)-(euphrates|master)'
- - '{installer}-.*-(baremetal|virtual|zte-pod9)-daily-(euphrates|master)'
+ - '{installer}-(build|deploy|test)-daily-(euphrates|master)'
+ - '{installer}-deploy-(baremetal|virtual|zte-pod9)-daily-(euphrates|master)'
- '(functest|yardstick)-{installer}-(baremetal|virtual|zte-pod9)-daily-(euphrates|master)'
block-level: 'NODE'
DOVETAIL_CONFIG=${DOVETAIL_HOME}/pre_config
mkdir -p ${DOVETAIL_CONFIG}
+DOVETAIL_IMAGES=${DOVETAIL_HOME}/images
+mkdir -p ${DOVETAIL_IMAGES}
+
ssh_options="-o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no"
sshkey=""
echo "Download image ubuntu-16.04-server-cloudimg-amd64-disk1.img ..."
wget -q -nc http://artifacts.opnfv.org/sdnvpn/ubuntu-16.04-server-cloudimg-amd64-disk1.img -P ${image_path}
fi
-sudo cp ${ubuntu_image} ${DOVETAIL_CONFIG}
+sudo cp ${ubuntu_image} ${DOVETAIL_IMAGES}
# functest needs to download this image first before running
cirros_image=${image_path}/cirros-0.3.5-x86_64-disk.img
echo "Download image cirros-0.3.5-x86_64-disk.img ..."
wget -q -nc http://download.cirros-cloud.net/0.3.5/cirros-0.3.5-x86_64-disk.img -P ${image_path}
fi
-sudo cp ${cirros_image} ${DOVETAIL_CONFIG}
+sudo cp ${cirros_image} ${DOVETAIL_IMAGES}
# snaps_smoke test case needs to download this image first before running
ubuntu14_image=${image_path}/ubuntu-14.04-server-cloudimg-amd64-disk1.img
echo "Download image ubuntu-14.04-server-cloudimg-amd64-disk1.img ..."
wget -q -nc https://cloud-images.ubuntu.com/releases/14.04/release/ubuntu-14.04-server-cloudimg-amd64-disk1.img -P ${image_path}
fi
-sudo cp ${ubuntu14_image} ${DOVETAIL_CONFIG}
+sudo cp ${ubuntu14_image} ${DOVETAIL_IMAGES}
# cloudify_ims test case needs to download these 2 images first before running
cloudify_image=${image_path}/cloudify-manager-premium-4.0.1.qcow2
echo "Download image cloudify-manager-premium-4.0.1.qcow2 ..."
wget -q -nc http://repository.cloudifysource.org/cloudify/4.0.1/sp-release/cloudify-manager-premium-4.0.1.qcow2 -P ${image_path}
fi
-sudo cp ${cloudify_image} ${DOVETAIL_CONFIG}
+sudo cp ${cloudify_image} ${DOVETAIL_IMAGES}
trusty_image=${image_path}/trusty-server-cloudimg-amd64-disk1.img
if [[ ! -f ${trusty_image} ]]; then
echo "Download image trusty-server-cloudimg-amd64-disk1.img ..."
wget -q -nc http://cloud-images.ubuntu.com/trusty/current/trusty-server-cloudimg-amd64-disk1.img -P ${image_path}
fi
-sudo cp ${trusty_image} ${DOVETAIL_CONFIG}
+sudo cp ${trusty_image} ${DOVETAIL_IMAGES}
opts="--privileged=true -id"
echo "Dovetail: Pulling image ${DOCKER_REPO}:${DOCKER_TAG}"
docker pull ${DOCKER_REPO}:$DOCKER_TAG >$redirect
-env4bgpvpn="-e INSTALLER_TYPE=${INSTALLER_TYPE} -e INSTALLER_IP=${INSTALLER_IP}"
-
cmd="docker run ${opts} -e DOVETAIL_HOME=${DOVETAIL_HOME} ${docker_volume} ${dovetail_home_volume} \
- ${sshkey} ${env4bgpvpn} ${DOCKER_REPO}:${DOCKER_TAG} /bin/bash"
+ ${sshkey} ${DOCKER_REPO}:${DOCKER_TAG} /bin/bash"
echo "Dovetail: running docker run command: ${cmd}"
${cmd} >${redirect}
sleep 5
set +e
-
-if [[ ${DEPLOY_SCENARIO} =~ ^os-.* ]]; then
- [[ ${BRANCH##*/} == "master" ]] && check_os_deployment
- if [ ${FUNCTEST_MODE} == 'testcase' ]; then
- echo "FUNCTEST_MODE=testcase, FUNCTEST_SUITE_NAME=${FUNCTEST_SUITE_NAME}"
- run_test ${FUNCTEST_SUITE_NAME}
- elif [ ${FUNCTEST_MODE} == 'tier' ]; then
- echo "FUNCTEST_MODE=tier, FUNCTEST_TIER=${FUNCTEST_TIER}"
- tiers=(${FUNCTEST_TIER})
- run_tiers ${tiers}
+[[ ${BRANCH##*/} == "master" ]] && check_os_deployment
+if [ ${FUNCTEST_MODE} == 'testcase' ]; then
+ echo "FUNCTEST_MODE=testcase, FUNCTEST_SUITE_NAME=${FUNCTEST_SUITE_NAME}"
+ run_test ${FUNCTEST_SUITE_NAME}
+elif [ ${FUNCTEST_MODE} == 'tier' ]; then
+ echo "FUNCTEST_MODE=tier, FUNCTEST_TIER=${FUNCTEST_TIER}"
+ tiers=(${FUNCTEST_TIER})
+ run_tiers ${tiers}
+else
+ if [ ${DEPLOY_TYPE} == 'baremetal' ]; then
+ tiers=(healthcheck smoke features vnf parser)
else
- if [ ${DEPLOY_TYPE} == 'baremetal' ]; then
- tiers=(healthcheck smoke features vnf parser)
- else
- tiers=(healthcheck smoke features)
- fi
- run_tiers ${tiers}
+ tiers=(healthcheck smoke features)
fi
-else
- echo "k8 deployment has not been supported by functest yet"
+ run_tiers ${tiers}
fi
DOCKER_TAG=`[[ ${BRANCH##*/} == "master" ]] && echo "latest" || echo ${BRANCH##*/}`
+set +e
+
FUNCTEST_IMAGE=opnfv/functest-kubernetes:${DOCKER_TAG}
echo "Pulling Docker image ${FUNCTEST_IMAGE} ..."
docker pull ${FUNCTEST_IMAGE}>/dev/null
exit 0
fi
-ssh -F $HOME/.ssh/${DISTRO}-xci-vm-config ${DISTRO}_xci_vm_opnfv "cd /root/releng-xci/xci/playbooks && ansible-playbook -i inventory prepare-functest.yml"
+ssh -F $HOME/.ssh/${DISTRO}-xci-vm-config ${DISTRO}_xci_vm "cd /root/releng-xci/xci && PATH=/home/devuser/.local/bin:$PATH ansible-playbook -i installer/osa/files/$XCI_FLAVOR/inventory playbooks/prepare-functest.yml"
echo "Running functest"
ssh -F $HOME/.ssh/${DISTRO}-xci-vm-config ${DISTRO}_xci_vm_opnfv "/root/run-functest.sh"
echo "Functest log"
--- /dev/null
+#!/usr/bin/env python
+
+# Copyright (c) 2018 HUAWEI TECHNOLOGIES CO.,LTD and others.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+
+from opnfv.deployment import manager
+from opnfv.utils import opnfv_logger as logger
+from opnfv.utils import ssh_utils
+
+import yaml
+import os
+
+logger = logger.Logger(__name__).getLogger()
+
+
+class ContainerizedCompassAdapter():
+
+ def __init__(self, installer_ip, installer_user, pkey_file):
+
+ self.installer = 'compass'
+ self.installer_ip = installer_ip
+ self.installer_user = installer_user
+ self.pkey_file = pkey_file
+ self.DST_PATH_UC = "/tmp/openstack_user_config.yml"
+ self.nodes = []
+ self.ROLES = {}
+
+ if pkey_file is not None and not os.path.isfile(pkey_file):
+ raise Exception(
+ 'The private key file %s does not exist!' % pkey_file)
+
+ def _find_nodes(self, file):
+ nodes = file['compute_hosts']
+ for compute in nodes:
+ self.ROLES[compute] = 'compute'
+ controllers = file['haproxy_hosts']
+ for controller in controllers:
+ nodes[controller] = controllers[controller]
+ self.ROLES[controller] = 'controller'
+ return nodes
+
+ def _process_nodes(self, raw_nodes):
+ nodes = []
+
+ for node in raw_nodes:
+ name = node
+ ip = raw_nodes[node]['ip']
+ status = 'active'
+ id = None
+ if self.ROLES[node] == 'controller':
+ roles = 'controller'
+ elif self.ROLES[node] == 'compute':
+ roles = 'compute'
+ ssh_client = ssh_utils.get_ssh_client(hostname=ip,
+ username=self.installer_user,
+ pkey_file=self.pkey_file)
+ node = manager.Node(id, ip, name, status, roles, ssh_client)
+ nodes.append(node)
+
+ return nodes
+
+ def get_nodes(self, options=None):
+ try:
+ # if we have retrieved previously all the nodes, don't do it again
+ # This fails the first time when the constructor calls this method
+ # therefore the try/except
+ if len(self.nodes) > 0:
+ return self.nodes
+ except:
+ pass
+
+ with open(self.DST_PATH_UC, 'r') as stream:
+ try:
+ file = yaml.load(stream)
+ raw_nodes = self._find_nodes(file)
+ except yaml.YAMLError as exc:
+ logger.error(exc)
+ self.nodes = self._process_nodes(raw_nodes)
+ return self.nodes
from opnfv.deployment.apex import adapter as apex_adapter
-from opnfv.deployment.compass import adapter as compass_adapter
+from opnfv.deployment.compass import adapter_container as compass_adapter
from opnfv.deployment.fuel import adapter as fuel_adapter
from opnfv.deployment.osa import adapter as osa_adapter
from opnfv.deployment.daisy import adapter as daisy_adapter
installer_user=installer_user,
installer_pwd=installer_pwd)
elif installer.lower() == "compass":
- return compass_adapter.CompassAdapter(
+ return compass_adapter.ContainerizedCompassAdapter(
installer_ip=installer_ip,
installer_user=installer_user,
- installer_pwd=installer_pwd)
+ pkey_file=pkey_file)
elif installer.lower() == "osa":
return osa_adapter.OSAAdapter(installer_ip=installer_ip,
installer_user=installer_user,