- 'posca_feature_moon_tenants'
- 'posca_feature_vnf_scale_out'
- 'posca_factor_soak_throughputs'
+ - 'kubestone_deployment_capacity'
jobs:
- 'bottlenecks-{installer}-{suite}-{pod}-daily-{stream}'
OS_CACERT=/tmp/os_cacert
BOTTLENECKS_CONFIG=/tmp
-
+KUBESTONE_TEST_DIR=/home/opnfv/bottlenecks/testsuites/kubestone/testcases
+
+# Pulling Bottlenecks docker and passing environment variables
+echo "INFO: pulling Bottlenecks docker ${DOCKER_TAG}"
+docker pull opnfv/bottlenecks:${DOCKER_TAG} >$redirect
+
+opts="--privileged=true -id"
+envs="-e INSTALLER_TYPE=${INSTALLER_TYPE} -e INSTALLER_IP=${INSTALLER_IP} \
+ -e NODE_NAME=${NODE_NAME} -e EXTERNAL_NET=${EXTERNAL_NETWORK} \
+ -e BRANCH=${BRANCH} -e GERRIT_REFSPEC_DEBUG=${GERRIT_REFSPEC_DEBUG} \
+ -e BOTTLENECKS_DB_TARGET=${BOTTLENECKS_DB_TARGET} -e PACKAGE_URL=${PACKAGE_URL} \
+ -e DEPLOY_SCENARIO=${DEPLOY_SCENARIO} -e BUILD_TAG=${BUILD_TAG}"
+docker_volume="-v /var/run/docker.sock:/var/run/docker.sock -v /tmp:/tmp"
+
+cmd="docker run ${opts} ${envs} --name bottlenecks-load-master ${docker_volume} opnfv/bottlenecks:${DOCKER_TAG} /bin/bash"
+echo "BOTTLENECKS INFO: running docker run commond: ${cmd}"
+${cmd} >$redirect
+sleep 5
+
+# Run test suite
if [[ $SUITE_NAME == *posca* ]]; then
POSCA_SCRIPT=/home/opnfv/bottlenecks/testsuites/posca
sudo rm -f ${OPENRC}
sudo ls -al ${BOTTLENECKS_CONFIG}
fi
- # Pulling Bottlenecks docker and passing environment variables
- echo "INFO: pulling Bottlenecks docker ${DOCKER_TAG}"
- docker pull opnfv/bottlenecks:${DOCKER_TAG} >$redirect
-
- opts="--privileged=true -id"
- envs="-e INSTALLER_TYPE=${INSTALLER_TYPE} -e INSTALLER_IP=${INSTALLER_IP} \
- -e NODE_NAME=${NODE_NAME} -e EXTERNAL_NET=${EXTERNAL_NETWORK} \
- -e BRANCH=${BRANCH} -e GERRIT_REFSPEC_DEBUG=${GERRIT_REFSPEC_DEBUG} \
- -e BOTTLENECKS_DB_TARGET=${BOTTLENECKS_DB_TARGET} -e PACKAGE_URL=${PACKAGE_URL} \
- -e DEPLOY_SCENARIO=${DEPLOY_SCENARIO} -e BUILD_TAG=${BUILD_TAG}"
- docker_volume="-v /var/run/docker.sock:/var/run/docker.sock -v /tmp:/tmp"
-
- cmd="docker run ${opts} ${envs} --name bottlenecks-load-master ${docker_volume} opnfv/bottlenecks:${DOCKER_TAG} /bin/bash"
- echo "BOTTLENECKS INFO: running docker run commond: ${cmd}"
- ${cmd} >$redirect
- sleep 5
-
# Running test cases through Bottlenecks docker
if [[ $SUITE_NAME == posca_stress_traffic ]]; then
TEST_CASE=posca_factor_system_bandwidth
testcase_cmd="docker exec bottlenecks-load-master python ${POSCA_SCRIPT}/../run_testsuite.py testcase $TEST_CASE $REPORT"
echo "BOTTLENECKS INFO: running test case ${TEST_CASE} with report indicator: ${testcase_cmd}"
${testcase_cmd} >$redirect
+elif [[ $SUITE_NAME == *kubestone* ]]; then
+ if [[ $SUITE_NAME == kubestone_deployment_capacity ]]; then
+ TEST_CASE=${KUBESTONE_TEST_DIR}/deployment_capacity.yaml
+ fi
+ testcase_cmd="docker exec bottlenecks-load-master python ${KUBESTONE_TEST_DIR}/../stress_test.py -c $TEST_CASE"
+ echo "BOTTLENECKS INFO: running test case ${TEST_CASE} with report indicator: ${testcase_cmd}"
+ ${testcase_cmd} >$redirect
fi
build-step-failure-threshold: 'never'
failure-threshold: 'never'
unstable-threshold: 'FAILURE'
+ - conditional-step:
+ condition-kind: and
+ condition-operands:
+ - condition-kind: regex-match
+ regex: k8-nosdn-nofeature-ha
+ label: '{scenario}'
+ steps:
+ - trigger-builds:
+ - project: 'bottlenecks-compass-kubestone_deployment_capacity-{pod}-daily-{stream}'
+ current-parameters: false
+ predefined-parameters:
+ DEPLOY_SCENARIO={scenario}
+ block: true
+ same-node: true
+ block-thresholds:
+ build-step-failure-threshold: 'never'
+ failure-threshold: 'never'
+ unstable-threshold: 'FAILURE'
- conditional-step:
condition-kind: and
condition-operands:
description: "Directory where the cache to be used during the build is located."
- string:
name: PPA_REPO
- default: "http://artifacts.opnfv.org/compass4nfv/package{ppa-pathname}"
+ default: "https://artifacts.opnfv.org/compass4nfv/package{ppa-pathname}"
- string:
name: PPA_CACHE
default: "$WORKSPACE/work/repo/"
description: "URL to Google Storage."
- string:
name: PPA_REPO
- default: "http://artifacts.opnfv.org/compass4nfv/package{ppa-pathname}"
+ default: "https://artifacts.opnfv.org/compass4nfv/package{ppa-pathname}"
- string:
name: PPA_CACHE
default: "$WORKSPACE/work/repo/"
description: "URL to Google Storage."
- string:
name: PPA_REPO
- default: "http://artifacts.opnfv.org/compass4nfv/package{ppa-pathname}"
+ default: "https://artifacts.opnfv.org/compass4nfv/package{ppa-pathname}"
- string:
name: PPA_CACHE
default: "$WORKSPACE/work/repo/"
sudo docker rm -f `docker ps | grep compass | cut -f1 -d' '` || true
-curl -s http://people.linaro.org/~yibo.cai/compass/compass4nfv-arm64-fixup.sh | bash -s {scenario} || true
+curl -s http://people.linaro.org/~yibo.cai/compass/compass4nfv-arm64-fixup.sh | bash || true
./build.sh
- job-template:
name: 'container4nfv-{scenario}-{pod}-daily-{stream}'
disabled: '{obj:disabled}'
+ node: '{slave-label}'
scm:
- git:
url: https://gerrit.opnfv.org/gerrit/container4nfv
+ branches:
+ - master
basedir: container4nfv
wipe-workspace: true
GERRIT_CHANGE_COMMIT_MESSAGE=$GERRIT_CHANGE_COMMIT_MESSAGE
kill-phase-on: FAILURE
git-revision: true
- - name: 'doctor-verify-daisy-congress-x86_64-{stream}'
- predefined-parameters: |
- PROJECT=$PROJECT
- GERRIT_BRANCH=$GERRIT_BRANCH
- GERRIT_REFSPEC=$GERRIT_REFSPEC
- GERRIT_CHANGE_NUMBER=$GERRIT_CHANGE_NUMBER
- GERRIT_CHANGE_COMMIT_MESSAGE=$GERRIT_CHANGE_COMMIT_MESSAGE
- kill-phase-on: FAILURE
- git-revision: true
- job-template:
name: 'doctor-verify-{installer}-{inspector}-{arch}-{stream}'
ubuntu_image=${image_path}/ubuntu-16.04-server-cloudimg-amd64-disk1.img
if [[ ! -f ${ubuntu_image} ]]; then
echo "Download image ubuntu-16.04-server-cloudimg-amd64-disk1.img ..."
- wget -q -nc http://artifacts.opnfv.org/sdnvpn/ubuntu-16.04-server-cloudimg-amd64-disk1.img -P ${image_path}
+ wget -q -nc https://artifacts.opnfv.org/sdnvpn/ubuntu-16.04-server-cloudimg-amd64-disk1.img -P ${image_path}
fi
sudo cp ${ubuntu_image} ${DOVETAIL_IMAGES}
testarea="--testarea ${TESTAREA}"
fi
-run_cmd="dovetail run ${testsuite} ${testarea} -d -r"
+run_cmd="dovetail run ${testsuite} ${testarea} --deploy-scenario ${DEPLOY_SCENARIO} -d -r"
echo "Container exec command: ${run_cmd}"
docker exec $container_id ${run_cmd}
#remove useless files to save disk space
sudo rm -rf ./results/workspace
sudo rm -f ./results/yardstick.img
-sudo rm -f ./results/tmp*
+sudo rm -f ./results/bottlenecks/tmp*
echo "Dovetail: done!"
- trigger:
name: 'fuel-os-nosdn-nofeature-ha-baremetal-daily-fraser-trigger'
triggers:
- - timed: '' # '0 20 * * *'
+ - timed: '0 20 * * 1,3,7'
- trigger:
name: 'fuel-os-odl-nofeature-ha-baremetal-daily-fraser-trigger'
triggers:
gs-pathname: '/{stream}'
disabled: false
+ project-name: '{project}'
+ project-pattern: '{project}'
+ rtd-build-url: https://readthedocs.org/api/v2/webhook/opnfv-fuel/47205/
+ rtd-token: 38f40bf6c08fd4bccb930871bc29b08404cf98b0
+
jobs:
- 'fuel-deploy-generic-daily-{stream}'
+ - '{project-name}-rtd-jobs'
########################
# job templates
case ${test_name} in
connection_check|tenantnetwork1|tenantnetwork2|vmready1|vmready2|singlevm1|singlevm2|vping_ssh|vping_userdata|cinder_test|odl|api_check|snaps_health_check)
FUNCTEST_IMAGE=${REPO}/functest-healthcheck:${DOCKER_TAG} ;;
- tempest_smoke|rally_sanity|refstack_defcore|patrole|shaker|snaps_smoke|neutron_trunk|barbican)
+ tempest_smoke|neutron-tempest-plugin-api|rally_sanity|refstack_defcore|patrole|snaps_smoke|neutron_trunk|networking-bgpvpn|networking-sfc|barbican)
FUNCTEST_IMAGE=${REPO}/functest-smoke:${DOCKER_TAG} ;;
- tempest_full|rally_full)
+ shaker|vmtp)
+ FUNCTEST_IMAGE=${REPO}/functest-benchmarking:${DOCKER_TAG} ;;
+ tempest_full|tempest_scenario|rally_full)
FUNCTEST_IMAGE=${REPO}/functest-components:${DOCKER_TAG} ;;
- cloudify_ims|vyos_vrouter|juju_epc)
+ cloudify|cloudify_ims|heat_ims|vyos_vrouter|juju_epc)
FUNCTEST_IMAGE=${REPO}/functest-vnf:${DOCKER_TAG} ;;
- doctor-notification|bgpvpn|functest-odl-sfc|barometercollectd|fds)
+ doctor-notification|bgpvpn|functest-odl-sfc|barometercollectd|fds|vgpu|stor4nfv_os)
FUNCTEST_IMAGE=${REPO}/functest-features:${DOCKER_TAG} ;;
*)
echo "Unkown test case $test_name"
tiers=(${FUNCTEST_TIER})
run_tiers ${tiers}
else
+ tests=(tempest_full tempest_scenario)
if [ ${DEPLOY_TYPE} == 'baremetal' ] && [ "${HOST_ARCH}" != "aarch64" ]; then
if [[ ${BRANCH} == "stable/fraser" ]]; then
tiers=(healthcheck smoke features vnf parser)
+ tests=(tempest_full_parallel)
else
- tiers=(healthcheck smoke features vnf)
+ tiers=(healthcheck smoke benchmarking features vnf)
fi
else
if [[ ${BRANCH} == "stable/fraser" ]]; then
tiers=(healthcheck smoke features parser)
+ tests=(tempest_full_parallel)
else
- tiers=(healthcheck smoke features)
+ tiers=(healthcheck smoke benchmarking features)
fi
fi
run_tiers ${tiers}
+ for test in "${tests[@]}"; do
+ run_test "$test"
+ done
fi
- 'patrole'
- 'odl'
- 'snaps_smoke'
+ - 'shaker'
+ - 'vmtp'
- 'neutron_trunk'
- 'doctor-notification'
- 'bgpvpn'
choices:
- 'healthcheck'
- 'smoke'
+ - 'benchmarking'
- 'features'
- 'components'
- 'vnf'
- 'components'
- 'parser'
- 'smoke'
+ - 'benchmarking'
- 'vnf'
exclude:
image: 'parser'
- stream: 'fraser'
image: 'tempest'
+ - stream: 'fraser'
+ image: 'benchmarking'
# settings for jobs run in multijob phases
build-job-settings: &build-job-settings
<<: *build-job-settings
- name: 'functest-smoke-docker-build-arm64-{stream}'
<<: *build-job-settings
+ - name: 'functest-benchmarking-docker-build-amd64-{stream}'
+ <<: *build-job-settings
+ - name: 'functest-benchmarking-docker-build-arm64-{stream}'
+ <<: *build-job-settings
- name: 'functest-vnf-docker-build-amd64-{stream}'
<<: *build-job-settings
- name: 'functest-vnf-docker-build-arm64-{stream}'
<<: *manifest-job-settings
- name: 'functest-smoke-docker-manifest-{stream}'
<<: *manifest-job-settings
+ - name: 'functest-benchmarking-docker-manifest-{stream}'
+ <<: *manifest-job-settings
- name: 'functest-vnf-docker-manifest-{stream}'
<<: *manifest-job-settings
node: master
+ # Defaults for global-jjb jobs
+ build-timeout: 60
+ build-node: 'opnfv-build'
+ gerrit-server-name: 'gerrit.opnfv.org'
+ jenkins-ssh-credential: 'd42411ac011ad6f3dd2e1fa34eaa5d87f910eb2e'
+ jjb-version: '2.0.3'
+ lftools-version: '<1.0.0'
+
properties:
- logrotate-default
dockerfile: 'Dockerfile'
<<: *master
<<: *other-receivers
+ - 'clover-sample-app':
+ project: 'clover'
+ dockerdir: 'samples/scenarios/sample_app'
+ dockerfile: 'Dockerfile'
+ <<: *master
+ <<: *other-receivers
- 'cperf':
project: 'cperf'
<<: *master
default-slaves:
- lf-build1
- lf-build2
+ - ericsson-build3
- ericsson-build4
allowed-multiselect: true
ignore-offline-nodes: true
- timed: '@midnight'
builders:
+ - description-setter:
+ description: '$NODE_NAME'
- install-ansible
- run-ansible-build-server-playbook
project-name: 'releng'
build-timeout: 60
- build-node: 'lf-build1'
- gerrit-server-name: 'gerrit.opnfv.org'
- jenkins-ssh-credential: 'd42411ac011ad6f3dd2e1fa34eaa5d87f910eb2e'
- jjb-version: '2.0.3'
- lftools-version: '<1.0.0'
rtd-build-url: 'https://readthedocs.org/api/v2/webhook/opnfv-releng/38594/'
rtd-token: '291c6a0109493b4457e566d06141212452c65784'
deploy-scenario: 'os-nosdn-nofeature'
installer-type: 'osa'
xci-flavor: 'mini'
+ - 'os-nosdn-osm-noha':
+ auto-trigger-name: 'daily-trigger-disabled'
+ deploy-scenario: 'os-nosdn-osm'
+ installer-type: 'osa'
+ xci-flavor: 'mini'
+ - 'os-odl-nofeature-noha':
+ auto-trigger-name: 'daily-trigger-disabled'
+ deploy-scenario: 'os-odl-nofeature'
+ installer-type: 'osa'
+ xci-flavor: 'mini'
+ - 'os-odl-bgpvpn-noha':
+ auto-trigger-name: 'daily-trigger-disabled'
+ deploy-scenario: 'os-odl-bgpvpn'
+ installer-type: 'osa'
+ xci-flavor: 'mini'
- 'os-odl-sfc-noha':
auto-trigger-name: 'daily-trigger-disabled'
deploy-scenario: 'os-odl-sfc'
installer-type: 'osa'
xci-flavor: 'mini'
+ - 'k8-nosdn-nofeature-noha':
+ auto-trigger-name: 'daily-trigger-disabled'
+ deploy-scenario: 'k8-nosdn-nofeature'
+ installer-type: 'kubespray'
+ xci-flavor: 'mini'
+ - 'k8-canal-nofeature-noha':
+ auto-trigger-name: 'daily-trigger-disabled'
+ deploy-scenario: 'k8-canal-nofeature'
+ installer-type: 'kubespray'
+ xci-flavor: 'mini'
+ - 'k8-calico-nofeature-noha':
+ auto-trigger-name: 'daily-trigger-disabled'
+ deploy-scenario: 'k8-calico-nofeature'
+ installer-type: 'kubespray'
+ xci-flavor: 'mini'
+ - 'k8-contiv-nofeature-noha':
+ auto-trigger-name: 'daily-trigger-disabled'
+ deploy-scenario: 'k8-contiv-nofeature'
+ installer-type: 'kubespray'
+ xci-flavor: 'mini'
+ - 'k8-flannel-nofeature-noha':
+ auto-trigger-name: 'daily-trigger-disabled'
+ deploy-scenario: 'k8-flannel-nofeature'
+ installer-type: 'kubespray'
+ xci-flavor: 'mini'
+ - 'k8-nosdn-istio-noha':
+ auto-trigger-name: 'daily-trigger-disabled'
+ deploy-scenario: 'k8-nosdn-istio'
+ installer-type: 'kubespray'
+ xci-flavor: 'mini'
# -------------------------------
# XCI PODs
phase:
- 'deploy'
- 'functest'
+ - 'yardstick'
# -------------------------------
# jobs
- '^xci-verify.*'
- '^xci-merge.*'
- '^xci-os.*'
+ - '^xci-k8.*'
- '^xci-deploy.*'
- '^xci-functest.*'
- '^bifrost-.*periodic.*'
default: 'tier'
- string:
name: FUNCTEST_SUITE_NAME
- default: 'healthcheck'
+ default: 'smoke'
- string:
name: CI_LOOP
default: 'daily'
XCI_FLAVOR=$XCI_FLAVOR
CI_LOOP=$CI_LOOP
XCI_DISTRO=$XCI_DISTRO
+ FUNCTEST_MODE=$FUNCTEST_MODE
+ FUNCTEST_SUITE_NAME=$FUNCTEST_SUITE_NAME
same-node: true
block: true
- trigger-builds:
build-step-failure-threshold: 'never'
failure-threshold: 'never'
unstable-threshold: 'FAILURE'
+ - trigger-builds:
+ - project: 'xci-yardstick-{pod}-{distro}-daily-{stream}'
+ current-parameters: false
+ predefined-parameters: |
+ DEPLOY_SCENARIO=$DEPLOY_SCENARIO
+ INSTALLER_TYPE=$INSTALLER_TYPE
+ XCI_FLAVOR=$XCI_FLAVOR
+ CI_LOOP=$CI_LOOP
+ XCI_DISTRO=$XCI_DISTRO
+ same-node: true
+ block: true
+ block-thresholds:
+ build-step-failure-threshold: 'never'
+ failure-threshold: 'never'
+ unstable-threshold: 'FAILURE'
publishers:
# yamllint disable rule:line-length
default: 'tier'
- string:
name: FUNCTEST_SUITE_NAME
- default: 'healthcheck'
+ default: 'smoke'
- string:
name: CI_LOOP
default: 'daily'
# --------------------------
# builder macros
# --------------------------
-# These need to be properly fixed once the basic deployment and functest
-# jobs are working outside of clean vm.
-# One of the ugly fixes is moving functest preparation step into the
-# deployment job itself since functest preparation requires some of the
+# These need to be properly fixed once the basic deployment, functest, and
+# yardstick jobs are working outside of clean vm.
+# One of the ugly fixes is moving test preparation step into the
+# deployment job itself since test preparation requires some of the
# things from deployment job. With clean VM, this wasn't an issue
# since everything was on clean VM. When we move things out of clean
# VM, things are done in workspaces of the jobs that are different.
set -o errexit
set -o pipefail
+ echo "Cleaning the leftovers from the earlier run"
+ echo "---------------------------------------------------------------------------------"
+ for vm in $(sudo virsh list --all --name | grep -v xci_vm); do
+ echo "Removing $vm"
+ sudo virsh destroy $vm > /dev/null 2>&1 || true
+ sudo virsh undefine $vm > /dev/null 2>&1 || true
+ sudo killall -r vbmc > /dev/null 2>&1 || true
+ sudo rm -rf /root/.vbmc > /dev/null 2>&1 || true
+ done
+ echo "---------------------------------------------------------------------------------"
+
cd $WORKSPACE/xci && ./xci-deploy.sh
- echo "Prepare OPNFV VM for Functest"
+ echo "Prepare OPNFV VM for Tests"
echo "---------------------------------------------------------------------------------"
export XCI_PATH=$WORKSPACE
export XCI_VENV=${XCI_PATH}/venv
declare -x "\${var}" 2>/dev/null
echo $var
done < ${XCI_PATH}/.cache/xci.env && cd ${XCI_PATH}/xci && \
- ansible-playbook -i playbooks/dynamic_inventory.py playbooks/prepare-functest.yml
-
- echo "Run Functest"
+ ansible-playbook -i playbooks/dynamic_inventory.py playbooks/prepare-tests.yml
+ ssh root@192.168.122.2 "/root/prepare-tests.sh"
echo "---------------------------------------------------------------------------------"
- builder:
builders:
- shell: |
#!/bin/bash
- set -o errexit
set -o pipefail
ssh root@192.168.122.2 "/root/run-functest.sh"
+ functest_exit=$?
+
+ case ${DEPLOY_SCENARIO[0]} in
+ os-*)
+ FUNCTEST_LOG=/root/functest-results/functest.log
+ ;;
+ k8-*)
+ FUNCTEST_LOG=/root/functest-results/functest-kubernetes.log
+ ;;
+ *)
+ echo "Unable to determine the installer. Exiting!"
+ exit $functest_exit
+ ;;
+ esac
+
+ echo "Functest log"
+ echo "---------------------------------------------------------------------------------"
+ ssh root@192.168.122.2 "cat $FUNCTEST_LOG"
+ echo "---------------------------------------------------------------------------------"
+ exit ${functest_exit}
+- builder:
+ name: 'xci-daily-yardstick-macro'
+ builders:
+ - shell: |
+ #!/bin/bash
+ set -o errexit
+ set -o pipefail
+
+ ssh root@192.168.122.2 "/root/run-yardstick.sh"
- 'xci-verify-{distro}-.*'
- 'xci-merge-{distro}-.*'
- 'xci-os.*'
+ - 'xci-k8.*'
- 'openstack-bifrost-verify-{distro}-.*'
- 'xci-osa-verify-{distro}-.*'
- 'xci-osa-periodic-{distro}-.*'
GERRIT_REFSPEC=$GERRIT_REFSPEC
GERRIT_CHANGE_NUMBER=$GERRIT_CHANGE_NUMBER
GERRIT_CHANGE_COMMIT_MESSAGE=$GERRIT_CHANGE_COMMIT_MESSAGE
+ FUNCTEST_MODE=$FUNCTEST_MODE
+ FUNCTEST_SUITE_NAME=$FUNCTEST_SUITE_NAME
node-parameters: true
kill-phase-on: '{kill-phase-on}'
abort-all-job: '{abort-all-job}'
export XCI_PATH=/home/devuser/releng-xci
export XCI_VENV=${XCI_PATH}/venv
-ssh -F $HOME/.ssh/${DISTRO}-xci-vm-config ${DISTRO}_xci_vm "source $XCI_VENV/bin/activate; while read var; do declare -x \"\${var}\" 2>/dev/null; done < ${XCI_PATH}/.cache/xci.env && cd releng-xci/xci && ansible-playbook -i playbooks/dynamic_inventory.py playbooks/prepare-functest.yml"
-echo "Running functest"
+ssh -F $HOME/.ssh/${DISTRO}-xci-vm-config ${DISTRO}_xci_vm "source $XCI_VENV/bin/activate; \
+ while read var; do declare -x \"\${var}\" 2>/dev/null; done < ${XCI_PATH}/.cache/xci.env && \
+ cd releng-xci/xci && ansible-playbook -i playbooks/dynamic_inventory.py playbooks/prepare-tests.yml"
+echo "Prepare OPNFV VM for Tests"
+ssh -F $HOME/.ssh/${DISTRO}-xci-vm-config ${DISTRO}_xci_vm_opnfv "/root/prepare-tests.sh"
+echo "Running Functest"
ssh -F $HOME/.ssh/${DISTRO}-xci-vm-config ${DISTRO}_xci_vm_opnfv "/root/run-functest.sh"
# Record exit code
functest_exit=$?
case ${DEPLOY_SCENARIO[0]} in
os-*)
- FUNCTEST_LOG=/root/results/functest.log
+ FUNCTEST_LOG=/root/functest-results/functest.log
;;
k8-*)
- FUNCTEST_LOG=/root/results/functest-kubernetes.log
+ FUNCTEST_LOG=/root/functest-results/functest-kubernetes.log
;;
*)
echo "Unable to determine the installer. Exiting!"
# http://www.apache.org/licenses/LICENSE-2.0
##############################################################################
set -o pipefail
+set -x
#----------------------------------------------------------------------
# This script is used by CI and executed by Jenkins jobs.
if [[ $GERRIT_PROJECT == "releng-xci" ]]; then
determine_default_scenario
+else
+ determine_scenario
fi
override_scenario
-determine_scenario
# ensure single scenario is impacted
if [[ $(IFS=$'\n' echo ${DEPLOY_SCENARIO[@]} | wc -w) != 1 ]]; then
- 'xci-verify-{distro}-.*'
- 'xci-merge-{distro}-.*'
- 'xci-os-.*'
+ - 'xci-k8-.*'
- 'openstack-bifrost-verify-{distro}-.*'
- 'xci-osa-verify-{distro}-.*'
- 'xci-osa-periodic-{distro}-.*'
GERRIT_CHANGE_NUMBER=$GERRIT_CHANGE_NUMBER
GERRIT_CHANGE_COMMIT_MESSAGE=$GERRIT_CHANGE_COMMIT_MESSAGE
CI_LOOP=$CI_LOOP
+ FUNCTEST_MODE=$FUNCTEST_MODE
+ FUNCTEST_SUITE_NAME=$FUNCTEST_SUITE_NAME
node-parameters: true
kill-phase-on: FAILURE
abort-all-job: true