<<: *master
- 'os-odl-sfc-ha':
<<: *master
- - 'os-odl-fdio-dvr-noha':
+ - 'os-odl-fdio_dvr-noha':
<<: *master
- - 'os-odl-fdio-dvr-ha':
+ - 'os-odl-fdio_dvr-ha':
<<: *master
- 'os-nosdn-nofeature-noha':
<<: *euphrates
<<: *euphrates
- 'os-odl-sfc-ha':
<<: *euphrates
- - 'os-odl-fdio-dvr-noha':
+ - 'os-odl-fdio_dvr-noha':
<<: *euphrates
- - 'os-odl-fdio-dvr-ha':
+ - 'os-odl-fdio_dvr-ha':
<<: *euphrates
platform:
abort-all-job: true
git-revision: false
- - name: 'apex-os-odl-fdio-dvr-noha-baremetal-master'
+ - name: 'apex-os-odl-fdio_dvr-noha-baremetal-master'
node-parameters: false
current-parameters: false
predefined-parameters: |
abort-all-job: true
git-revision: false
- - name: 'apex-os-odl-fdio-dvr-ha-baremetal-master'
+ - name: 'apex-os-odl-fdio_dvr-ha-baremetal-master'
node-parameters: false
current-parameters: false
predefined-parameters: |
abort-all-job: true
git-revision: false
- - name: 'apex-os-odl-fdio-dvr-noha-baremetal-euphrates'
+ - name: 'apex-os-odl-fdio_dvr-noha-baremetal-euphrates'
node-parameters: false
current-parameters: false
predefined-parameters: |
abort-all-job: true
git-revision: false
- - name: 'apex-os-odl-fdio-dvr-ha-baremetal-euphrates'
+ - name: 'apex-os-odl-fdio_dvr-ha-baremetal-euphrates'
node-parameters: false
current-parameters: false
predefined-parameters: |
- 'os-nosdn-kvm_ovs_dpdk-ha'
- 'os-odl-sfc-noha'
- 'os-odl-sfc-ha'
- - 'os-odl-fdio-dvr-noha'
- - 'os-odl-fdio-dvr-ha'
+ - 'os-odl-fdio_dvr-noha'
+ - 'os-odl-fdio_dvr-ha'
euphrates:
- 'os-nosdn-nofeature-noha'
- 'os-nosdn-nofeature-ha'
- 'os-nosdn-kvm_ovs_dpdk-ha'
- 'os-odl-sfc-noha'
- 'os-odl-sfc-ha'
- - 'os-odl-fdio-dvr-noha'
- - 'os-odl-fdio-dvr-ha'
+ - 'os-odl-fdio_dvr-noha'
+ - 'os-odl-fdio_dvr-ha'
danube:
- 'os-nosdn-nofeature-noha'
- 'os-nosdn-nofeature-ha'
build-step-failure-threshold: 'never'
failure-threshold: 'never'
unstable-threshold: 'FAILURE'
+ - conditional-step:
+ condition-kind: and
+ condition-operands:
+ - condition-kind: regex-match
+ regex: os-nosdn-nofeature-ha
+ label: '{scenario}'
+ steps:
+ - trigger-builds:
+ - project: 'bottlenecks-compass-posca_stress_ping-{pod}-daily-{stream}'
+ current-parameters: false
+ predefined-parameters:
+ DEPLOY_SCENARIO={scenario}
+ block: true
+ same-node: true
+ block-thresholds:
+ build-step-failure-threshold: 'never'
+ failure-threshold: 'never'
+ unstable-threshold: 'FAILURE'
- job-template:
name: 'compass-deploy-{pod}-daily-{stream}'
echo "Dovetail: Pulling image opnfv/dovetail:${DOCKER_TAG}"
docker pull opnfv/dovetail:$DOCKER_TAG >$redirect
+env4bgpvpn="-e INSTALLER_TYPE=${INSTALLER_TYPE} -e INSTALLER_IP=${INSTALLER_IP}"
+
cmd="docker run ${opts} -e DOVETAIL_HOME=${DOVETAIL_HOME} ${docker_volume} ${dovetail_home_volume} \
- ${sshkey} opnfv/dovetail:${DOCKER_TAG} /bin/bash"
+ ${sshkey} ${env4bgpvpn} opnfv/dovetail:${DOCKER_TAG} /bin/bash"
echo "Dovetail: running docker run command: ${cmd}"
${cmd} >${redirect}
sleep 5
sudo scp $ssh_options root@$installer_ip:/home/stack/overcloudrc.v3 $dest_path
elif [ "$installer_type" == "compass" ]; then
- if [ "${BRANCH}" == "master" ]; then
- sudo docker cp compass-tasks:/opt/openrc $dest_path &> /dev/null
- sudo chown $(whoami):$(whoami) $dest_path
- sudo docker cp compass-tasks:/opt/os_cacert $os_cacert
- else
+ if [ "${BRANCH}" == "stable/danube" ]; then
verify_connectivity $installer_ip
controller_ip=$(sshpass -p'root' ssh 2>/dev/null $ssh_options root@${installer_ip} \
'mysql -ucompass -pcompass -Dcompass -e"select * from cluster;"' \
fi
info "public_ip: $public_ip"
swap_to_public $public_ip
+ else
+ sudo docker cp compass-tasks:/opt/openrc $dest_path &> /dev/null
+ sudo chown $(whoami):$(whoami) $dest_path
+ sudo docker cp compass-tasks:/opt/os_cacert $os_cacert
fi
elif [ "$installer_type" == "joid" ]; then
# For all the installers
for installer in INSTALLERS:
# get scenarios results data
+ if version != 'master':
+ new_version = "stable/{}".format(version)
+ else:
+ new_version = version
scenario_results = rp_utils.getScenarios("bottlenecks",
"posca_factor_ping",
installer,
- version)
+ new_version)
LOGGER.info("scenario_results: %s", scenario_results)
scenario_stats = rp_utils.getScenarioStats(scenario_results)
import jinja2
-import testCase as tc
-import scenarioResult as sr
+import reporting.functest.testCase as tc
+import reporting.functest.scenarioResult as sr
import reporting.utils.reporting_utils as rp_utils
"""
LOGGER.debug("No results found")
items[s] = testCases2BeDisplayed
- except Exception:
+ except Exception: # pylint: disable=broad-except
LOGGER.error("Error installer %s, version %s, scenario %s",
installer, version, s)
LOGGER.error("No data available: %s", sys.exc_info()[0])