Merge "re-enabling ODL-L3 tests"
authorTim Rozet <trozet@redhat.com>
Mon, 8 Feb 2016 21:41:47 +0000 (21:41 +0000)
committerGerrit Code Review <gerrit@172.30.200.206>
Mon, 8 Feb 2016 21:41:47 +0000 (21:41 +0000)
22 files changed:
jjb/apex/apex.yml
jjb/compass4nfv/compass-ci-jobs.yml
jjb/compass4nfv/compass-deploy.sh
jjb/compass4nfv/compass-project-jobs.yml
jjb/doctor/doctor.yml
jjb/fuel/fuel-ci-jobs.yml
jjb/fuel/fuel-download-artifact.sh
jjb/fuel/fuel-upload-artifact.sh
jjb/functest/functest-ci-jobs.yml
jjb/opnfv/installer-params.yml
jjb/opnfv/slave-params.yml
jjb/qtip/qtip-ci-jobs.yml [moved from jjb/qtip/qtip.yml with 53% similarity]
jjb/qtip/qtip-project-jobs.yml [new file with mode: 0644]
jjb/releng-macros.yaml
jjb/yardstick/yardstick-ci-jobs.yml
jjb/yardstick/yardstick-project-jobs.yml
utils/docs-build.sh
utils/fetch_os_creds.sh
utils/push-test-logs.sh
utils/test/result_collection_api/dashboard/dashboard_utils.py
utils/test/result_collection_api/dashboard/functest2Dashboard.py
utils/test/result_collection_api/dashboard/promise2Dashboard.py [new file with mode: 0644]

index 785056f..1ab7f64 100644 (file)
@@ -84,9 +84,6 @@
                     pattern: 'lib/**'
                   - compare-type: ANT
                     pattern: 'config/**'
-                forbidden-file-paths:
-                  - compare-type: ANT
-                    pattern: 'docs/**'
 
     properties:
         - build-blocker:
             git-revision: false
             block: true
         - trigger-builds:
-          - project: 'functest-apex-opnfv-jump-1-daily-{stream}'
-            predefined-parameters:
-              DEPLOY_SCENARIO={scenario}
+          - project: 'yardstick-apex-opnfv-jump-1-daily-{stream}'
             block: true
             block-thresholds:
                 build-step-failure-threshold: 'never'
                 failure-threshold: 'never'
                 unstable-threshold: 'FAILURE'
         - trigger-builds:
-          - project: 'yardstick-apex-opnfv-jump-1-daily-{stream}'
+          - project: 'functest-apex-opnfv-jump-1-daily-{stream}'
+            predefined-parameters:
+              DEPLOY_SCENARIO={scenario}
             block: true
             block-thresholds:
                 build-step-failure-threshold: 'never'
index a4be24e..5564d69 100644 (file)
 
     builders:
         - shell:
-            !include-raw: ./compass-workspace-precleanup.sh
+            !include-raw-escape: ./compass-download-artifact.sh
         - shell:
-            !include-raw: ./compass-download-artifact.sh
-        - shell:
-            !include-raw: ./compass-deploy.sh
+            !include-raw-escape: ./compass-deploy.sh
 
     publishers:
         - archive:
index 697a1d5..860f5da 100644 (file)
@@ -6,24 +6,38 @@ echo "Starting the deployment on baremetal environment using $INSTALLER_TYPE. Th
 echo "--------------------------------------------------------"
 echo
 
+# source the properties file so we get OPNFV vars
+source $BUILD_DIRECTORY/latest.properties
+
+# echo the info about artifact that is used during the deployment
+echo "Using ${OPNFV_ARTIFACT_URL/*\/} for deployment"
+
+if [[ ! "$JOB_NAME" =~ (verify|merge) ]]; then
+    # for none-merge deployments
+    # checkout the commit that was used for building the downloaded artifact
+    # to make sure the ISO and deployment mechanism uses same versions
+    echo "Checking out $OPNFV_GIT_SHA1"
+    git checkout $OPNFV_GIT_SHA1 --quiet
+fi
+
 echo 1 > /proc/sys/vm/drop_caches
 
 export CONFDIR=$WORKSPACE/deploy/conf
 export ISO_URL=file://$BUILD_DIRECTORY/compass.iso
 if [[ "$NODE_NAME" =~ "-vm" ]]; then
     export NETWORK_CONF=$CONFDIR/vm_environment/$NODE_NAME/network.yml
-    export DHA_CONF=$CONFDIR/vm_environment/${{DEPLOY_SCENARIO}}.yml
+    export DHA_CONF=$CONFDIR/vm_environment/${DEPLOY_SCENARIO}.yml
 else
     export INSTALL_NIC=eth1
     export NETWORK_CONF=$CONFDIR/hardware_environment/$NODE_NAME/network.yml
-    export DHA_CONF=$CONFDIR/hardware_environment/$NODE_NAME/${{DEPLOY_SCENARIO}}.yml
+    export DHA_CONF=$CONFDIR/hardware_environment/$NODE_NAME/${DEPLOY_SCENARIO}.yml
 fi
 
 cd $WORKSPACE
 
-export OS_VERSION=${{COMPASS_OS_VERSION}}
-export OPENSTACK_VERSION=${{COMPASS_OPENSTACK_VERSION}}
-./deploy.sh --dha ${{DHA_CONF}} --network ${{NETWORK_CONF}}
+export OS_VERSION=${COMPASS_OS_VERSION}
+export OPENSTACK_VERSION=${COMPASS_OPENSTACK_VERSION}
+./deploy.sh --dha ${DHA_CONF} --network ${NETWORK_CONF}
 if [ $? -ne 0 ]; then
     echo "depolyment failed!"
     deploy_ret=1
@@ -34,6 +48,6 @@ echo "--------------------------------------------------------"
 echo "Done!"
 
 ssh_options="-o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no"
-sshpass -p root scp 2>/dev/null $ssh_options root@${{INSTALLER_IP}}:/var/ansible/run/openstack_${{COMPASS_OPENSTACK_VERSION}}-opnfv2/ansible.log ./  &> /dev/null
+sshpass -p root scp 2>/dev/null $ssh_options root@${INSTALLER_IP}:/var/ansible/run/openstack_${COMPASS_OPENSTACK_VERSION}-opnfv2/ansible.log ./  &> /dev/null
 
 exit $deploy_ret
index 8943afb..11e1d5f 100644 (file)
 
     builders:
         - shell:
-            !include-raw: ./compass-build.sh
+            !include-raw-escape: ./compass-build.sh
         - shell:
-            !include-raw: ./compass-deploy.sh
+            !include-raw-escape: ./compass-deploy.sh
 
     publishers:
         - archive:
 
     builders:
         - shell:
-            !include-raw: ./compass-build.sh
+            !include-raw-escape: ./compass-build.sh
         - shell:
-            !include-raw: ./compass-upload-artifact.sh
+            !include-raw-escape: ./compass-upload-artifact.sh
         - shell:
-            !include-raw: ./compass-workspace-cleanup.sh
+            !include-raw-escape: ./compass-workspace-cleanup.sh
 
 - job-template:
     name: 'compass-build-ppa-{stream}'
 
     builders:
         - shell:
-            !include-raw: ./compass-makeppa.sh
+            !include-raw-escape: ./compass-makeppa.sh
 
 
 
index 167b696..b2cbb02 100644 (file)
@@ -53,4 +53,4 @@
                     pattern: 'tests/**'
 
     builders:
-        - shell: "[ -e tests/run.sh ] && ./tests/run.sh"
+        - shell: "[ -e tests/run.sh ] && bash -n ./tests/run.sh"
index 33c54ec..6b69afe 100644 (file)
@@ -35,6 +35,8 @@
 #--------------------------------
 #        master
 #--------------------------------
+        - opnfv-jump-2:
+            <<: *master
         - ericsson-pod1:
             <<: *master
         - ericsson-pod2:
 #--------------------------------
     scenario:
         # HA scenarios
-        - 'os-onos-nofeature-ha':
-#            auto-trigger-name: 'fuel-{scenario}-{pod}-trigger'
-            auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
         - 'os-odl_l2-nofeature-ha':
-            auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
+            auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
         - 'os-odl_l3-nofeature-ha':
-            auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
-        - 'os-odl_l2-bgpvpn-ha':
-            auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
-        - 'os-nosdn-nofeature-ha':
-            auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
+            auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
+        - 'os-onos-nofeature-ha':
+            auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
         - 'os-nosdn-kvm-ha':
+            auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
+        - 'os-nosdn-ovs-ha':
+            auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
+        - 'os-nosdn-nofeature-ha':
+            auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
+        - 'os-odl_l2-bgpvpn-ha':
             auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
         - 'os-nosdn-kvm_ovs-ha':
             auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
-        - 'os-nosdn-ovs-ha':
-            auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
         - 'os-nosdn-vlan-ha':
             auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
         # NOHA scenarios
 ########################
 # trigger macros
 ########################
-# os-onos3-nofeature-ha trigger
+# os-nosdn-nofeature-ha trigger
+- trigger:
+    name: 'fuel-os-nosdn-nofeature-ha-opnfv-jump-2-master-trigger'
+    triggers:
+        - timed: '0 14 * * *'
+- trigger:
+    name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod1-master-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod2-master-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-nofeature-ha-virtual-master-trigger'
+    triggers:
+        - timed: ''
 - trigger:
-    name: 'fuel-os-onos-nofeature-ha-opnfv-jump-2-trigger'
+    name: 'fuel-os-nosdn-nofeature-ha-opnfv-jump-2-brahmaputra-trigger'
     triggers:
-        - timed: '0 20 * * *'
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod1-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod2-brahmaputra-trigger'
+    triggers:
+        - timed: '0 14 * * *'
+- trigger:
+    name: 'fuel-os-nosdn-nofeature-ha-virtual-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+
+# os-nosdn-kvm-ha trigger
 - trigger:
-    name: 'fuel-os-onos-nofeature-ha-ericsson-pod1-trigger'
+    name: 'fuel-os-nosdn-kvm-ha-opnfv-jump-2-master-trigger'
+    triggers:
+        - timed: '0 18 * * *'
+- trigger:
+    name: 'fuel-os-nosdn-kvm-ha-ericsson-pod1-master-trigger'
     triggers:
         - timed: ''
 - trigger:
-    name: 'fuel-os-onos-nofeature-ha-ericsson-pod2-trigger'
+    name: 'fuel-os-nosdn-kvm-ha-ericsson-pod2-master-trigger'
     triggers:
         - timed: ''
 - trigger:
-    name: 'fuel-os-onos-nofeature-ha-virtual-trigger'
+    name: 'fuel-os-nosdn-kvm-ha-virtual-master-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-kvm-ha-opnfv-jump-2-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-kvm-ha-ericsson-pod1-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-kvm-ha-ericsson-pod2-brahmaputra-trigger'
+    triggers:
+        - timed: '0 18 * * *'
+- trigger:
+    name: 'fuel-os-nosdn-kvm-ha-virtual-brahmaputra-trigger'
     triggers:
         - timed: ''
 
 # os-odl_l2-nofeature-ha trigger
 - trigger:
-    name: 'fuel-os-odl_l2-nofeature-ha-opnfv-jump-2-trigger'
+    name: 'fuel-os-odl_l2-nofeature-ha-opnfv-jump-2-master-trigger'
+    triggers:
+        - timed: '0 2 * * *'
+- trigger:
+    name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod1-master-trigger'
     triggers:
-        - timed: '0 0 * * *'
+        - timed: ''
 - trigger:
-    name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod1-trigger'
+    name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod2-master-trigger'
     triggers:
         - timed: ''
 - trigger:
-    name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod2-trigger'
+    name: 'fuel-os-odl_l2-nofeature-ha-virtual-master-trigger'
     triggers:
         - timed: ''
 - trigger:
-    name: 'fuel-os-odl_l2-nofeature-ha-virtual-trigger'
+    name: 'fuel-os-odl_l2-nofeature-ha-opnfv-jump-2-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod1-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod2-brahmaputra-trigger'
+    triggers:
+        - timed: '0 2 * * *'
+- trigger:
+    name: 'fuel-os-odl_l2-nofeature-ha-virtual-brahmaputra-trigger'
     triggers:
         - timed: ''
 
 # os-odl_l3-nofeature-ha trigger
 - trigger:
-    name: 'fuel-os-odl_l3-nofeature-ha-opnfv-jump-2-trigger'
+    name: 'fuel-os-odl_l3-nofeature-ha-opnfv-jump-2-master-trigger'
     triggers:
-        - timed: '0 4 * * *'
+        - timed: '0 6 * * *'
 - trigger:
-    name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod1-trigger'
+    name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod1-master-trigger'
     triggers:
         - timed: ''
 - trigger:
-    name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod2-trigger'
+    name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod2-master-trigger'
     triggers:
         - timed: ''
 - trigger:
-    name: 'fuel-os-odl_l3-nofeature-ha-virtual-trigger'
+    name: 'fuel-os-odl_l3-nofeature-ha-virtual-master-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-odl_l3-nofeature-ha-opnfv-jump-2-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod1-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod2-brahmaputra-trigger'
+    triggers:
+        - timed: '0 6 * * *'
+- trigger:
+    name: 'fuel-os-odl_l3-nofeature-ha-virtual-brahmaputra-trigger'
     triggers:
         - timed: ''
 
-# os-odl_l2-bgpvpn-ha trigger
+# os-onos-nofeature-ha trigger
 - trigger:
-    name: 'fuel-os-odl_l2-bgpvpn-ha-opnfv-jump-2-trigger'
+    name: 'fuel-os-onos-nofeature-ha-opnfv-jump-2-master-trigger'
     triggers:
-        - timed: '0 8 * * *'
+        - timed: '0 10 * * *'
+- trigger:
+    name: 'fuel-os-onos-nofeature-ha-ericsson-pod1-master-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-onos-nofeature-ha-ericsson-pod2-master-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-onos-nofeature-ha-virtual-master-trigger'
+    triggers:
+        - timed: ''
 - trigger:
-    name: 'fuel-os-odl_l2-bgpvpn-ha-ericsson-pod1-trigger'
+    name: 'fuel-os-onos-nofeature-ha-opnfv-jump-2-brahmaputra-trigger'
     triggers:
         - timed: ''
 - trigger:
-    name: 'fuel-os-odl_l2-bgpvpn-ha-ericsson-pod2-trigger'
+    name: 'fuel-os-onos-nofeature-ha-ericsson-pod1-brahmaputra-trigger'
     triggers:
         - timed: ''
 - trigger:
-    name: 'fuel-os-odl_l2-bgpvpn-ha-virtual-trigger'
+    name: 'fuel-os-onos-nofeature-ha-ericsson-pod2-brahmaputra-trigger'
+    triggers:
+        - timed: '0 10 * * *'
+- trigger:
+    name: 'fuel-os-onos-nofeature-ha-virtual-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+
+
+# os-nosdn-ovs-ha trigger
+- trigger:
+    name: 'fuel-os-nosdn-ovs-ha-opnfv-jump-2-master-trigger'
+    triggers:
+        - timed: '0 22 * * *'
+- trigger:
+    name: 'fuel-os-nosdn-ovs-ha-ericsson-pod1-master-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-ovs-ha-ericsson-pod2-master-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-ovs-ha-virtual-master-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-ovs-ha-opnfv-jump-2-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-ovs-ha-ericsson-pod1-brahmaputra-trigger'
+    triggers:
+        - timed: ''
+- trigger:
+    name: 'fuel-os-nosdn-ovs-ha-ericsson-pod2-brahmaputra-trigger'
+    triggers:
+        - timed: '0 22 * * *'
+- trigger:
+    name: 'fuel-os-nosdn-ovs-ha-virtual-brahmaputra-trigger'
     triggers:
         - timed: ''
index ceac859..d78ddc5 100755 (executable)
@@ -36,7 +36,7 @@ if [[ ! "$JOB_NAME" =~ (verify|merge) ]]; then
     ISOSTORE="/iso_mount/opnfv_ci/${GIT_BRANCH##*/}"
     if [[ -f "$ISOSTORE/$OPNFV_ARTIFACT" ]]; then
         echo "ISO exists locally. Skipping the download and using the file from ISO store"
-        /bin/cp -f $ISOSTORE/$OPNFV_ARTIFACT $WORKSPACE/opnfv.iso
+        ln -s $ISOSTORE/$OPNFV_ARTIFACT $WORKSPACE/opnfv.iso
         echo "--------------------------------------------------------"
         echo
         ls -al $WORKSPACE/opnfv.iso
index 6dd5f34..47cf898 100755 (executable)
@@ -29,17 +29,17 @@ if [[ ! "$JOB_NAME" =~ (verify|merge) ]]; then
         ls -tp | grep -v '/' | tail -n +6 | xargs -d '\n' /bin/rm -f --
 
         # store ISO
-        echo "Storing $INSTALLER_TYPE artifact on NFS first"
+        echo "Storing $INSTALLER_TYPE artifact on NFS..."
         /bin/cp -f $BUILD_DIRECTORY/opnfv-$OPNFV_ARTIFACT_VERSION.iso \
             $ISOSTORE/opnfv-$OPNFV_ARTIFACT_VERSION.iso
     fi
 fi
 
 # log info to console
-echo
 echo "Uploading $INSTALLER_TYPE artifact. This could take some time..."
 echo
 
+cd $WORKSPACE
 # upload artifact and additional files to google storage
 gsutil cp $BUILD_DIRECTORY/opnfv-$OPNFV_ARTIFACT_VERSION.iso \
     gs://$GS_URL/opnfv-$OPNFV_ARTIFACT_VERSION.iso > gsutil.iso.log 2>&1
index de05922..1165b2a 100644 (file)
@@ -53,6 +53,9 @@
 #--------------------------------
 #        master
 #--------------------------------
+        - opnfv-jump-2:
+            installer: fuel
+            <<: *master
         - intel-pod6:
             installer: joid
             <<: *master
index e03b474..57ca986 100644 (file)
@@ -47,7 +47,7 @@
             description: 'Installer used for deploying OPNFV on this POD'
         - string:
             name: EXTERNAL_NETWORK
-            default: 'net04_ext'
+            default: 'admin_floating_net'
             description: 'external network for test'
 
 - parameter:
index b2c6e7c..5e0274d 100644 (file)
             name: GIT_BASE
             default: https://gerrit.opnfv.org/gerrit/$PROJECT
             description: 'Git URL to use on this Jenkins Slave'
-        - string:
-            name: EXTERNAL_NETWORK
-            default: 'admin_floating_net'
-            description: 'external network'
 
 - parameter:
     name: 'intelpod2-jumphost-defaults'
             name: SSH_KEY
             default: /root/.ssh/id_rsa
             description: 'SSH key to use for Apex'
+
+- parameter:
+    name: 'dell-us-testing-bm-1-defaults'
+    parameters:
+        - node:
+            name: SLAVE_NAME
+            description: 'Slave name on Jenkins'
+            allowed-slaves:
+                - dell-us-testing-bm-1
+            default-slaves:
+                - dell-us-testing-bm-1
+        - string:
+            name: INSTALLER_VERSION
+            default: latest
+            description: 'Version of the installer to deploy'
+        - string:
+            name: GIT_BASE
+            default: https://gerrit.opnfv.org/gerrit/$PROJECT
+            description: 'Git URL to use on this Jenkins Slave'
+
+- parameter:
+    name: 'dell-us-deploying-bm3-defaults'
+    parameters:
+        - node:
+            name: SLAVE_NAME
+            description: 'Slave name on Jenkins'
+            allowed-slaves:
+                - dell-us-deploying-bm3
+            default-slaves:
+                - dell-us-deploying-bm3
+        - string:
+            name: INSTALLER_VERSION
+            default: latest
+            description: 'Version of the installer to deploy'
+        - string:
+            name: GIT_BASE
+            default: https://gerrit.opnfv.org/gerrit/$PROJECT
+            description: 'Git URL to use on this Jenkins Slave'
similarity index 53%
rename from jjb/qtip/qtip.yml
rename to jjb/qtip/qtip-ci-jobs.yml
index 450ba93..d33bdda 100644 (file)
+####################################
+# job configuration for qtip
+####################################
 - project:
-    name: qtip
-
-    project: '{name}'
-
-    jobs:
-        - 'qtip-verify-{stream}'
-        - 'qtip-{installer_type}-{pod}-daily-{stream}'
-
+    name: qtip-ci-jobs
+
+    project: 'qtip'
+
+#--------------------------------
+# BRANCH ANCHORS
+#--------------------------------
+    master: &master
+        stream: master
+        branch: '{stream}'
+        gs-pathname: ''
+    brahmaputra: &brahmaputra
+        stream: brahmaputra
+        branch: 'stable/{stream}'
+        gs-pathname: '{stream}'
+#--------------------------------
+# POD, INSTALLER, AND BRANCH MAPPING
+#--------------------------------
+#        brahmaputra
+#--------------------------------
     pod:
         - dell-us-testing-bm-1:
-           node: 'dell-us-testing-bm-1'
-           installer_type: 'fuel'
-           installer_ip: '10.20.0.2'
+            installer: compass
+            <<: *brahmaputra
+        - orange-pod2:
+            installer: joid
+            <<: *brahmaputra
         - dell-us-deploying-bm3:
-           node: 'dell-us-deploying-bm3'
-           installer_type: 'fuel'
-           installer_ip: '10.20.0.2'
-
-# only master branch is enabled at the moment to keep no of jobs sane
-    stream:
-        - master:
-            branch: 'master'
-            gs-pathname: ''
-#        - brahmaputra:
-#            branch: 'stable/brahmaputra'
-#            gs-pathname: '/brahmaputra'
-
-- job-template:
-    name: 'qtip-verify-{stream}'
-
-    parameters:
-        - project-parameter:
-            project: '{project}'
-        - gerrit-parameter:
-            branch: '{branch}'
-        - 'opnfv-build-defaults'
-
-    scm:
-        - gerrit-trigger-scm:
-            credentials-id: '{ssh-credentials}'
-            refspec: '$GERRIT_REFSPEC'
-            choosing-strategy: 'gerrit'
-
-    triggers:
-        - gerrit:
-            trigger-on:
-                - patchset-created-event:
-                    exclude-drafts: 'false'
-                    exclude-trivial-rebase: 'false'
-                    exclude-no-code-change: 'false'
-                - draft-published-event
-                - comment-added-contains-event:
-                    comment-contains-value: 'recheck'
-                - comment-added-contains-event:
-                    comment-contains-value: 'reverify'
-            projects:
-              - project-compare-type: 'ANT'
-                project-pattern: '{project}'
-                branches:
-                  - branch-compare-type: 'ANT'
-                    branch-pattern: '**/{branch}'
-                forbidden-file-paths:
-                  - compare-type: ANT
-                    pattern: 'docs/**|.gitignore'
-
-    builders:
-        - shell: |
-            echo "Nothing to verify!"
+            installer: fuel
+            <<: *brahmaputra
+
+#--------------------------------
+#        master
+#--------------------------------
+        - juniper-pod1:
+            installer: joid
+            <<: *master
+#      - zte-build-1:   #would be confirmed with the ZTE lab by tomorrow
+#            installer: fuel
+#            <<: *master
+
+#--------------------------------
+    jobs:
+        - 'qtip-{installer}-{pod}-daily-{stream}'
 
+################################
+# job templates
+################################
 - job-template:
-    name: 'qtip-{installer_type}-{pod}-daily-{stream}'
+    name: 'qtip-{installer}-{pod}-daily-{stream}'
 
     disabled: false
 
-    node: '{node}'
-
     parameters:
       - project-parameter:
           project: '{project}'
+      - '{pod}-defaults'
+      - '{installer}-defaults'
       - string:
-          name: INSTALLER_TYPE
-          default: '{installer_type}'
-          description: "Installer name that is used for deployment."
-      - string:
-          name: INSTALLER_IP
-          default: '{installer_ip}'
-          description: "Installer IP"
-      - string:
-          name: NODE_NAME
-          default: '{node}'
-          description: "Name of the POD"
-      - string:
-          name: GIT_BASE
-          default: https://gerrit.opnfv.org/gerrit/$PROJECT
-          description: "Used for overriding the GIT URL coming from Global Jenkins configuration in case if the stuff is done on none-LF HW."
+          name: DEPLOY_SCENARIO
+          default: 'os-nosdn-nofeature-ha'
 
     scm:
        - git-scm:
            credentials-id: '{ssh-credentials}'
            refspec: ''
-           branch: master
+           branch: '{branch}'
 
     builders:
         - 'qtip-cleanup'
         - 'qtip-run-suite'
         - 'qtip-pushtoDB'
 
-    triggers:
-      - timed: '0 0,13,15,20 * * *'
-
+    publishers:
+        - email:
+            recipients: nauman.ahad@xflowresearch.com, mofassir.arif@xflowresearch.com, vikram@nvirters.com
 
 ###########################
 #biuilder macros
             envs="INSTALLER_TYPE=${INSTALLER_TYPE} -e INSTALLER_IP=${INSTALLER_IP} -e NODE_NAME=${NODE_NAME}"
             suite="TEST_CASE=all"
             docker pull opnfv/qtip:latest
-
             cmd=" docker run -id -e $envs -e $suite opnfv/qtip:latest /bin/bash"
             echo "Qtip: Running docker run command: ${cmd}"
             ${cmd}
                     docker rmi opnfv/qtip:$tag
                 done
             fi
-
-- builder:
-    name: qtip-fetch-os-cred
-    builders:
-        - shell: |
-            #!/bin/bash
-
-            echo $HOME
-            echo $WORKSPACE
-            cd $WORKSPACE
-            git clone https://gerrit.opnfv.org/gerrit/releng
-            cd $WORKSPACE/releng/utils
-            #./fetch_os_creds.sh                  #To test the problem with creating the file in $HOME
-            ./fetch_os_creds.sh -d $WORKSPACE/opnfv-stack.sh    #To test if  the file can be created in $WORKSPACE
diff --git a/jjb/qtip/qtip-project-jobs.yml b/jjb/qtip/qtip-project-jobs.yml
new file mode 100644 (file)
index 0000000..c77bac7
--- /dev/null
@@ -0,0 +1,58 @@
+- project:
+    name: qtip
+
+    project: '{name}'
+
+    jobs:
+        - 'qtip-verify-{stream}'
+
+# only master branch is enabled at the moment to keep no of jobs sane
+    stream:
+        - master:
+            branch: '{stream}'
+            gs-pathname: ''
+        - brahmaputra:
+            branch: 'stable/{stream}'
+            gs-pathname: '/{stream}'
+
+- job-template:
+    name: 'qtip-verify-{stream}'
+
+    parameters:
+        - project-parameter:
+            project: '{project}'
+        - gerrit-parameter:
+            branch: '{branch}'
+        - 'opnfv-build-defaults'
+
+    scm:
+        - gerrit-trigger-scm:
+            credentials-id: '{ssh-credentials}'
+            refspec: '$GERRIT_REFSPEC'
+            choosing-strategy: 'gerrit'
+
+    triggers:
+        - gerrit:
+            trigger-on:
+                - patchset-created-event:
+                    exclude-drafts: 'false'
+                    exclude-trivial-rebase: 'false'
+                    exclude-no-code-change: 'false'
+                - draft-published-event
+                - comment-added-contains-event:
+                    comment-contains-value: 'recheck'
+                - comment-added-contains-event:
+                    comment-contains-value: 'reverify'
+            projects:
+              - project-compare-type: 'ANT'
+                project-pattern: '{project}'
+                branches:
+                  - branch-compare-type: 'ANT'
+                    branch-pattern: '**/{branch}'
+                forbidden-file-paths:
+                  - compare-type: ANT
+                    pattern: 'docs/**|.gitignore'
+
+    builders:
+        - shell: |
+            echo "Nothing to verify!"
index 5d95707..b43a76d 100644 (file)
             echo "###########################"
             echo
 
-            gs_base="artifacts.opnfv.org/review"
+            gs_base="artifacts.opnfv.org/$PROJECT/review"
             gs_path="$gs_base/$GERRIT_CHANGE_NUMBER"
             local_path="upload/$GERRIT_CHANGE_NUMBER"
 
 
             [[ $GERRIT_CHANGE_NUMBER =~ .+ ]]
 
+            gs_path="artifacts.opnfv.org/$PROJECT/review/$GERRIT_CHANGE_NUMBER"
+
+            if gsutil ls "gs://$gs_path" > /dev/null 2>&1 ; then
+                echo
+                echo "Deleting Out-of-dated Documents..."
+                gsutil -m rm -r "gs://$gs_path"
+            fi
             gs_path="artifacts.opnfv.org/review/$GERRIT_CHANGE_NUMBER"
 
             if gsutil ls "gs://$gs_path" > /dev/null 2>&1 ; then
index ec5c547..e7221e7 100644 (file)
 #--------------------------------
 #        master
 #--------------------------------
+        - opnfv-jump-2:
+            installer: fuel
+            suite: daily
+            auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
+            <<: *master
         - ericsson-pod1:
             installer: fuel
             suite: daily
     parameters:
         - string:
             name: YARDSTICK_DB_BACKEND
-            default: '-r 213.77.62.197/results'
+            default: '-i 10.118.36.90:8086'
             description: 'Arguments to use in order to choose the backend DB'
 
 - parameter:
     parameters:
         - string:
             name: YARDSTICK_DB_BACKEND
-            default: '-r 213.77.62.197/results'
+            default: '-i 10.118.36.90:8086'
             description: 'Arguments to use in order to choose the backend DB'
 
 - parameter:
index 749ebf5..f8f1abb 100644 (file)
 
             # install python packages
             easy_install -U setuptools
-            pip install Sphinx==1.3.1 doc8 docutils sphinxcontrib-httpdomain
             python setup.py develop
 
             # unit tests
             ./run_tests.sh
 
+            pip install Sphinx==1.3.1 doc8 docutils sphinxcontrib-httpdomain
+
             # generate api doc sources
             sphinx-apidoc -o docs/apidocs yardstick
 
index e5d6b4a..3c0e8bd 100755 (executable)
@@ -41,7 +41,7 @@ revision="$(git rev-parse --short HEAD)"
 rev_full="$(git rev-parse HEAD)"
 version="$(git describe --abbrev=0 2> /dev/null || echo draft) ($revision)"
 project="$(basename $(git rev-parse --show-toplevel))"
-html_notes="\n    Revision: $rev_full\n\n    Build date: |today|"
+html_notes="    Revision: $rev_full\n    Build date: |today|"
 default_conf='releng/docs/etc/conf.py'
 opnfv_logo='releng/docs/etc/opnfv-logo.png'
 
index d1b192d..a5c601b 100755 (executable)
@@ -94,7 +94,7 @@ if [ "$installer_type" == "fuel" ]; then
     #This file contains the mgmt keystone API, we need the public one for our rc file
     admin_ip=$(cat $dest_path | grep "OS_AUTH_URL" | sed 's/^.*\=//' | sed "s/^\([\"']\)\(.*\)\1\$/\2/g" | sed s'/\/$//')
     public_ip=$(sshpass -p r00tme ssh $ssh_options root@${installer_ip} \
-        "ssh ${controller_ip} 'source openrc; keystone endpoint-list'" \
+        "ssh ${controller_ip} 'source openrc; openstack endpoint list --long'" \
         | grep $admin_ip | sed 's/ /\n/g' | grep ^http | head -1) &> /dev/null
         #| grep http | head -1 | cut -d '|' -f 4 | sed 's/v1\/.*/v1\//' | sed 's/ //g') &> /dev/null
     #NOTE: this is super ugly sed 's/v1\/.*/v1\//'OS_AUTH_URL
index 331cbd9..68f8087 100644 (file)
@@ -36,9 +36,9 @@ if [ -d "$dir_result" ]; then
         if [ $? != 0 ]; then
             echo "Not possible to push results to artifact: gsutil not installed";
         else
-            gsutil ls gs://artifacts.opnfv.org/"$project_artifact"/ &>/dev/null
+            gsutil ls gs://artifacts.opnfv.org/"$project"/ &>/dev/null
             if [ $? != 0 ]; then
-                echo "Not possible to push results to artifact: wrong credentials.";
+                echo "Not possible to push results to artifact: gsutil not installed.";
             else
                 echo "copy result files to artifact $project_artifact"
                 gsutil -m cp -r "$dir_result" gs://artifacts.opnfv.org/"$project_artifact"/
index 8d83b00..3252c3a 100644 (file)
@@ -26,6 +26,8 @@ from bottlenecks2Dashboard import format_bottlenecks_for_dashboard, \
     check_bottlenecks_case_exist
 from qtip2Dashboard import format_qtip_for_dashboard, \
     check_qtip_case_exist
+from promise2Dashboard import format_promise_for_dashboard, \
+    check_promise_case_exist
 
 # any project test project wishing to provide dashboard ready values
 # must include at least 2 methods
index 65dbca6..a817f75 100644 (file)
 import re
 import datetime
 
+
 def get_functest_cases():
     """
     get the list of the supported test cases
     TODO: update the list when adding a new test case for the dashboard
     """
-    return ["status", "vPing", "vPing_userdata", "vIMS", "Tempest", "ODL", "ONOS", "Rally"]
+    return ["status", "vPing", "vPing_userdata", "vIMS", "Tempest", "ODL",
+            "ONOS", "Rally"]
 
 
 def format_functest_for_dashboard(case, results):
@@ -66,7 +68,9 @@ def format_status_for_dashboard(results):
     test_data.append({'nb test suite(s) run': len(testcases)-1})
     test_data.append({'vPing': '100%'})
     test_data.append({'VIM status': '82%'})
-    test_data.append({'SDN Controllers': {'odl':'92%', 'onos':'95%', 'opencontrail':'93%'}})
+    test_data.append({'SDN Controllers': {'odl': '92%',
+                                          'onos': '95%',
+                                          'ocl': '93%'}})
     test_data.append({'VNF deployment': '95%'})
 
     return test_data
@@ -227,7 +231,7 @@ def format_ODL_for_dashboard(results):
         nbFailures = 0
         for odl in odl_results:
             if (odl['test_status']['@status'] == "FAIL"):
-                 nbFailures+=1
+                nbFailures += 1
         new_element.append({'x': data['creation_date'],
                             'y1': len(odl_results),
                             'y2': nbFailures})
@@ -246,25 +250,91 @@ def format_ONOS_for_dashboard(results):
     Post processing for the odl test case
     """
     test_data = [{'description': 'ONOS results for Dashboard'}]
-    # Graph 1: (duration)=f(time)
+    # Graph 1: (duration FUNCvirtNet)=f(time)
     # ***************************************
     new_element = []
 
     # default duration 0:00:08.999904
     # consider only seconds => 09
     for data in results:
-        t = data['details']['duration']
-        h,m,s = re.split(':',t)
+        t = data['details']['FUNCvirNet']['duration']
+        h, m, s = re.split(':', t)
         s = round(float(s))
-        new_duration = int(datetime.timedelta(hours=int(h),minutes=int(m),seconds=int(s)).total_seconds())
+        new_duration = int(datetime.timedelta(hours=int(h),
+                                              minutes=int(m),
+                                              seconds=int(s)).total_seconds())
         new_element.append({'x': data['creation_date'],
                             'y': new_duration})
 
-    test_data.append({'name': "ONOS duration",
+    test_data.append({'name': "ONOS FUNCvirNet duration ",
                       'info': {'type': "graph",
                                'xlabel': 'time (s)',
                                'ylabel': 'duration (s)'},
                       'data_set': new_element})
+
+    # Graph 2: (Nb test, nb failure)FuncvirtNet=f(time)
+    # ***************************************
+    new_element = []
+
+    for data in results:
+        onos_results = data['details']['FUNCvirNet']['status']
+        nbFailures = 0
+        for onos in onos_results:
+            if (onos['Case result'] == "FAIL"):
+                nbFailures += 1
+        new_element.append({'x': data['creation_date'],
+                            'y1': len(onos_results),
+                            'y2': nbFailures})
+
+    test_data.append({'name': "ONOS FUNCvirNet nb tests/nb failures",
+                      'info': {'type': "graph",
+                               'xlabel': 'time',
+                               'y1label': 'Number of tests',
+                               'y2label': 'Number of failures'},
+                      'data_set': new_element})
+
+    # Graph 3: (duration FUNCvirtNetL3)=f(time)
+    # ***************************************
+    new_element = []
+
+    # default duration 0:00:08.999904
+    # consider only seconds => 09
+    for data in results:
+        t = data['details']['FUNCvirNetL3']['duration']
+        h, m, s = re.split(':', t)
+        s = round(float(s))
+        new_duration = int(datetime.timedelta(hours=int(h),
+                                              minutes=int(m),
+                                              seconds=int(s)).total_seconds())
+        new_element.append({'x': data['creation_date'],
+                            'y': new_duration})
+
+    test_data.append({'name': "ONOS FUNCvirNetL3 duration",
+                      'info': {'type': "graph",
+                               'xlabel': 'time (s)',
+                               'ylabel': 'duration (s)'},
+                      'data_set': new_element})
+
+    # Graph 4: (Nb test, nb failure)FuncvirtNetL3=f(time)
+    # ***************************************
+    new_element = []
+
+    for data in results:
+        onos_results = data['details']['FUNCvirNetL3']['status']
+        nbFailures = 0
+        for onos in onos_results:
+            if (onos['Case result'] == "FAIL"):
+                nbFailures += 1
+        new_element.append({'x': data['creation_date'],
+                            'y1': len(onos_results),
+                            'y2': nbFailures})
+
+    test_data.append({'name': "ONOS FUNCvirNetL3 nb tests/nb failures",
+                      'info': {'type': "graph",
+                               'xlabel': 'time',
+                               'y1label': 'Number of tests',
+                               'y2label': 'Number of failures'},
+                      'data_set': new_element})
     return test_data
 
 
@@ -312,6 +382,7 @@ def format_vPing_for_dashboard(results):
 
     return test_data
 
+
 def format_vPing_userdata_for_dashboard(results):
     """
     Post processing for the vPing_userdata test case
@@ -347,3 +418,4 @@ def format_vPing_userdata_for_dashboard(results):
                                     'Nb Success': nbTestOk}]})
 
     return test_data
+
diff --git a/utils/test/result_collection_api/dashboard/promise2Dashboard.py b/utils/test/result_collection_api/dashboard/promise2Dashboard.py
new file mode 100644 (file)
index 0000000..84f43a7
--- /dev/null
@@ -0,0 +1,103 @@
+ #!/usr/bin/python
+#
+# Copyright (c) 2015 Orange
+# morgan.richomme@orange.com
+#
+# This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# This script is used to build dashboard ready json results
+# It may be used for all the test case of the Promise project
+# a new method format_<Test_case>_for_dashboard(results)
+# v0.1: basic example with methods for odl, Tempest, Rally and vPing
+#
+import re
+import datetime
+
+
+def get_promise_cases():
+    """
+    get the list of the supported test cases
+    TODO: update the list when adding a new test case for the dashboard
+    """
+    return ["promise"]
+
+
+def format_promise_for_dashboard(case, results):
+    """
+    generic method calling the method corresponding to the test case
+    check that the testcase is properly declared first
+    then build the call to the specific method
+    """
+    if check_promise_case_exist(case):
+        # note we add _case because testcase and project had the same name
+        # TODO refactoring...looks fine at the beginning wit only 1 project
+        # not very ugly now and clearly not optimized...
+        cmd = "format_" + case + "_case_for_dashboard(results)"
+        res = eval(cmd)
+    else:
+        res = []
+        print "Test cases not declared"
+    return res
+
+
+def check_promise_case_exist(case):
+    """
+    check if the testcase exists
+    if the test case is not defined or not declared in the list
+    return False
+    """
+    promise_cases = get_promise_cases()
+
+    if (case is None or case not in promise_cases):
+        return False
+    else:
+        return True
+
+
+
+
+
+def format_promise_case_for_dashboard(results):
+    """
+    Post processing for the promise test case
+    """
+    test_data = [{'description': 'Promise results for Dashboard'}]
+    # Graph 1: (duration)=f(time)
+    # ***************************************
+    new_element = []
+
+    # default duration 0:00:08.999904
+    # consider only seconds => 09
+    for data in results:
+        t = data['details']['duration']
+        new_element.append({'x': data['creation_date'],
+                            'y': t})
+
+    test_data.append({'name': "Promise duration ",
+                      'info': {'type': "graph",
+                               'xlabel': 'time (s)',
+                               'ylabel': 'duration (s)'},
+                      'data_set': new_element})
+
+    # Graph 2: (Nb test, nb failure)=f(time)
+    # ***************************************
+    new_element = []
+
+    for data in results:
+        promise_results = data['details']
+        new_element.append({'x': data['creation_date'],
+                            'y1': promise_results['tests'],
+                            'y2': promise_results['failures']})
+
+    test_data.append({'name': "Promise nb tests/nb failures",
+                      'info': {'type': "graph",
+                               'xlabel': 'time',
+                               'y1label': 'Number of tests',
+                               'y2label': 'Number of failures'},
+                      'data_set': new_element})
+
+    return test_data