X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=ci%2Fdeploy.sh;h=6005eb1de53626a4343701e405defd384e6c3d27;hb=e6f1588dbeb8f4d02398b34b621d285d5112e4c5;hp=1280625314a4f699248d183bc98481d3a9350165;hpb=85e65d41f1cdc9c9291fb5b2de29c02f0e57b405;p=joid.git diff --git a/ci/deploy.sh b/ci/deploy.sh index 12806253..6005eb1d 100755 --- a/ci/deploy.sh +++ b/ci/deploy.sh @@ -10,6 +10,7 @@ opnfvtype=nonha openstack=liberty opnfvlab=default opnfvrel=b +opnfvfeature=odl_l2 read_config() { opnfvrel=`grep release: deploy.yaml | cut -d ":" -f2` @@ -23,9 +24,10 @@ usage() { echo "Usage: $0 [-s ] [-t ] [-o ] [-l ] + [-f ] [-r ]" 1>&2 exit 1; } -while getopts ":s:t:o:l:h:r:" opt; do +while getopts ":s:t:o:l:h:r:f:" opt; do case "${opt}" in s) opnfvsdn=${OPTARG} @@ -42,6 +44,9 @@ while getopts ":s:t:o:l:h:r:" opt; do r) opnfvrel=${OPTARG} ;; + f) + opnfvfeature=${OPTARG} + ;; h) usage ;; @@ -58,6 +63,45 @@ deploy_dep() { cp environments.yaml ~/.juju/ } +#by default maas creates two VMs in case of three more VM needed. +createresource() { + maas_ip=`grep " ip_address" deployment.yaml | cut -d " " -f 10` + apikey=`grep maas-oauth: environments.yaml | cut -d "'" -f 2` + maas login maas http://${maas_ip}/MAAS/api/1.0 ${apikey} + + nodeexist=`maas maas nodes list hostname=node3-control` + + if [ $nodeexist != *node3* ]; then + sudo virt-install --connect qemu:///system --name node3-control --ram 8192 --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node3-control + + sudo virt-install --connect qemu:///system --name node4-control --ram 8192 --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node4-control + + sudo virt-install --connect qemu:///system --name node5-compute --ram 8192 --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node5-compute + + node3controlmac=`grep "mac address" node3-control | head -1 | cut -d "'" -f 2` + node4controlmac=`grep "mac address" node4-control | head -1 | cut -d "'" -f 2` + node5computemac=`grep "mac address" node5-compute | head -1 | cut -d "'" -f 2` + + sudo virsh -c qemu:///system define --file node3-control + sudo virsh -c qemu:///system define --file node4-control + sudo virsh -c qemu:///system define --file node5-compute + + controlnodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node3-control' tags='control' hostname='node3-control' power_type='virsh' mac_addresses=$node3controlmac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node3-control' | grep system_id | cut -d '"' -f 4 ` + + maas maas tag update-nodes control add=$controlnodeid + + controlnodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node4-control' tags='control' hostname='node4-control' power_type='virsh' mac_addresses=$node4controlmac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node4-control' | grep system_id | cut -d '"' -f 4 ` + + maas maas tag update-nodes control add=$controlnodeid + + computenodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node5-compute' tags='compute' hostname='node5-compute' power_type='virsh' mac_addresses=$node5computemac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node5-compute' | grep system_id | cut -d '"' -f 4 ` + + maas maas tag update-nodes compute add=$computenodeid + fi +} + +#copy the files and create extra resources needed for HA deployment +# in case of default VM labs. deploy() { #copy the script which needs to get deployed as part of ofnfv release echo "...... deploying now ......" @@ -69,13 +113,18 @@ deploy() { cp environments.yaml ~/.juju/ - cp ./$opnfvsdn/01-deploybundle.sh ./01-deploybundle.sh + if [[ "$opnfvtype" = "ha" && "$opnfvlab" = "default" ]]; then + createresource + fi + + #cp ./$opnfvsdn/01-deploybundle.sh ./01-deploybundle.sh ./00-bootstrap.sh #case default: - ./01-deploybundle.sh $opnfvtype $openstack $opnfvlab + ./01-deploybundle.sh $opnfvtype $openstack $opnfvlab $opnfvsdn $opnfvfeature } +#check whether charms are still executing the code even juju-deployer says installed. check_status() { retval=0 timeoutiter=0 @@ -84,7 +133,7 @@ check_status() { juju status > status.txt if [ "$(grep -c "executing" status.txt )" -ge 1 ]; then echo " still executing the reltionship within charms ..." - if [ $timeoutiter -ge 60 ] then + if [ $timeoutiter -ge 60 ]; then retval=1 fi timeoutiter=$((timeoutiter+1)) @@ -95,9 +144,10 @@ check_status() { echo "...... deployment finishing ......." } +#create config RC file to consume by various tests. configOpenrc() { - echo" " > ./cloud/admin-openrc + echo " " > ./cloud/admin-openrc echo "export OS_USERNAME=$1" >> ./cloud/admin-openrc echo "export OS_PASSWORD=$2" >> ./cloud/admin-openrc echo "export OS_TENANT_NAME=$3" >> ./cloud/admin-openrc @@ -105,6 +155,7 @@ configOpenrc() echo "export OS_REGION_NAME=$5" >> ./cloud/admin-openrc } +#to get the address of a service using juju unitAddress() { juju status | python -c "import yaml; import sys; print yaml.load(sys.stdin)[\"services\"][\"$1\"][\"units\"][\"$1/$2\"][\"public-address\"]" 2> /dev/null @@ -120,8 +171,8 @@ createopenrc() } if [ "$#" -eq 0 ]; then - echo "This installtion will use deploy.yaml" - read_config + echo "This installtion will use default options" + #read_config fi echo "...... deployment started ......"