X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=src%2Fceph%2Fqa%2Fworkunits%2Frest%2Ftest.py;fp=src%2Fceph%2Fqa%2Fworkunits%2Frest%2Ftest.py;h=8b55378cf8b6d588c2def15de554b8a4f7fcb628;hb=812ff6ca9fcd3e629e49d4328905f33eee8ca3f5;hp=0000000000000000000000000000000000000000;hpb=15280273faafb77777eab341909a3f495cf248d9;p=stor4nfv.git diff --git a/src/ceph/qa/workunits/rest/test.py b/src/ceph/qa/workunits/rest/test.py new file mode 100755 index 0000000..8b55378 --- /dev/null +++ b/src/ceph/qa/workunits/rest/test.py @@ -0,0 +1,424 @@ +#!/usr/bin/python + +from __future__ import print_function + +import json +import os +import requests +import subprocess +import sys +import time +import uuid +import xml.etree.ElementTree + +BASEURL = os.environ.get('BASEURL', 'http://localhost:5000/api/v0.1') + + +def fail(r, msg): + print('FAILURE: url ', r.url, file=sys.stderr) + print(msg, file=sys.stderr) + print('Response content: ', r.text, file=sys.stderr) + print('Headers: ', r.headers, file=sys.stderr) + sys.exit(1) + + +def expect(url, method, respcode, contenttype, extra_hdrs=None, data=None): + failmsg, r = expect_nofail(url, method, respcode, contenttype, extra_hdrs, + data) + if failmsg: + fail(r, failmsg) + return r + + +def expect_nofail(url, method, respcode, contenttype, extra_hdrs=None, + data=None): + + fdict = {'get':requests.get, 'put':requests.put} + f = fdict[method.lower()] + r = f(BASEURL + '/' + url, headers=extra_hdrs, data=data) + + print('{0} {1}: {2} {3}'.format(method, url, contenttype, r.status_code)) + + if r.status_code != respcode: + return 'expected {0}, got {1}'.format(respcode, r.status_code), r + + r_contenttype = r.headers['content-type'] + + if contenttype in ['json', 'xml']: + contenttype = 'application/' + contenttype + elif contenttype: + contenttype = 'text/' + contenttype + + if contenttype and r_contenttype != contenttype: + return 'expected {0}, got "{1}"'.format(contenttype, r_contenttype), r + + if contenttype.startswith('application'): + if r_contenttype == 'application/json': + try: + # older requests.py doesn't create r.myjson; create it myself + r.myjson = json.loads(r.text) + assert(r.myjson is not None) + except Exception as e: + return 'Invalid JSON returned: "{0}"'.format(str(e)), r + + if r_contenttype == 'application/xml': + try: + # if it's there, squirrel it away for use in the caller + r.tree = xml.etree.ElementTree.fromstring(r.text) + except Exception as e: + return 'Invalid XML returned: "{0}"'.format(str(e)), r + + return '', r + + +JSONHDR={'accept':'application/json'} +XMLHDR={'accept':'application/xml'} + +if __name__ == '__main__': + expect('auth/export', 'GET', 200, 'plain') + expect('auth/export.json', 'GET', 200, 'json') + expect('auth/export.xml', 'GET', 200, 'xml') + expect('auth/export', 'GET', 200, 'json', JSONHDR) + expect('auth/export', 'GET', 200, 'xml', XMLHDR) + + expect('auth/add?entity=client.xx&' + 'caps=mon&caps=allow&caps=osd&caps=allow+*', 'PUT', 200, 'json', + JSONHDR) + + r = expect('auth/export?entity=client.xx', 'GET', 200, 'plain') + # must use text/plain; default is application/x-www-form-urlencoded + expect('auth/add?entity=client.xx', 'PUT', 200, 'plain', + {'Content-Type':'text/plain'}, data=r.text) + + r = expect('auth/list', 'GET', 200, 'plain') + assert('client.xx' in r.text) + + r = expect('auth/list.json', 'GET', 200, 'json') + dictlist = r.myjson['output']['auth_dump'] + xxdict = [d for d in dictlist if d['entity'] == 'client.xx'][0] + assert(xxdict) + assert('caps' in xxdict) + assert('mon' in xxdict['caps']) + assert('osd' in xxdict['caps']) + + expect('auth/get-key?entity=client.xx', 'GET', 200, 'json', JSONHDR) + expect('auth/print-key?entity=client.xx', 'GET', 200, 'json', JSONHDR) + expect('auth/print_key?entity=client.xx', 'GET', 200, 'json', JSONHDR) + + expect('auth/caps?entity=client.xx&caps=osd&caps=allow+rw', 'PUT', 200, + 'json', JSONHDR) + r = expect('auth/list.json', 'GET', 200, 'json') + dictlist = r.myjson['output']['auth_dump'] + xxdict = [d for d in dictlist if d['entity'] == 'client.xx'][0] + assert(xxdict) + assert('caps' in xxdict) + assert(not 'mon' in xxdict['caps']) + assert('osd' in xxdict['caps']) + assert(xxdict['caps']['osd'] == 'allow rw') + + # export/import/export, compare + r = expect('auth/export', 'GET', 200, 'plain') + exp1 = r.text + assert('client.xx' in exp1) + r = expect('auth/import', 'PUT', 200, 'plain', + {'Content-Type':'text/plain'}, data=r.text) + r2 = expect('auth/export', 'GET', 200, 'plain') + assert(exp1 == r2.text) + expect('auth/del?entity=client.xx', 'PUT', 200, 'json', JSONHDR) + + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert('epoch' in r.myjson['output']) + + assert('GLOBAL' in expect('df', 'GET', 200, 'plain').text) + assert('DIRTY' in expect('df?detail=detail', 'GET', 200, 'plain').text) + # test param with no value (treated as param=param) + assert('DIRTY' in expect('df?detail', 'GET', 200, 'plain').text) + + r = expect('df', 'GET', 200, 'json', JSONHDR) + assert('total_used_bytes' in r.myjson['output']['stats']) + r = expect('df', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/stats/stats/total_used_bytes') is not None) + + r = expect('df?detail', 'GET', 200, 'json', JSONHDR) + assert('rd_bytes' in r.myjson['output']['pools'][0]['stats']) + r = expect('df?detail', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/stats/pools/pool/stats/rd_bytes') is not None) + + expect('fsid', 'GET', 200, 'json', JSONHDR) + expect('health', 'GET', 200, 'json', JSONHDR) + expect('health?detail', 'GET', 200, 'json', JSONHDR) + expect('health?detail', 'GET', 200, 'plain') + + # XXX no ceph -w equivalent yet + + expect('mds/cluster_down', 'PUT', 200, '') + expect('mds/cluster_down', 'PUT', 200, '') + expect('mds/cluster_up', 'PUT', 200, '') + expect('mds/cluster_up', 'PUT', 200, '') + + expect('mds/compat/rm_incompat?feature=4', 'PUT', 200, '') + expect('mds/compat/rm_incompat?feature=4', 'PUT', 200, '') + + r = expect('mds/compat/show', 'GET', 200, 'json', JSONHDR) + assert('incompat' in r.myjson['output']) + r = expect('mds/compat/show', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/mds_compat/incompat') is not None) + + # EEXIST from CLI + expect('mds/deactivate?who=2', 'PUT', 400, '') + + r = expect('mds/dump.xml', 'GET', 200, 'xml') + assert(r.tree.find('output/mdsmap/created') is not None) + + expect('fs/flag/set?flag_name=enable_multiple&val=true', 'PUT', 200, '') + expect('osd/pool/create?pg_num=1&pool=my_cephfs_metadata', 'PUT', 200, '') + expect('osd/pool/create?pg_num=1&pool=my_cephfs_data', 'PUT', 200, '') + expect('fs/new?fs_name=mycephfs&metadata=my_cephfs_metadata&data=my_cephfs_data', 'PUT', 200, '') + expect('osd/pool/create?pool=data2&pg_num=10', 'PUT', 200, '') + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + pools = r.myjson['output']['pools'] + poolnum = None + for p in pools: + if p['pool_name'] == 'data2': + poolnum = p['pool'] + assert(p['pg_num'] == 10) + break + assert(poolnum is not None) + expect('mds/add_data_pool?pool={0}'.format(poolnum), 'PUT', 200, '') + expect('mds/remove_data_pool?pool={0}'.format(poolnum), 'PUT', 200, '') + expect('osd/pool/delete?pool=data2&pool2=data2' + '&sure=--yes-i-really-really-mean-it', 'PUT', 200, '') + expect('mds/set?var=allow_multimds&val=true&confirm=--yes-i-really-mean-it', 'PUT', 200, '') + expect('mds/set_max_mds?maxmds=4', 'PUT', 200, '') + expect('mds/set?var=max_mds&val=4', 'PUT', 200, '') + expect('mds/set?var=max_file_size&val=1048576', 'PUT', 200, '') + expect('mds/set?var=allow_new_snaps&val=true&confirm=--yes-i-really-mean-it', 'PUT', 200, '') + expect('mds/set?var=allow_new_snaps&val=0', 'PUT', 200, '') + expect('mds/set?var=inline_data&val=true&confirm=--yes-i-really-mean-it', 'PUT', 200, '') + expect('mds/set?var=inline_data&val=0', 'PUT', 200, '') + r = expect('mds/dump.json', 'GET', 200, 'json') + assert(r.myjson['output']['max_mds'] == 4) + expect('mds/set_max_mds?maxmds=3', 'PUT', 200, '') + r = expect('mds/stat.json', 'GET', 200, 'json') + expect('mds/set?var=max_mds&val=2', 'PUT', 200, '') + r = expect('mds/stat.json', 'GET', 200, 'json') + assert('epoch' in r.myjson['output']['fsmap']) + r = expect('mds/stat.xml', 'GET', 200, 'xml') + assert(r.tree.find('output/mds_stat/fsmap/epoch') is not None) + + # more content tests below, just check format here + expect('mon/dump.json', 'GET', 200, 'json') + expect('mon/dump.xml', 'GET', 200, 'xml') + + r = expect('mon/getmap', 'GET', 200, '') + assert(len(r.text) != 0) + r = expect('mon_status.json', 'GET', 200, 'json') + assert('name' in r.myjson['output']) + r = expect('mon_status.xml', 'GET', 200, 'xml') + assert(r.tree.find('output/mon_status/name') is not None) + + bl = '192.168.0.1:0/1000' + expect('osd/blacklist?blacklistop=add&addr=' + bl, 'PUT', 200, '') + r = expect('osd/blacklist/ls.json', 'GET', 200, 'json') + assert([b for b in r.myjson['output'] if b['addr'] == bl]) + expect('osd/blacklist?blacklistop=rm&addr=' + bl, 'PUT', 200, '') + r = expect('osd/blacklist/ls.json', 'GET', 200, 'json') + assert([b for b in r.myjson['output'] if b['addr'] == bl] == []) + + expect('osd/crush/tunables?profile=legacy', 'PUT', 200, '') + expect('osd/crush/tunables?profile=bobtail', 'PUT', 200, '') + + expect('osd/scrub?who=0', 'PUT', 200, '') + expect('osd/deep-scrub?who=0', 'PUT', 200, '') + expect('osd/repair?who=0', 'PUT', 200, '') + + expect('osd/set?key=noup', 'PUT', 200, '') + + expect('osd/down?ids=0', 'PUT', 200, '') + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert(r.myjson['output']['osds'][0]['osd'] == 0) + assert(r.myjson['output']['osds'][0]['up'] == 0) + + expect('osd/unset?key=noup', 'PUT', 200, '') + + for i in range(0,100): + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert(r.myjson['output']['osds'][0]['osd'] == 0) + if r.myjson['output']['osds'][0]['up'] == 1: + break + else: + print("waiting for osd.0 to come back up", file=sys.stderr) + time.sleep(10) + + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert(r.myjson['output']['osds'][0]['osd'] == 0) + assert(r.myjson['output']['osds'][0]['up'] == 1) + + r = expect('osd/find?id=1', 'GET', 200, 'json', JSONHDR) + assert(r.myjson['output']['osd'] == 1) + + expect('osd/out?ids=1', 'PUT', 200, '') + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert(r.myjson['output']['osds'][1]['osd'] == 1) + assert(r.myjson['output']['osds'][1]['in'] == 0) + + expect('osd/in?ids=1', 'PUT', 200, '') + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert(r.myjson['output']['osds'][1]['osd'] == 1) + assert(r.myjson['output']['osds'][1]['in'] == 1) + + r = expect('osd/find?id=0', 'GET', 200, 'json', JSONHDR) + assert(r.myjson['output']['osd'] == 0) + + r = expect('osd/getmaxosd', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/getmaxosd/max_osd') is not None) + r = expect('osd/getmaxosd', 'GET', 200, 'json', JSONHDR) + saved_maxosd = r.myjson['output']['max_osd'] + expect('osd/setmaxosd?newmax=10', 'PUT', 200, '') + r = expect('osd/getmaxosd', 'GET', 200, 'json', JSONHDR) + assert(r.myjson['output']['max_osd'] == 10) + expect('osd/setmaxosd?newmax={0}'.format(saved_maxosd), 'PUT', 200, '') + r = expect('osd/getmaxosd', 'GET', 200, 'json', JSONHDR) + assert(r.myjson['output']['max_osd'] == saved_maxosd) + + osd_uuid=uuid.uuid1() + r = expect('osd/create?uuid={0}'.format(osd_uuid), 'PUT', 200, 'json', JSONHDR) + assert('osdid' in r.myjson['output']) + osdid = r.myjson['output']['osdid'] + expect('osd/lost?id={0}'.format(osdid), 'PUT', 400, '') + expect('osd/lost?id={0}&sure=--yes-i-really-mean-it'.format(osdid), + 'PUT', 200, 'json', JSONHDR) + expect('osd/rm?ids={0}'.format(osdid), 'PUT', 200, '') + r = expect('osd/ls', 'GET', 200, 'json', JSONHDR) + assert(isinstance(r.myjson['output'], list)) + r = expect('osd/ls', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/osds/osd') is not None) + + expect('osd/pause', 'PUT', 200, '') + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert('pauserd,pausewr' in r.myjson['output']['flags']) + expect('osd/unpause', 'PUT', 200, '') + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert('pauserd,pausewr' not in r.myjson['output']['flags']) + + r = expect('osd/tree', 'GET', 200, 'json', JSONHDR) + assert('nodes' in r.myjson['output']) + r = expect('osd/tree', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/tree/nodes') is not None) + + expect('osd/pool/create?pool=data2&pg_num=10', 'PUT', 200, '') + r = expect('osd/lspools', 'GET', 200, 'json', JSONHDR) + assert([p for p in r.myjson['output'] if p['poolname'] == 'data2']) + expect('osd/pool/rename?srcpool=data2&destpool=data3', 'PUT', 200, '') + r = expect('osd/lspools', 'GET', 200, 'json', JSONHDR) + assert([p for p in r.myjson['output'] if p['poolname'] == 'data3']) + expect('osd/pool/mksnap?pool=data3&snap=datasnap', 'PUT', 200, '') + r = subprocess.call('rados -p data3 lssnap | grep -q datasnap', shell=True) + assert(r == 0) + expect('osd/pool/rmsnap?pool=data3&snap=datasnap', 'PUT', 200, '') + expect('osd/pool/delete?pool=data3', 'PUT', 400, '') + expect('osd/pool/delete?pool=data3&pool2=data3&sure=--yes-i-really-really-mean-it', 'PUT', 200, '') + + r = expect('osd/stat', 'GET', 200, 'json', JSONHDR) + assert('num_up_osds' in r.myjson['output']) + r = expect('osd/stat', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/osdmap/num_up_osds') is not None) + + r = expect('osd/ls', 'GET', 200, 'json', JSONHDR) + for osdid in r.myjson['output']: + expect('tell/osd.{0}/version'.format(osdid), 'GET', 200, '') + + expect('pg/debug?debugop=unfound_objects_exist', 'GET', 200, '') + expect('pg/debug?debugop=degraded_pgs_exist', 'GET', 200, '') + expect('pg/deep-scrub?pgid=1.0', 'PUT', 200, '') + r = expect('pg/dump', 'GET', 200, 'json', JSONHDR) + assert('pg_stats_sum' in r.myjson['output']) + r = expect('pg/dump', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/pg_map/pg_stats_sum') is not None) + + expect('pg/dump_json', 'GET', 200, 'json', JSONHDR) + expect('pg/dump_pools_json', 'GET', 200, 'json', JSONHDR) + expect('pg/dump_stuck?stuckops=inactive', 'GET', 200, '') + expect('pg/dump_stuck?stuckops=unclean', 'GET', 200, '') + expect('pg/dump_stuck?stuckops=stale', 'GET', 200, '') + + r = expect('pg/getmap', 'GET', 200, '') + assert(len(r.text) != 0) + + r = expect('pg/map?pgid=1.0', 'GET', 200, 'json', JSONHDR) + assert('acting' in r.myjson['output']) + assert(r.myjson['output']['pgid'] == '1.0') + r = expect('pg/map?pgid=1.0', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/pg_map/acting') is not None) + assert(r.tree.find('output/pg_map/pgid').text == '1.0') + + expect('pg/repair?pgid=1.0', 'PUT', 200, '') + expect('pg/scrub?pgid=1.0', 'PUT', 200, '') + + expect('osd/set-full-ratio?ratio=0.90', 'PUT', 200, '') + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert(float(r.myjson['output']['full_ratio']) == 0.90) + expect('osd/set-full-ratio?ratio=0.95', 'PUT', 200, '') + expect('osd/set-backfillfull-ratio?ratio=0.88', 'PUT', 200, '') + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert(float(r.myjson['output']['backfillfull_ratio']) == 0.88) + expect('osd/set-backfillfull-ratio?ratio=0.90', 'PUT', 200, '') + expect('osd/set-nearfull-ratio?ratio=0.90', 'PUT', 200, '') + r = expect('osd/dump', 'GET', 200, 'json', JSONHDR) + assert(float(r.myjson['output']['nearfull_ratio']) == 0.90) + expect('osd/set-nearfull-ratio?ratio=0.85', 'PUT', 200, '') + + r = expect('pg/stat', 'GET', 200, 'json', JSONHDR) + assert('num_pgs' in r.myjson['output']) + r = expect('pg/stat', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/pg_summary/num_pgs') is not None) + + expect('tell/1.0/query', 'GET', 200, 'json', JSONHDR) + expect('quorum?quorumcmd=enter', 'PUT', 200, 'json', JSONHDR) + expect('quorum?quorumcmd=enter', 'PUT', 200, 'xml', XMLHDR) + expect('quorum_status', 'GET', 200, 'json', JSONHDR) + expect('quorum_status', 'GET', 200, 'xml', XMLHDR) + + # report's CRC needs to be handled + # r = expect('report', 'GET', 200, 'json', JSONHDR) + # assert('osd_stats' in r.myjson['output']) + # r = expect('report', 'GET', 200, 'xml', XMLHDR) + # assert(r.tree.find('output/report/osdmap') is not None) + + r = expect('status', 'GET', 200, 'json', JSONHDR) + assert('osdmap' in r.myjson['output']) + r = expect('status', 'GET', 200, 'xml', XMLHDR) + assert(r.tree.find('output/status/osdmap') is not None) + + r = expect('tell/osd.0/version', 'GET', 200, '') + assert('ceph version' in r.text) + expect('tell/osd.999/version', 'GET', 400, '') + expect('tell/osd.foo/version', 'GET', 400, '') + + r = expect('tell/osd.0/dump_pg_recovery_stats', 'GET', 200, '') + assert('Started' in r.text) + + expect('osd/reweight?id=0&weight=0.9', 'PUT', 200, '') + expect('osd/reweight?id=0&weight=-1', 'PUT', 400, '') + expect('osd/reweight?id=0&weight=1', 'PUT', 200, '') + + for v in ['pg_num', 'pgp_num', 'size', 'min_size', + 'crush_rule']: + r = expect('osd/pool/get.json?pool=rbd&var=' + v, 'GET', 200, 'json') + assert(v in r.myjson['output']) + + r = expect('osd/pool/get.json?pool=rbd&var=size', 'GET', 200, 'json') + assert(r.myjson['output']['size'] >= 2) + + expect('osd/pool/set?pool=rbd&var=size&val=3', 'PUT', 200, 'plain') + r = expect('osd/pool/get.json?pool=rbd&var=size', 'GET', 200, 'json') + assert(r.myjson['output']['size'] == 3) + + expect('osd/pool/set?pool=rbd&var=size&val=2', 'PUT', 200, 'plain') + r = expect('osd/pool/get.json?pool=rbd&var=size', 'GET', 200, 'json') + assert(r.myjson['output']['size'] == 2) + + r = expect('osd/pool/get.json?pool=rbd&var=crush_rule', 'GET', 200, 'json') + assert(r.myjson['output']['crush_rule'] == "replicated_rule") + + print('OK')