X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=src%2Fceph%2Fsrc%2Ftest%2Frgw%2Frgw_multi%2Fzone_rados.py;fp=src%2Fceph%2Fsrc%2Ftest%2Frgw%2Frgw_multi%2Fzone_rados.py;h=0000000000000000000000000000000000000000;hb=7da45d65be36d36b880cc55c5036e96c24b53f00;hp=8fae22ea8a997b24e47567991b887202696e89ae;hpb=691462d09d0987b47e112d6ee8740375df3c51b2;p=stor4nfv.git diff --git a/src/ceph/src/test/rgw/rgw_multi/zone_rados.py b/src/ceph/src/test/rgw/rgw_multi/zone_rados.py deleted file mode 100644 index 8fae22e..0000000 --- a/src/ceph/src/test/rgw/rgw_multi/zone_rados.py +++ /dev/null @@ -1,89 +0,0 @@ -import logging - -try: - from itertools import izip_longest as zip_longest -except ImportError: - from itertools import zip_longest - -from nose.tools import eq_ as eq - -from .multisite import * - -log = logging.getLogger(__name__) - -def check_object_eq(k1, k2, check_extra = True): - assert k1 - assert k2 - log.debug('comparing key name=%s', k1.name) - eq(k1.name, k2.name) - eq(k1.get_contents_as_string(), k2.get_contents_as_string()) - eq(k1.metadata, k2.metadata) - eq(k1.cache_control, k2.cache_control) - eq(k1.content_type, k2.content_type) - eq(k1.content_encoding, k2.content_encoding) - eq(k1.content_disposition, k2.content_disposition) - eq(k1.content_language, k2.content_language) - eq(k1.etag, k2.etag) - eq(k1.last_modified, k2.last_modified) - if check_extra: - eq(k1.owner.id, k2.owner.id) - eq(k1.owner.display_name, k2.owner.display_name) - eq(k1.storage_class, k2.storage_class) - eq(k1.size, k2.size) - eq(k1.version_id, k2.version_id) - eq(k1.encrypted, k2.encrypted) - - -class RadosZone(Zone): - def __init__(self, name, zonegroup = None, cluster = None, data = None, zone_id = None, gateways = None): - super(RadosZone, self).__init__(name, zonegroup, cluster, data, zone_id, gateways) - - def tier_type(self): - return "rados" - - - class Conn(ZoneConn): - def __init__(self, zone, credentials): - super(RadosZone.Conn, self).__init__(zone, credentials) - - def get_bucket(self, name): - return self.conn.get_bucket(name) - - def create_bucket(self, name): - return self.conn.create_bucket(name) - - def check_bucket_eq(self, zone_conn, bucket_name): - log.info('comparing bucket=%s zones={%s, %s}', bucket_name, self.name, zone_conn.name) - b1 = self.get_bucket(bucket_name) - b2 = zone_conn.get_bucket(bucket_name) - - log.debug('bucket1 objects:') - for o in b1.get_all_versions(): - log.debug('o=%s', o.name) - log.debug('bucket2 objects:') - for o in b2.get_all_versions(): - log.debug('o=%s', o.name) - - for k1, k2 in zip_longest(b1.get_all_versions(), b2.get_all_versions()): - if k1 is None: - log.critical('key=%s is missing from zone=%s', k2.name, self.name) - assert False - if k2 is None: - log.critical('key=%s is missing from zone=%s', k1.name, zone_conn.name) - assert False - - check_object_eq(k1, k2) - - # now get the keys through a HEAD operation, verify that the available data is the same - k1_head = b1.get_key(k1.name) - k2_head = b2.get_key(k2.name) - - check_object_eq(k1_head, k2_head, False) - - log.info('success, bucket identical: bucket=%s zones={%s, %s}', bucket_name, self.name, zone_conn.name) - - return True - - def get_conn(self, credentials): - return self.Conn(self, credentials) -