
Since Python 2.7, unittest in the standard library has included mulitple facilities for skipping tests by decorators as well as an exception. Switch to that directly, rather than importing nose. Change-Id: I4009033473ea24f0d0faed3670db844f40051f30
382 lines
16 KiB
Python
382 lines
16 KiB
Python
#!/usr/bin/python -u
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
|
# implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
import json
|
|
import uuid
|
|
import random
|
|
import unittest
|
|
|
|
from six.moves.urllib.parse import urlparse
|
|
from swiftclient import client, ClientException
|
|
|
|
from swift.common.http import HTTP_NOT_FOUND
|
|
from swift.common.manager import Manager
|
|
from test.probe.brain import BrainSplitter
|
|
from test.probe.common import ReplProbeTest, ENABLED_POLICIES
|
|
|
|
|
|
def get_current_realm_cluster(url):
|
|
parts = urlparse(url)
|
|
url = parts.scheme + '://' + parts.netloc + '/info'
|
|
http_conn = client.http_connection(url)
|
|
try:
|
|
info = client.get_capabilities(http_conn)
|
|
except client.ClientException:
|
|
raise unittest.SkipTest('Unable to retrieve cluster info')
|
|
try:
|
|
realms = info['container_sync']['realms']
|
|
except KeyError:
|
|
raise unittest.SkipTest('Unable to find container sync realms')
|
|
for realm, realm_info in realms.items():
|
|
for cluster, options in realm_info['clusters'].items():
|
|
if options.get('current', False):
|
|
return realm, cluster
|
|
raise unittest.SkipTest('Unable find current realm cluster')
|
|
|
|
|
|
class TestContainerSync(ReplProbeTest):
|
|
|
|
def setUp(self):
|
|
super(TestContainerSync, self).setUp()
|
|
self.realm, self.cluster = get_current_realm_cluster(self.url)
|
|
|
|
def _setup_synced_containers(
|
|
self, source_overrides=None, dest_overrides=None):
|
|
# these defaults are used to create both source and dest containers
|
|
# unless overridden by source_overrides and/or dest_overrides
|
|
default_params = {'url': self.url,
|
|
'token': self.token,
|
|
'account': self.account,
|
|
'sync_key': 'secret'}
|
|
|
|
# setup dest container
|
|
dest = dict(default_params)
|
|
dest['name'] = 'dest-container-%s' % uuid.uuid4()
|
|
dest.update(dest_overrides or {})
|
|
dest_headers = {}
|
|
dest_policy = None
|
|
if len(ENABLED_POLICIES) > 1:
|
|
dest_policy = random.choice(ENABLED_POLICIES)
|
|
dest_headers['X-Storage-Policy'] = dest_policy.name
|
|
if dest['sync_key'] is not None:
|
|
dest_headers['X-Container-Sync-Key'] = dest['sync_key']
|
|
client.put_container(dest['url'], dest['token'], dest['name'],
|
|
headers=dest_headers)
|
|
|
|
# setup source container
|
|
source = dict(default_params)
|
|
source['name'] = 'source-container-%s' % uuid.uuid4()
|
|
source.update(source_overrides or {})
|
|
source_headers = {}
|
|
sync_to = '//%s/%s/%s/%s' % (self.realm, self.cluster, dest['account'],
|
|
dest['name'])
|
|
source_headers['X-Container-Sync-To'] = sync_to
|
|
if source['sync_key'] is not None:
|
|
source_headers['X-Container-Sync-Key'] = source['sync_key']
|
|
if dest_policy:
|
|
source_policy = random.choice([p for p in ENABLED_POLICIES
|
|
if p is not dest_policy])
|
|
source_headers['X-Storage-Policy'] = source_policy.name
|
|
client.put_container(source['url'], source['token'], source['name'],
|
|
headers=source_headers)
|
|
|
|
return source['name'], dest['name']
|
|
|
|
def test_sync(self):
|
|
source_container, dest_container = self._setup_synced_containers()
|
|
|
|
# upload to source
|
|
object_name = 'object-%s' % uuid.uuid4()
|
|
put_headers = {'X-Object-Meta-Test': 'put_value'}
|
|
client.put_object(self.url, self.token, source_container, object_name,
|
|
'test-body', headers=put_headers)
|
|
|
|
# cycle container-sync
|
|
Manager(['container-sync']).once()
|
|
|
|
resp_headers, body = client.get_object(self.url, self.token,
|
|
dest_container, object_name)
|
|
self.assertEqual(body, 'test-body')
|
|
self.assertIn('x-object-meta-test', resp_headers)
|
|
self.assertEqual('put_value', resp_headers['x-object-meta-test'])
|
|
|
|
# update metadata with a POST
|
|
post_headers = {'Content-Type': 'image/jpeg',
|
|
'X-Object-Meta-Test': 'post_value'}
|
|
int_client = self.make_internal_client()
|
|
int_client.set_object_metadata(self.account, source_container,
|
|
object_name, post_headers)
|
|
# sanity checks...
|
|
resp_headers = client.head_object(
|
|
self.url, self.token, source_container, object_name)
|
|
self.assertIn('x-object-meta-test', resp_headers)
|
|
self.assertEqual('post_value', resp_headers['x-object-meta-test'])
|
|
self.assertEqual('image/jpeg', resp_headers['content-type'])
|
|
|
|
# cycle container-sync
|
|
Manager(['container-sync']).once()
|
|
|
|
# verify that metadata changes were sync'd
|
|
resp_headers, body = client.get_object(self.url, self.token,
|
|
dest_container, object_name)
|
|
self.assertEqual(body, 'test-body')
|
|
self.assertIn('x-object-meta-test', resp_headers)
|
|
self.assertEqual('post_value', resp_headers['x-object-meta-test'])
|
|
self.assertEqual('image/jpeg', resp_headers['content-type'])
|
|
|
|
# delete the object
|
|
client.delete_object(
|
|
self.url, self.token, source_container, object_name)
|
|
with self.assertRaises(ClientException) as cm:
|
|
client.get_object(
|
|
self.url, self.token, source_container, object_name)
|
|
self.assertEqual(404, cm.exception.http_status) # sanity check
|
|
|
|
# cycle container-sync
|
|
Manager(['container-sync']).once()
|
|
|
|
# verify delete has been sync'd
|
|
with self.assertRaises(ClientException) as cm:
|
|
client.get_object(
|
|
self.url, self.token, dest_container, object_name)
|
|
self.assertEqual(404, cm.exception.http_status) # sanity check
|
|
|
|
def test_sync_slo_manifest(self):
|
|
# Verify that SLO manifests are sync'd even if their segments can not
|
|
# be found in the destination account at time of sync'ing.
|
|
# Create source and dest containers for manifest in separate accounts.
|
|
dest_account = self.account_2
|
|
source_container, dest_container = self._setup_synced_containers(
|
|
dest_overrides=dest_account
|
|
)
|
|
|
|
# Create source and dest containers for segments in separate accounts.
|
|
# These containers must have same name for the destination SLO manifest
|
|
# to be able to resolve segments. Initially the destination has no sync
|
|
# key so segments will not sync.
|
|
segs_container = 'segments-%s' % uuid.uuid4()
|
|
dest_segs_info = dict(dest_account)
|
|
dest_segs_info.update({'name': segs_container, 'sync_key': None})
|
|
self._setup_synced_containers(
|
|
source_overrides={'name': segs_container, 'sync_key': 'segs_key'},
|
|
dest_overrides=dest_segs_info)
|
|
|
|
# upload a segment to source
|
|
segment_name = 'segment-%s' % uuid.uuid4()
|
|
segment_data = 'segment body' # it's ok for first segment to be small
|
|
segment_etag = client.put_object(
|
|
self.url, self.token, segs_container, segment_name,
|
|
segment_data)
|
|
|
|
manifest = [{'etag': segment_etag,
|
|
'size_bytes': len(segment_data),
|
|
'path': '/%s/%s' % (segs_container, segment_name)}]
|
|
manifest_name = 'manifest-%s' % uuid.uuid4()
|
|
put_headers = {'X-Object-Meta-Test': 'put_value'}
|
|
client.put_object(
|
|
self.url, self.token, source_container, manifest_name,
|
|
json.dumps(manifest), headers=put_headers,
|
|
query_string='multipart-manifest=put')
|
|
|
|
resp_headers, manifest_body = client.get_object(
|
|
self.url, self.token, source_container, manifest_name,
|
|
query_string='multipart-manifest=get')
|
|
int_manifest = json.loads(manifest_body)
|
|
|
|
# cycle container-sync
|
|
Manager(['container-sync']).once()
|
|
|
|
# verify manifest was sync'd
|
|
resp_headers, dest_listing = client.get_container(
|
|
dest_account['url'], dest_account['token'], dest_container)
|
|
self.assertFalse(dest_listing[1:])
|
|
self.assertEqual(manifest_name, dest_listing[0]['name'])
|
|
|
|
# verify manifest body
|
|
resp_headers, body = client.get_object(
|
|
dest_account['url'], dest_account['token'], dest_container,
|
|
manifest_name, query_string='multipart-manifest=get')
|
|
self.assertEqual(int_manifest, json.loads(body))
|
|
self.assertIn('x-object-meta-test', resp_headers)
|
|
self.assertEqual('put_value', resp_headers['x-object-meta-test'])
|
|
|
|
# attempt to GET the SLO will fail because the segment wasn't sync'd
|
|
with self.assertRaises(ClientException) as cm:
|
|
client.get_object(dest_account['url'], dest_account['token'],
|
|
dest_container, manifest_name)
|
|
self.assertEqual(409, cm.exception.http_status)
|
|
|
|
# now set sync key on destination segments container
|
|
client.put_container(
|
|
dest_account['url'], dest_account['token'], segs_container,
|
|
headers={'X-Container-Sync-Key': 'segs_key'})
|
|
|
|
# cycle container-sync
|
|
Manager(['container-sync']).once()
|
|
|
|
# sanity check - verify manifest body
|
|
resp_headers, body = client.get_object(
|
|
dest_account['url'], dest_account['token'], dest_container,
|
|
manifest_name, query_string='multipart-manifest=get')
|
|
self.assertEqual(int_manifest, json.loads(body))
|
|
self.assertIn('x-object-meta-test', resp_headers)
|
|
self.assertEqual('put_value', resp_headers['x-object-meta-test'])
|
|
|
|
# verify GET of SLO manifest now succeeds
|
|
resp_headers, body = client.get_object(
|
|
dest_account['url'], dest_account['token'], dest_container,
|
|
manifest_name)
|
|
self.assertEqual(segment_data, body)
|
|
|
|
def test_sync_lazy_skey(self):
|
|
# Create synced containers, but with no key at source
|
|
source_container, dest_container =\
|
|
self._setup_synced_containers(source_overrides={'sync_key': None})
|
|
|
|
# upload to source
|
|
object_name = 'object-%s' % uuid.uuid4()
|
|
client.put_object(self.url, self.token, source_container, object_name,
|
|
'test-body')
|
|
|
|
# cycle container-sync, nothing should happen
|
|
Manager(['container-sync']).once()
|
|
with self.assertRaises(ClientException) as err:
|
|
_junk, body = client.get_object(self.url, self.token,
|
|
dest_container, object_name)
|
|
self.assertEqual(err.exception.http_status, HTTP_NOT_FOUND)
|
|
|
|
# amend source key
|
|
source_headers = {'X-Container-Sync-Key': 'secret'}
|
|
client.put_container(self.url, self.token, source_container,
|
|
headers=source_headers)
|
|
# cycle container-sync, should replicate
|
|
Manager(['container-sync']).once()
|
|
_junk, body = client.get_object(self.url, self.token,
|
|
dest_container, object_name)
|
|
self.assertEqual(body, 'test-body')
|
|
|
|
def test_sync_lazy_dkey(self):
|
|
# Create synced containers, but with no key at dest
|
|
source_container, dest_container =\
|
|
self._setup_synced_containers(dest_overrides={'sync_key': None})
|
|
|
|
# upload to source
|
|
object_name = 'object-%s' % uuid.uuid4()
|
|
client.put_object(self.url, self.token, source_container, object_name,
|
|
'test-body')
|
|
|
|
# cycle container-sync, nothing should happen
|
|
Manager(['container-sync']).once()
|
|
with self.assertRaises(ClientException) as err:
|
|
_junk, body = client.get_object(self.url, self.token,
|
|
dest_container, object_name)
|
|
self.assertEqual(err.exception.http_status, HTTP_NOT_FOUND)
|
|
|
|
# amend dest key
|
|
dest_headers = {'X-Container-Sync-Key': 'secret'}
|
|
client.put_container(self.url, self.token, dest_container,
|
|
headers=dest_headers)
|
|
# cycle container-sync, should replicate
|
|
Manager(['container-sync']).once()
|
|
_junk, body = client.get_object(self.url, self.token,
|
|
dest_container, object_name)
|
|
self.assertEqual(body, 'test-body')
|
|
|
|
def test_sync_with_stale_container_rows(self):
|
|
source_container, dest_container = self._setup_synced_containers()
|
|
brain = BrainSplitter(self.url, self.token, source_container,
|
|
None, 'container')
|
|
|
|
# upload to source
|
|
object_name = 'object-%s' % uuid.uuid4()
|
|
client.put_object(self.url, self.token, source_container, object_name,
|
|
'test-body')
|
|
|
|
# check source container listing
|
|
_, listing = client.get_container(
|
|
self.url, self.token, source_container)
|
|
for expected_obj_dict in listing:
|
|
if expected_obj_dict['name'] == object_name:
|
|
break
|
|
else:
|
|
self.fail('Failed to find source object %r in container listing %r'
|
|
% (object_name, listing))
|
|
|
|
# stop all container servers
|
|
brain.stop_primary_half()
|
|
brain.stop_handoff_half()
|
|
|
|
# upload new object content to source - container updates will fail
|
|
client.put_object(self.url, self.token, source_container, object_name,
|
|
'new-test-body')
|
|
source_headers = client.head_object(
|
|
self.url, self.token, source_container, object_name)
|
|
|
|
# start all container servers
|
|
brain.start_primary_half()
|
|
brain.start_handoff_half()
|
|
|
|
# sanity check: source container listing should not have changed
|
|
_, listing = client.get_container(
|
|
self.url, self.token, source_container)
|
|
for actual_obj_dict in listing:
|
|
if actual_obj_dict['name'] == object_name:
|
|
self.assertDictEqual(expected_obj_dict, actual_obj_dict)
|
|
break
|
|
else:
|
|
self.fail('Failed to find source object %r in container listing %r'
|
|
% (object_name, listing))
|
|
|
|
# cycle container-sync - object should be correctly sync'd despite
|
|
# stale info in container row
|
|
Manager(['container-sync']).once()
|
|
|
|
# verify sync'd object has same content and headers
|
|
dest_headers, body = client.get_object(self.url, self.token,
|
|
dest_container, object_name)
|
|
self.assertEqual(body, 'new-test-body')
|
|
mismatched_headers = []
|
|
for k in ('etag', 'content-length', 'content-type', 'x-timestamp',
|
|
'last-modified'):
|
|
if source_headers[k] == dest_headers[k]:
|
|
continue
|
|
mismatched_headers.append((k, source_headers[k], dest_headers[k]))
|
|
if mismatched_headers:
|
|
msg = '\n'.join([('Mismatched header %r, expected %r but got %r'
|
|
% item) for item in mismatched_headers])
|
|
self.fail(msg)
|
|
|
|
def test_sync_newer_remote(self):
|
|
source_container, dest_container = self._setup_synced_containers()
|
|
|
|
# upload to source
|
|
object_name = 'object-%s' % uuid.uuid4()
|
|
client.put_object(self.url, self.token, source_container, object_name,
|
|
'old-source-body')
|
|
|
|
# upload to dest with same name
|
|
client.put_object(self.url, self.token, dest_container, object_name,
|
|
'new-test-body')
|
|
|
|
# cycle container-sync
|
|
Manager(['container-sync']).once()
|
|
|
|
# verify that the remote object did not change
|
|
resp_headers, body = client.get_object(self.url, self.token,
|
|
dest_container, object_name)
|
|
self.assertEqual(body, 'new-test-body')
|
|
|
|
|
|
if __name__ == "__main__":
|
|
unittest.main()
|