
If the global configuration option 'enable_open_expired' is set to true in the config, then the client will be able to make a request with the header 'x-open-expired' set to true in order to access an object that has expired, provided it is in its grace period. If this config flag is set to false, the client will not be able to access any expired objects, even with the header, which is the default behavior unless the flag is set. When a client sets a 'x-open-expired' header to a true value for a GET/HEAD/POST request the proxy will forward x-backend-open-expired to storage server. The storage server will allow clients that set x-backend-open-expired to open and read an object that has not yet been reaped by the object-expirer, even after the x-delete-at time has passed. The header is always ignored when used with temporary URLs. Co-Authored-By: Anish Kachinthaya <akachinthaya@nvidia.com> Related-Change: I106103438c4162a561486ac73a09436e998ae1f0 Change-Id: Ibe7dde0e3bf587d77e14808b169c02f8fb3dddb3
7947 lines
344 KiB
Python
7947 lines
344 KiB
Python
#!/usr/bin/env python
|
|
# Copyright (c) 2010-2012 OpenStack Foundation
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
|
# implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
import collections
|
|
import itertools
|
|
import math
|
|
import random
|
|
import time
|
|
import unittest
|
|
import argparse
|
|
from collections import defaultdict
|
|
from contextlib import contextmanager
|
|
import json
|
|
|
|
import mock
|
|
from eventlet import Timeout, sleep
|
|
from eventlet.queue import Empty
|
|
|
|
import six
|
|
from six import StringIO
|
|
from six.moves import range
|
|
from six.moves.urllib.parse import quote
|
|
if six.PY2:
|
|
from email.parser import FeedParser as EmailFeedParser
|
|
else:
|
|
from email.parser import BytesFeedParser as EmailFeedParser
|
|
|
|
import swift
|
|
from swift.common import utils, swob, exceptions
|
|
from swift.common.exceptions import ChunkWriteTimeout, ShortReadError, \
|
|
ChunkReadTimeout, RangeAlreadyComplete
|
|
from swift.common.utils import Timestamp, list_from_csv, md5, FileLikeIter, \
|
|
ShardRange, Namespace, NamespaceBoundList
|
|
from swift.proxy import server as proxy_server
|
|
from swift.proxy.controllers import obj
|
|
from swift.proxy.controllers.base import \
|
|
get_container_info as _real_get_container_info, GetterSource, \
|
|
NodeIter
|
|
from swift.common.storage_policy import POLICIES, ECDriverError, \
|
|
StoragePolicy, ECStoragePolicy
|
|
from swift.common.swob import Request, wsgi_to_str
|
|
from test.debug_logger import debug_logger
|
|
from test.unit import (
|
|
FakeRing, fake_http_connect, patch_policies, SlowBody, FakeStatus,
|
|
DEFAULT_TEST_EC_TYPE, encode_frag_archive_bodies, make_ec_object_stub,
|
|
fake_ec_node_response, StubResponse, mocked_http_conn,
|
|
quiet_eventlet_exceptions, FakeSource, make_timestamp_iter, FakeMemcache,
|
|
node_error_count, node_error_counts)
|
|
|
|
|
|
def unchunk_body(chunked_body):
|
|
body = b''
|
|
remaining = chunked_body
|
|
while remaining:
|
|
hex_length, remaining = remaining.split(b'\r\n', 1)
|
|
length = int(hex_length, 16)
|
|
body += remaining[:length]
|
|
remaining = remaining[length + 2:]
|
|
return body
|
|
|
|
|
|
@contextmanager
|
|
def set_http_connect(*args, **kwargs):
|
|
old_connect = swift.proxy.controllers.base.http_connect
|
|
new_connect = fake_http_connect(*args, **kwargs)
|
|
try:
|
|
swift.proxy.controllers.base.http_connect = new_connect
|
|
swift.proxy.controllers.obj.http_connect = new_connect
|
|
swift.proxy.controllers.account.http_connect = new_connect
|
|
swift.proxy.controllers.container.http_connect = new_connect
|
|
yield new_connect
|
|
left_over_status = list(new_connect.code_iter)
|
|
if left_over_status:
|
|
raise AssertionError('left over status %r' % left_over_status)
|
|
finally:
|
|
swift.proxy.controllers.base.http_connect = old_connect
|
|
swift.proxy.controllers.obj.http_connect = old_connect
|
|
swift.proxy.controllers.account.http_connect = old_connect
|
|
swift.proxy.controllers.container.http_connect = old_connect
|
|
|
|
|
|
class PatchedObjControllerApp(proxy_server.Application):
|
|
"""
|
|
This patch is just a hook over the proxy server's __call__ to ensure
|
|
that calls to get_container_info will return the stubbed value for
|
|
container_info if it's a container info call.
|
|
"""
|
|
|
|
container_info = {}
|
|
per_container_info = {}
|
|
|
|
def __call__(self, *args, **kwargs):
|
|
|
|
def _fake_get_container_info(env, app, swift_source=None):
|
|
_vrs, account, container, _junk = utils.split_path(
|
|
swob.wsgi_to_str(env['PATH_INFO']), 3, 4)
|
|
|
|
# Seed the cache with our container info so that the real
|
|
# get_container_info finds it.
|
|
ic = env.setdefault('swift.infocache', {})
|
|
cache_key = "container/%s/%s" % (account, container)
|
|
|
|
old_value = ic.get(cache_key)
|
|
|
|
# Copy the container info so we don't hand out a reference to a
|
|
# mutable thing that's set up only once at compile time. Nothing
|
|
# *should* mutate it, but it's better to be paranoid than wrong.
|
|
if container in self.per_container_info:
|
|
ic[cache_key] = self.per_container_info[container].copy()
|
|
else:
|
|
ic[cache_key] = self.container_info.copy()
|
|
|
|
real_info = _real_get_container_info(env, app, swift_source)
|
|
|
|
if old_value is None:
|
|
del ic[cache_key]
|
|
else:
|
|
ic[cache_key] = old_value
|
|
|
|
return real_info
|
|
|
|
with mock.patch('swift.proxy.server.get_container_info',
|
|
new=_fake_get_container_info), \
|
|
mock.patch('swift.proxy.controllers.base.get_container_info',
|
|
new=_fake_get_container_info):
|
|
return super(
|
|
PatchedObjControllerApp, self).__call__(*args, **kwargs)
|
|
|
|
|
|
def make_footers_callback(body=None):
|
|
# helper method to create a footers callback that will generate some fake
|
|
# footer metadata
|
|
cont_etag = 'container update etag may differ'
|
|
crypto_etag = '20242af0cd21dd7195a10483eb7472c9'
|
|
etag_crypto_meta = \
|
|
'{"cipher": "AES_CTR_256", "iv": "sD+PSw/DfqYwpsVGSo0GEw=="}'
|
|
etag = md5(body,
|
|
usedforsecurity=False).hexdigest() if body is not None else None
|
|
footers_to_add = {
|
|
'X-Object-Sysmeta-Container-Update-Override-Etag': cont_etag,
|
|
'X-Object-Sysmeta-Crypto-Etag': crypto_etag,
|
|
'X-Object-Sysmeta-Crypto-Meta-Etag': etag_crypto_meta,
|
|
'X-I-Feel-Lucky': 'Not blocked',
|
|
'Etag': etag}
|
|
|
|
def footers_callback(footers):
|
|
footers.update(footers_to_add)
|
|
|
|
return footers_callback
|
|
|
|
|
|
class BaseObjectControllerMixin(object):
|
|
container_info = {
|
|
'status': 200,
|
|
'write_acl': None,
|
|
'read_acl': None,
|
|
'storage_policy': None,
|
|
'sync_key': None,
|
|
'versions': None,
|
|
}
|
|
|
|
# this needs to be set on the test case
|
|
controller_cls = None
|
|
|
|
def setUp(self):
|
|
# setup fake rings with handoffs
|
|
for policy in POLICIES:
|
|
policy.object_ring.max_more_nodes = policy.object_ring.replicas
|
|
|
|
self.logger = debug_logger('proxy-server')
|
|
self.logger.thread_locals = ('txn1', '127.0.0.2')
|
|
# increase connection timeout to avoid intermittent failures
|
|
conf = {'conn_timeout': 1.0}
|
|
self.app = PatchedObjControllerApp(
|
|
conf, account_ring=FakeRing(),
|
|
container_ring=FakeRing(), logger=self.logger)
|
|
self.logger.clear() # startup/loading debug msgs not helpful
|
|
|
|
# you can over-ride the container_info just by setting it on the app
|
|
# (see PatchedObjControllerApp for details)
|
|
self.app.container_info = dict(self.container_info)
|
|
|
|
# default policy and ring references
|
|
self.policy = POLICIES.default
|
|
self.obj_ring = self.policy.object_ring
|
|
self._ts_iter = (utils.Timestamp(t) for t in
|
|
itertools.count(int(time.time())))
|
|
|
|
def ts(self):
|
|
return next(self._ts_iter)
|
|
|
|
def replicas(self, policy=None):
|
|
policy = policy or POLICIES.default
|
|
return policy.object_ring.replicas
|
|
|
|
def quorum(self, policy=None):
|
|
policy = policy or POLICIES.default
|
|
return policy.quorum
|
|
|
|
|
|
class CommonObjectControllerMixin(BaseObjectControllerMixin):
|
|
# defines tests that are common to all storage policy types
|
|
|
|
def test_GET_all_primaries_error_limited(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
obj_ring = self.app.get_object_ring(int(self.policy))
|
|
_part, primary_nodes = obj_ring.get_nodes('a', 'c', 'o')
|
|
for dev in primary_nodes:
|
|
self.app.error_limiter.limit(dev)
|
|
|
|
num_handoff = (
|
|
2 * self.policy.object_ring.replica_count) - len(primary_nodes)
|
|
codes = [404] * num_handoff
|
|
with mocked_http_conn(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
def test_iter_nodes_local_first_noops_when_no_affinity(self):
|
|
# this test needs a stable node order - most don't
|
|
self.app.sort_nodes = lambda l, *args, **kwargs: l
|
|
controller = self.controller_cls(
|
|
self.app, 'a', 'c', 'o')
|
|
policy = self.policy
|
|
self.app.get_policy_options(policy).write_affinity_is_local_fn = None
|
|
object_ring = policy.object_ring
|
|
all_nodes = object_ring.get_part_nodes(1)
|
|
all_nodes.extend(object_ring.get_more_nodes(1))
|
|
|
|
for node in all_nodes:
|
|
node['use_replication'] = False
|
|
local_first_nodes = list(controller.iter_nodes_local_first(
|
|
object_ring, 1, Request.blank('')))
|
|
|
|
self.maxDiff = None
|
|
|
|
self.assertEqual(all_nodes, local_first_nodes)
|
|
|
|
def test_iter_nodes_local_first_moves_locals_first(self):
|
|
controller = self.controller_cls(
|
|
self.app, 'a', 'c', 'o')
|
|
policy_conf = self.app.get_policy_options(self.policy)
|
|
policy_conf.write_affinity_is_local_fn = (
|
|
lambda node: node['region'] == 1)
|
|
# we'll write to one more than replica count local nodes
|
|
policy_conf.write_affinity_node_count_fn = lambda r: r + 1
|
|
|
|
object_ring = self.policy.object_ring
|
|
# make our fake ring have plenty of nodes, and not get limited
|
|
# artificially by the proxy max request node count
|
|
object_ring.max_more_nodes = 100000
|
|
# nothing magic about * 2 + 3, just a way to make it bigger
|
|
self.app.request_node_count = lambda r: r * 2 + 3
|
|
|
|
all_nodes = object_ring.get_part_nodes(1)
|
|
all_nodes.extend(object_ring.get_more_nodes(1))
|
|
for node in all_nodes:
|
|
node['use_replication'] = False
|
|
|
|
# limit to the number we're going to look at in this request
|
|
nodes_requested = self.app.request_node_count(object_ring.replicas)
|
|
all_nodes = all_nodes[:nodes_requested]
|
|
|
|
# make sure we have enough local nodes (sanity)
|
|
all_local_nodes = [n for n in all_nodes if
|
|
policy_conf.write_affinity_is_local_fn(n)]
|
|
self.assertGreaterEqual(len(all_local_nodes), self.replicas() + 1)
|
|
|
|
# finally, create the local_first_nodes iter and flatten it out
|
|
local_first_nodes = list(controller.iter_nodes_local_first(
|
|
object_ring, 1, Request.blank('')))
|
|
|
|
# the local nodes move up in the ordering
|
|
self.assertEqual([1] * (self.replicas() + 1), [
|
|
node['region'] for node in local_first_nodes[
|
|
:self.replicas() + 1]])
|
|
# we don't skip any nodes
|
|
self.assertEqual(len(all_nodes), len(local_first_nodes))
|
|
self.assertEqual(sorted(all_nodes, key=lambda dev: dev['id']),
|
|
sorted(local_first_nodes, key=lambda dev: dev['id']))
|
|
|
|
for node in all_nodes:
|
|
node['use_replication'] = True
|
|
|
|
req = Request.blank(
|
|
'/v1/a/c', headers={'x-backend-use-replication-network': 'yes'})
|
|
local_first_nodes = list(controller.iter_nodes_local_first(
|
|
object_ring, 1, request=req))
|
|
self.assertEqual([1] * (self.replicas() + 1), [
|
|
node['region'] for node in local_first_nodes[
|
|
:self.replicas() + 1]])
|
|
# we don't skip any nodes
|
|
self.assertEqual(len(all_nodes), len(local_first_nodes))
|
|
self.assertEqual(sorted(all_nodes, key=lambda dev: dev['id']),
|
|
sorted(local_first_nodes, key=lambda dev: dev['id']))
|
|
|
|
def test_iter_nodes_local_first_best_effort(self):
|
|
controller = self.controller_cls(
|
|
self.app, 'a', 'c', 'o')
|
|
policy_conf = self.app.get_policy_options(self.policy)
|
|
policy_conf.write_affinity_is_local_fn = (
|
|
lambda node: node['region'] == 1)
|
|
|
|
object_ring = self.policy.object_ring
|
|
all_nodes = object_ring.get_part_nodes(1)
|
|
all_nodes.extend(object_ring.get_more_nodes(1))
|
|
for node in all_nodes:
|
|
node['use_replication'] = False
|
|
|
|
local_first_nodes = list(controller.iter_nodes_local_first(
|
|
object_ring, 1, request=Request.blank('')))
|
|
|
|
# we won't have quite enough local nodes...
|
|
self.assertEqual(len(all_nodes), self.replicas() +
|
|
POLICIES.default.object_ring.max_more_nodes)
|
|
all_local_nodes = [n for n in all_nodes if
|
|
policy_conf.write_affinity_is_local_fn(n)]
|
|
self.assertEqual(len(all_local_nodes), self.replicas())
|
|
# but the local nodes we do have are at the front of the local iter
|
|
first_n_local_first_nodes = local_first_nodes[:len(all_local_nodes)]
|
|
self.assertEqual(sorted(all_local_nodes, key=lambda dev: dev['id']),
|
|
sorted(first_n_local_first_nodes,
|
|
key=lambda dev: dev['id']))
|
|
# but we *still* don't *skip* any nodes
|
|
self.assertEqual(len(all_nodes), len(local_first_nodes))
|
|
self.assertEqual(sorted(all_nodes, key=lambda dev: dev['id']),
|
|
sorted(local_first_nodes, key=lambda dev: dev['id']))
|
|
|
|
def test_iter_nodes_local_handoff_first_noops_when_no_affinity(self):
|
|
# this test needs a stable node order - most don't
|
|
self.app.sort_nodes = lambda l, *args, **kwargs: l
|
|
controller = self.controller_cls(
|
|
self.app, 'a', 'c', 'o')
|
|
policy = self.policy
|
|
self.app.get_policy_options(policy).write_affinity_is_local_fn = None
|
|
object_ring = policy.object_ring
|
|
all_nodes = object_ring.get_part_nodes(1)
|
|
all_nodes.extend(object_ring.get_more_nodes(1))
|
|
for node in all_nodes:
|
|
node['use_replication'] = False
|
|
|
|
local_first_nodes = list(controller.iter_nodes_local_first(
|
|
object_ring, 1, local_handoffs_first=True,
|
|
request=Request.blank('')))
|
|
|
|
self.maxDiff = None
|
|
|
|
self.assertEqual(all_nodes, local_first_nodes)
|
|
|
|
def test_iter_nodes_handoff_local_first_default(self):
|
|
controller = self.controller_cls(
|
|
self.app, 'a', 'c', 'o')
|
|
policy_conf = self.app.get_policy_options(self.policy)
|
|
policy_conf.write_affinity_is_local_fn = (
|
|
lambda node: node['region'] == 1)
|
|
|
|
object_ring = self.policy.object_ring
|
|
primary_nodes = object_ring.get_part_nodes(1)
|
|
handoff_nodes_iter = object_ring.get_more_nodes(1)
|
|
all_nodes = primary_nodes + list(handoff_nodes_iter)
|
|
for node in all_nodes:
|
|
node['use_replication'] = False
|
|
handoff_nodes_iter = object_ring.get_more_nodes(1)
|
|
local_handoffs = [n for n in handoff_nodes_iter if
|
|
policy_conf.write_affinity_is_local_fn(n)]
|
|
|
|
prefered_nodes = list(controller.iter_nodes_local_first(
|
|
object_ring, 1, local_handoffs_first=True,
|
|
request=Request.blank('')))
|
|
|
|
self.assertEqual(len(all_nodes), self.replicas() +
|
|
POLICIES.default.object_ring.max_more_nodes)
|
|
|
|
first_primary_nodes = prefered_nodes[:len(primary_nodes)]
|
|
self.assertEqual(sorted(primary_nodes, key=lambda dev: dev['id']),
|
|
sorted(first_primary_nodes,
|
|
key=lambda dev: dev['id']))
|
|
|
|
handoff_count = self.replicas() - len(primary_nodes)
|
|
first_handoffs = prefered_nodes[len(primary_nodes):][:handoff_count]
|
|
self.assertEqual(first_handoffs, local_handoffs[:handoff_count])
|
|
|
|
def test_iter_nodes_handoff_local_first_non_default(self):
|
|
# Obviously this test doesn't work if we're testing 1 replica.
|
|
# In that case, we don't have any failovers to check.
|
|
if self.replicas() == 1:
|
|
return
|
|
|
|
controller = self.controller_cls(
|
|
self.app, 'a', 'c', 'o')
|
|
policy_conf = self.app.get_policy_options(self.policy)
|
|
policy_conf.write_affinity_is_local_fn = (
|
|
lambda node: node['region'] == 1)
|
|
policy_conf.write_affinity_handoff_delete_count = 1
|
|
|
|
object_ring = self.policy.object_ring
|
|
primary_nodes = object_ring.get_part_nodes(1)
|
|
handoff_nodes_iter = object_ring.get_more_nodes(1)
|
|
all_nodes = primary_nodes + list(handoff_nodes_iter)
|
|
for node in all_nodes:
|
|
node['use_replication'] = False
|
|
handoff_nodes_iter = object_ring.get_more_nodes(1)
|
|
local_handoffs = [n for n in handoff_nodes_iter if
|
|
policy_conf.write_affinity_is_local_fn(n)]
|
|
for node in local_handoffs:
|
|
node['use_replication'] = False
|
|
|
|
prefered_nodes = list(controller.iter_nodes_local_first(
|
|
object_ring, 1, local_handoffs_first=True,
|
|
request=Request.blank('')))
|
|
|
|
self.assertEqual(len(all_nodes), self.replicas() +
|
|
POLICIES.default.object_ring.max_more_nodes)
|
|
|
|
first_primary_nodes = prefered_nodes[:len(primary_nodes)]
|
|
self.assertEqual(sorted(primary_nodes, key=lambda dev: dev['id']),
|
|
sorted(first_primary_nodes,
|
|
key=lambda dev: dev['id']))
|
|
|
|
handoff_count = policy_conf.write_affinity_handoff_delete_count
|
|
first_handoffs = prefered_nodes[len(primary_nodes):][:handoff_count]
|
|
self.assertEqual(first_handoffs, local_handoffs[:handoff_count])
|
|
|
|
def test_connect_put_node_timeout(self):
|
|
controller = self.controller_cls(
|
|
self.app, 'a', 'c', 'o')
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
self.app.conn_timeout = 0.05
|
|
with set_http_connect(slow_connect=True):
|
|
nodes = [dict(ip='', port='', device='')]
|
|
res = controller._connect_put_node(nodes, '', req, {}, ('', ''))
|
|
self.assertIsNone(res)
|
|
|
|
def test_DELETE_simple(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
codes = [204] * self.replicas()
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
def test_object_DELETE_backend_update_container_ip_default(self):
|
|
self.policy.object_ring = FakeRing(separate_replication=True)
|
|
self.app.container_ring = FakeRing(separate_replication=True)
|
|
# sanity, devs have different ip & replication_ip
|
|
for ring in (self.policy.object_ring, self.app.container_ring):
|
|
for dev in ring.devs:
|
|
self.assertNotEqual(dev['ip'], dev['replication_ip'])
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
codes = [204] * self.replicas()
|
|
with mocked_http_conn(*codes) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
# sanity, object hosts use node ip/port
|
|
object_hosts = {'%(ip)s:%(port)s' % req for req in log.requests}
|
|
object_ring = self.app.get_object_ring(int(self.policy))
|
|
part, object_nodes = object_ring.get_nodes('a', 'c', 'o')
|
|
expected_object_hosts = {'%(ip)s:%(port)s' % n for n in object_nodes}
|
|
self.assertEqual(object_hosts, expected_object_hosts)
|
|
|
|
# container hosts use node ip/port
|
|
container_hosts = {req['headers']['x-container-host']
|
|
for req in log.requests}
|
|
part, container_nodes = self.app.container_ring.get_nodes('a', 'c')
|
|
expected_container_hosts = {'%(ip)s:%(port)s' % n
|
|
for n in container_nodes}
|
|
self.assertEqual(container_hosts, expected_container_hosts)
|
|
|
|
def test_repl_object_DELETE_backend_update_container_repl_ip(self):
|
|
self.policy.object_ring = FakeRing(separate_replication=True)
|
|
self.app.container_ring = FakeRing(separate_replication=True)
|
|
# sanity, devs have different ip & replication_ip
|
|
for ring in (self.policy.object_ring, self.app.container_ring):
|
|
for dev in ring.devs:
|
|
self.assertNotEqual(dev['ip'], dev['replication_ip'])
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='DELETE', headers={
|
|
'x-backend-use-replication-network': 'true'})
|
|
codes = [204] * self.replicas()
|
|
with mocked_http_conn(*codes) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
# sanity, object hosts use node replication ip/port
|
|
object_hosts = {'%(ip)s:%(port)s' % req for req in log.requests}
|
|
object_ring = self.app.get_object_ring(int(self.policy))
|
|
part, object_nodes = object_ring.get_nodes('a', 'c', 'o')
|
|
expected_object_hosts = {
|
|
'%(replication_ip)s:%(replication_port)s' % n
|
|
for n in object_nodes}
|
|
self.assertEqual(object_hosts, expected_object_hosts)
|
|
|
|
# container hosts use node replication ip/port
|
|
container_hosts = {req['headers']['x-container-host']
|
|
for req in log.requests}
|
|
part, container_nodes = self.app.container_ring.get_nodes('a', 'c')
|
|
expected_container_hosts = {
|
|
'%(replication_ip)s:%(replication_port)s' % n
|
|
for n in container_nodes}
|
|
self.assertEqual(container_hosts, expected_container_hosts)
|
|
|
|
def test_DELETE_missing_one(self):
|
|
# Obviously this test doesn't work if we're testing 1 replica.
|
|
# In that case, we don't have any failovers to check.
|
|
if self.replicas() == 1:
|
|
return
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
codes = [404] + [204] * (self.replicas() - 1)
|
|
random.shuffle(codes)
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
def test_DELETE_not_found(self):
|
|
# Obviously this test doesn't work if we're testing 1 replica.
|
|
# In that case, we don't have any failovers to check.
|
|
if self.replicas() == 1:
|
|
return
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
codes = [404] * (self.replicas() - 1) + [204]
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
def test_DELETE_mostly_found(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
mostly_204s = [204] * self.quorum()
|
|
codes = mostly_204s + [404] * (self.replicas() - len(mostly_204s))
|
|
self.assertEqual(len(codes), self.replicas())
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
def test_DELETE_mostly_not_found(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
mostly_404s = [404] * self.quorum()
|
|
codes = mostly_404s + [204] * (self.replicas() - len(mostly_404s))
|
|
self.assertEqual(len(codes), self.replicas())
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
def test_DELETE_half_not_found_statuses(self):
|
|
self.obj_ring.set_replicas(4)
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
with set_http_connect(404, 204, 404, 204):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
def test_DELETE_half_not_found_headers_and_body(self):
|
|
# Transformed responses have bogus bodies and headers, so make sure we
|
|
# send the client headers and body from a real node's response.
|
|
self.obj_ring.set_replicas(4)
|
|
|
|
status_codes = (404, 404, 204, 204)
|
|
bodies = (b'not found', b'not found', b'', b'')
|
|
headers = [{}, {}, {'Pick-Me': 'yes'}, {'Pick-Me': 'yes'}]
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
with set_http_connect(*status_codes, body_iter=bodies,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
self.assertEqual(resp.headers.get('Pick-Me'), 'yes')
|
|
self.assertEqual(resp.body, b'')
|
|
|
|
def test_DELETE_handoff(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
codes = [204] * self.replicas()
|
|
with set_http_connect(507, *codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
def test_DELETE_limits_expirer_queue_updates(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
codes = [204] * self.replicas()
|
|
captured_headers = []
|
|
|
|
def capture_headers(ip, port, device, part, method, path,
|
|
headers=None, **kwargs):
|
|
captured_headers.append(headers)
|
|
|
|
with set_http_connect(*codes, give_connect=capture_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204) # sanity check
|
|
|
|
counts = {True: 0, False: 0, None: 0}
|
|
for headers in captured_headers:
|
|
v = headers.get('X-Backend-Clean-Expiring-Object-Queue')
|
|
norm_v = None if v is None else utils.config_true_value(v)
|
|
counts[norm_v] += 1
|
|
|
|
max_queue_updates = 2
|
|
o_replicas = self.replicas()
|
|
self.assertEqual(counts, {
|
|
True: min(max_queue_updates, o_replicas),
|
|
False: max(o_replicas - max_queue_updates, 0),
|
|
None: 0,
|
|
})
|
|
|
|
def test_expirer_DELETE_suppresses_expirer_queue_updates(self):
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='DELETE', headers={
|
|
'X-Backend-Clean-Expiring-Object-Queue': 'no'})
|
|
codes = [204] * self.replicas()
|
|
captured_headers = []
|
|
|
|
def capture_headers(ip, port, device, part, method, path,
|
|
headers=None, **kwargs):
|
|
captured_headers.append(headers)
|
|
|
|
with set_http_connect(*codes, give_connect=capture_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204) # sanity check
|
|
|
|
counts = {True: 0, False: 0, None: 0}
|
|
for headers in captured_headers:
|
|
v = headers.get('X-Backend-Clean-Expiring-Object-Queue')
|
|
norm_v = None if v is None else utils.config_true_value(v)
|
|
counts[norm_v] += 1
|
|
|
|
o_replicas = self.replicas()
|
|
self.assertEqual(counts, {
|
|
True: 0,
|
|
False: o_replicas,
|
|
None: 0,
|
|
})
|
|
|
|
# Make sure we're not sending any expirer-queue update headers here.
|
|
# Since we're not updating the expirer queue, these headers would be
|
|
# superfluous.
|
|
for headers in captured_headers:
|
|
self.assertNotIn('X-Delete-At-Container', headers)
|
|
self.assertNotIn('X-Delete-At-Partition', headers)
|
|
self.assertNotIn('X-Delete-At-Host', headers)
|
|
self.assertNotIn('X-Delete-At-Device', headers)
|
|
|
|
def test_DELETE_write_affinity_after_replication(self):
|
|
policy_conf = self.app.get_policy_options(self.policy)
|
|
policy_conf.write_affinity_handoff_delete_count = self.replicas() // 2
|
|
policy_conf.write_affinity_is_local_fn = (
|
|
lambda node: node['region'] == 1)
|
|
handoff_count = policy_conf.write_affinity_handoff_delete_count
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
codes = [204] * self.replicas() + [404] * handoff_count
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
def test_DELETE_write_affinity_before_replication(self):
|
|
policy_conf = self.app.get_policy_options(self.policy)
|
|
policy_conf.write_affinity_handoff_delete_count = self.replicas() // 2
|
|
policy_conf.write_affinity_is_local_fn = (
|
|
lambda node: node['region'] == 1)
|
|
handoff_count = policy_conf.write_affinity_handoff_delete_count
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
codes = ([204] * (self.replicas() - handoff_count) +
|
|
[404] * handoff_count +
|
|
[204] * handoff_count)
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
def test_PUT_limits_expirer_queue_deletes(self):
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', body=b'',
|
|
headers={'Content-Type': 'application/octet-stream'})
|
|
codes = [201] * self.replicas()
|
|
captured_headers = []
|
|
|
|
def capture_headers(ip, port, device, part, method, path,
|
|
headers=None, **kwargs):
|
|
captured_headers.append(headers)
|
|
|
|
expect_headers = {
|
|
'X-Obj-Metadata-Footer': 'yes',
|
|
'X-Obj-Multiphase-Commit': 'yes'
|
|
}
|
|
with set_http_connect(*codes, give_connect=capture_headers,
|
|
expect_headers=expect_headers):
|
|
# this req may or may not succeed depending on the Putter type used
|
|
# but that's ok because we're only interested in verifying the
|
|
# headers that were sent
|
|
req.get_response(self.app)
|
|
|
|
counts = {True: 0, False: 0, None: 0}
|
|
for headers in captured_headers:
|
|
v = headers.get('X-Backend-Clean-Expiring-Object-Queue')
|
|
norm_v = None if v is None else utils.config_true_value(v)
|
|
counts[norm_v] += 1
|
|
|
|
max_queue_updates = 2
|
|
o_replicas = self.replicas()
|
|
self.assertEqual(counts, {
|
|
True: min(max_queue_updates, o_replicas),
|
|
False: max(o_replicas - max_queue_updates, 0),
|
|
None: 0,
|
|
})
|
|
|
|
def test_POST_limits_expirer_queue_deletes(self):
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='POST', body=b'',
|
|
headers={'Content-Type': 'application/octet-stream'})
|
|
codes = [201] * self.replicas()
|
|
captured_headers = []
|
|
|
|
def capture_headers(ip, port, device, part, method, path,
|
|
headers=None, **kwargs):
|
|
captured_headers.append(headers)
|
|
|
|
with set_http_connect(*codes, give_connect=capture_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201) # sanity check
|
|
|
|
counts = {True: 0, False: 0, None: 0}
|
|
for headers in captured_headers:
|
|
v = headers.get('X-Backend-Clean-Expiring-Object-Queue')
|
|
norm_v = None if v is None else utils.config_true_value(v)
|
|
counts[norm_v] += 1
|
|
|
|
max_queue_updates = 2
|
|
o_replicas = self.replicas()
|
|
self.assertEqual(counts, {
|
|
True: min(max_queue_updates, o_replicas),
|
|
False: max(o_replicas - max_queue_updates, 0),
|
|
None: 0,
|
|
})
|
|
|
|
def test_POST_non_int_delete_after(self):
|
|
t = str(int(time.time() + 100)) + '.1'
|
|
req = swob.Request.blank('/v1/a/c/o', method='POST',
|
|
headers={'Content-Type': 'foo/bar',
|
|
'X-Delete-After': t})
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 400)
|
|
self.assertEqual(b'Non-integer X-Delete-After', resp.body)
|
|
|
|
def test_PUT_non_int_delete_after(self):
|
|
t = str(int(time.time() + 100)) + '.1'
|
|
req = swob.Request.blank('/v1/a/c/o', method='PUT', body=b'',
|
|
headers={'Content-Type': 'foo/bar',
|
|
'X-Delete-After': t})
|
|
with set_http_connect():
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 400)
|
|
self.assertEqual(b'Non-integer X-Delete-After', resp.body)
|
|
|
|
def test_POST_negative_delete_after(self):
|
|
req = swob.Request.blank('/v1/a/c/o', method='POST',
|
|
headers={'Content-Type': 'foo/bar',
|
|
'X-Delete-After': '-60'})
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 400)
|
|
self.assertEqual(b'X-Delete-After in past', resp.body)
|
|
|
|
def test_PUT_negative_delete_after(self):
|
|
req = swob.Request.blank('/v1/a/c/o', method='PUT', body=b'',
|
|
headers={'Content-Type': 'foo/bar',
|
|
'X-Delete-After': '-60'})
|
|
with set_http_connect():
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 400)
|
|
self.assertEqual(b'X-Delete-After in past', resp.body)
|
|
|
|
def test_POST_delete_at_non_integer(self):
|
|
t = str(int(time.time() + 100)) + '.1'
|
|
req = swob.Request.blank('/v1/a/c/o', method='POST',
|
|
headers={'Content-Type': 'foo/bar',
|
|
'X-Delete-At': t})
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 400)
|
|
self.assertEqual(b'Non-integer X-Delete-At', resp.body)
|
|
|
|
def test_PUT_delete_at_non_integer(self):
|
|
t = str(int(time.time() - 100)) + '.1'
|
|
req = swob.Request.blank('/v1/a/c/o', method='PUT', body=b'',
|
|
headers={'Content-Type': 'foo/bar',
|
|
'X-Delete-At': t})
|
|
with set_http_connect():
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 400)
|
|
self.assertEqual(b'Non-integer X-Delete-At', resp.body)
|
|
|
|
def test_POST_delete_at_in_past(self):
|
|
t = str(int(time.time() - 100))
|
|
req = swob.Request.blank('/v1/a/c/o', method='POST',
|
|
headers={'Content-Type': 'foo/bar',
|
|
'X-Delete-At': t})
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 400)
|
|
self.assertEqual(b'X-Delete-At in past', resp.body)
|
|
|
|
def test_PUT_delete_at_in_past(self):
|
|
t = str(int(time.time() - 100))
|
|
req = swob.Request.blank('/v1/a/c/o', method='PUT', body=b'',
|
|
headers={'Content-Type': 'foo/bar',
|
|
'X-Delete-At': t})
|
|
with set_http_connect():
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 400)
|
|
self.assertEqual(b'X-Delete-At in past', resp.body)
|
|
|
|
def _test_x_open_expired(self, method, num_reqs, headers=None):
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method=method, headers=headers)
|
|
codes = [404] * num_reqs
|
|
with mocked_http_conn(*codes) as fake_conn:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
return fake_conn.requests
|
|
|
|
def test_x_open_expired_default_config(self):
|
|
for method, num_reqs in (
|
|
('GET',
|
|
self.obj_ring.replicas + self.obj_ring.max_more_nodes),
|
|
('HEAD',
|
|
self.obj_ring.replicas + self.obj_ring.max_more_nodes),
|
|
('POST', self.obj_ring.replicas)):
|
|
requests = self._test_x_open_expired(method, num_reqs)
|
|
for r in requests:
|
|
self.assertNotIn('X-Open-Expired', r['headers'])
|
|
self.assertNotIn('X-Backend-Open-Expired', r['headers'])
|
|
|
|
requests = self._test_x_open_expired(
|
|
method, num_reqs, headers={'X-Open-Expired': 'true'})
|
|
for r in requests:
|
|
self.assertEqual(r['headers']['X-Open-Expired'], 'true')
|
|
self.assertNotIn('X-Backend-Open-Expired', r['headers'])
|
|
|
|
requests = self._test_x_open_expired(
|
|
method, num_reqs, headers={'X-Open-Expired': 'false'})
|
|
for r in requests:
|
|
self.assertEqual(r['headers']['X-Open-Expired'], 'false')
|
|
self.assertNotIn('X-Backend-Open-Expired', r['headers'])
|
|
|
|
def test_x_open_expired_custom_config(self):
|
|
# helper to check that PUT is not supported in all cases
|
|
def test_put_unsupported():
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': '0',
|
|
'X-Open-Expired': 'true'})
|
|
codes = [201] * self.obj_ring.replicas
|
|
expect_headers = {
|
|
'X-Obj-Metadata-Footer': 'yes',
|
|
'X-Obj-Multiphase-Commit': 'yes'
|
|
}
|
|
with mocked_http_conn(
|
|
*codes, expect_headers=expect_headers) as fake_conn:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
for r in fake_conn.requests:
|
|
self.assertEqual(r['headers']['X-Open-Expired'], 'true')
|
|
self.assertNotIn('X-Backend-Open-Expired', r['headers'])
|
|
|
|
# Allow open expired
|
|
# Override app configuration
|
|
conf = {'allow_open_expired': 'true'}
|
|
# Create a new proxy instance for test with config
|
|
self.app = PatchedObjControllerApp(
|
|
conf, account_ring=FakeRing(),
|
|
container_ring=FakeRing(), logger=None)
|
|
# Use the same container info as the app used in other tests
|
|
self.app.container_info = dict(self.container_info)
|
|
self.obj_ring = self.app.get_object_ring(int(self.policy))
|
|
|
|
for method, num_reqs in (
|
|
('GET',
|
|
self.obj_ring.replicas + self.obj_ring.max_more_nodes),
|
|
('HEAD',
|
|
self.obj_ring.replicas + self.obj_ring.max_more_nodes),
|
|
('POST', self.obj_ring.replicas)):
|
|
requests = self._test_x_open_expired(
|
|
method, num_reqs, headers={'X-Open-Expired': 'true'})
|
|
for r in requests:
|
|
# If the proxy server config is has allow_open_expired set
|
|
# to true, then we set x-backend-open-expired to true
|
|
self.assertEqual(r['headers']['X-Open-Expired'], 'true')
|
|
self.assertEqual(r['headers']['X-Backend-Open-Expired'],
|
|
'true')
|
|
|
|
for method, num_reqs in (
|
|
('GET',
|
|
self.obj_ring.replicas + self.obj_ring.max_more_nodes),
|
|
('HEAD',
|
|
self.obj_ring.replicas + self.obj_ring.max_more_nodes),
|
|
('POST', self.obj_ring.replicas)):
|
|
requests = self._test_x_open_expired(
|
|
method, num_reqs, headers={'X-Open-Expired': 'false'})
|
|
for r in requests:
|
|
# If the proxy server config has allow_open_expired set
|
|
# to false, then we set x-backend-open-expired to false
|
|
self.assertEqual(r['headers']['X-Open-Expired'], 'false')
|
|
self.assertNotIn('X-Backend-Open-Expired', r['headers'])
|
|
|
|
# we don't support x-open-expired on PUT when allow_open_expired
|
|
test_put_unsupported()
|
|
|
|
# Disallow open expired
|
|
conf = {'allow_open_expired': 'false'}
|
|
# Create a new proxy instance for test with config
|
|
self.app = PatchedObjControllerApp(
|
|
conf, account_ring=FakeRing(),
|
|
container_ring=FakeRing(), logger=None)
|
|
# Use the same container info as the app used in other tests
|
|
self.app.container_info = dict(self.container_info)
|
|
self.obj_ring = self.app.get_object_ring(int(self.policy))
|
|
|
|
for method, num_reqs in (
|
|
('GET',
|
|
self.obj_ring.replicas + self.obj_ring.max_more_nodes),
|
|
('HEAD',
|
|
self.obj_ring.replicas + self.obj_ring.max_more_nodes),
|
|
('POST', self.obj_ring.replicas)):
|
|
# This case is different: we never add the 'X-Backend-Open-Expired'
|
|
# header if the proxy server config disables this feature
|
|
requests = self._test_x_open_expired(
|
|
method, num_reqs, headers={'X-Open-Expired': 'true'})
|
|
for r in requests:
|
|
self.assertEqual(r['headers']['X-Open-Expired'], 'true')
|
|
self.assertNotIn('X-Backend-Open-Expired', r['headers'])
|
|
|
|
# we don't support x-open-expired on PUT when not allow_open_expired
|
|
test_put_unsupported()
|
|
|
|
def test_HEAD_simple(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='HEAD')
|
|
with set_http_connect(200):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertIn('Accept-Ranges', resp.headers)
|
|
|
|
def test_HEAD_x_newest(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='HEAD',
|
|
headers={'X-Newest': 'true'})
|
|
with set_http_connect(*([200] * self.replicas())):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
def test_HEAD_x_newest_different_timestamps(self):
|
|
req = swob.Request.blank('/v1/a/c/o', method='HEAD',
|
|
headers={'X-Newest': 'true'})
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
timestamps = [next(ts) for i in range(self.replicas())]
|
|
newest_timestamp = timestamps[-1]
|
|
random.shuffle(timestamps)
|
|
backend_response_headers = [{
|
|
'X-Backend-Timestamp': t.internal,
|
|
'X-Timestamp': t.normal
|
|
} for t in timestamps]
|
|
with set_http_connect(*([200] * self.replicas()),
|
|
headers=backend_response_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['x-timestamp'], newest_timestamp.normal)
|
|
|
|
def test_HEAD_x_newest_with_two_vector_timestamps(self):
|
|
req = swob.Request.blank('/v1/a/c/o', method='HEAD',
|
|
headers={'X-Newest': 'true'})
|
|
ts = (utils.Timestamp.now(offset=offset)
|
|
for offset in itertools.count())
|
|
timestamps = [next(ts) for i in range(self.replicas())]
|
|
newest_timestamp = timestamps[-1]
|
|
random.shuffle(timestamps)
|
|
backend_response_headers = [{
|
|
'X-Backend-Timestamp': t.internal,
|
|
'X-Timestamp': t.normal
|
|
} for t in timestamps]
|
|
with set_http_connect(*([200] * self.replicas()),
|
|
headers=backend_response_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['x-backend-timestamp'],
|
|
newest_timestamp.internal)
|
|
|
|
def test_HEAD_x_newest_with_some_missing(self):
|
|
req = swob.Request.blank('/v1/a/c/o', method='HEAD',
|
|
headers={'X-Newest': 'true'})
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
request_count = self.app.request_node_count(self.obj_ring.replicas)
|
|
backend_response_headers = [{
|
|
'x-timestamp': next(ts).normal,
|
|
} for i in range(request_count)]
|
|
responses = [404] * (request_count - 1)
|
|
responses.append(200)
|
|
request_log = []
|
|
|
|
def capture_requests(ip, port, device, part, method, path,
|
|
headers=None, **kwargs):
|
|
req = {
|
|
'ip': ip,
|
|
'port': port,
|
|
'device': device,
|
|
'part': part,
|
|
'method': method,
|
|
'path': path,
|
|
'headers': headers,
|
|
}
|
|
request_log.append(req)
|
|
with set_http_connect(*responses,
|
|
headers=backend_response_headers,
|
|
give_connect=capture_requests):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
for req in request_log:
|
|
self.assertEqual(req['method'], 'HEAD')
|
|
self.assertEqual(req['path'], '/a/c/o')
|
|
|
|
def test_some_404s_and_507s(self):
|
|
self.policy.object_ring.max_more_nodes = (3 * self.replicas())
|
|
req = swob.Request.blank('/v1/a/c/o', method='HEAD')
|
|
responses = [StubResponse(
|
|
404, headers={'X-Backend-Timestamp': '2'})] * self.replicas()
|
|
responses += [StubResponse(507, headers={})] * (
|
|
self.policy.object_ring.max_more_nodes - self.replicas())
|
|
self.assertEqual(len(responses), 3 * self.replicas()) # sanity
|
|
|
|
def get_response(req):
|
|
return responses.pop(0)
|
|
|
|
with capture_http_requests(get_response):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
self.assertEqual(resp.headers['X-Backend-Timestamp'], '2')
|
|
|
|
def test_container_sync_delete(self):
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
test_indexes = [None] + [int(p) for p in POLICIES]
|
|
for policy_index in test_indexes:
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='DELETE', headers={
|
|
'X-Timestamp': next(ts).internal})
|
|
codes = [409] * self.obj_ring.replicas
|
|
ts_iter = itertools.repeat(next(ts).internal)
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 409)
|
|
|
|
def test_PUT_requires_length(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT')
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 411)
|
|
|
|
def test_container_update_backend_requests(self):
|
|
for policy in POLICIES:
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT',
|
|
headers={'Content-Length': '0',
|
|
'X-Backend-Storage-Policy-Index': int(policy)})
|
|
controller = self.controller_cls(self.app, 'a', 'c', 'o')
|
|
|
|
# This is the number of container updates we're doing, simulating
|
|
# 1 to 15 container replicas.
|
|
for num_containers in range(1, 16):
|
|
containers = [{'ip': '1.0.0.%s' % i,
|
|
'port': '60%s' % str(i).zfill(2),
|
|
'device': 'sdb'} for i in range(num_containers)]
|
|
|
|
backend_headers = controller._backend_requests(
|
|
req, self.replicas(policy), 1, containers)
|
|
|
|
# how many of the backend headers have a container update
|
|
n_container_updates = len(
|
|
[headers for headers in backend_headers
|
|
if 'X-Container-Partition' in headers])
|
|
|
|
# how many object-server PUTs can fail and still let the
|
|
# client PUT succeed
|
|
n_can_fail = self.replicas(policy) - self.quorum(policy)
|
|
n_expected_updates = (
|
|
n_can_fail + utils.quorum_size(num_containers))
|
|
|
|
# you get at least one update per container no matter what
|
|
n_expected_updates = max(
|
|
n_expected_updates, num_containers)
|
|
|
|
# you can't have more object requests with updates than you
|
|
# have object requests (the container stuff gets doubled up,
|
|
# but that's not important for purposes of durability)
|
|
n_expected_updates = min(
|
|
n_expected_updates, self.replicas(policy))
|
|
self.assertEqual(n_expected_updates, n_container_updates)
|
|
|
|
def test_delete_at_backend_requests(self):
|
|
t = str(int(time.time() + 100))
|
|
for policy in POLICIES:
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT',
|
|
headers={'Content-Length': '0',
|
|
'X-Backend-Storage-Policy-Index': int(policy),
|
|
'X-Delete-At': t})
|
|
controller = self.controller_cls(self.app, 'a', 'c', 'o')
|
|
|
|
for num_del_at_nodes in range(1, 16):
|
|
containers = [
|
|
{'ip': '2.0.0.%s' % i, 'port': '70%s' % str(i).zfill(2),
|
|
'device': 'sdc'} for i in range(num_del_at_nodes)]
|
|
del_at_nodes = [
|
|
{'ip': '1.0.0.%s' % i, 'port': '60%s' % str(i).zfill(2),
|
|
'device': 'sdb'} for i in range(num_del_at_nodes)]
|
|
|
|
backend_headers = controller._backend_requests(
|
|
req, self.replicas(policy), 1, containers,
|
|
delete_at_container='dac', delete_at_partition=2,
|
|
delete_at_nodes=del_at_nodes)
|
|
|
|
devices = []
|
|
hosts = []
|
|
part = ctr = 0
|
|
for given_headers in backend_headers:
|
|
self.assertEqual(given_headers.get('X-Delete-At'), t)
|
|
if 'X-Delete-At-Partition' in given_headers:
|
|
self.assertEqual(
|
|
given_headers.get('X-Delete-At-Partition'), '2')
|
|
part += 1
|
|
if 'X-Delete-At-Container' in given_headers:
|
|
self.assertEqual(
|
|
given_headers.get('X-Delete-At-Container'), 'dac')
|
|
ctr += 1
|
|
devices += (
|
|
list_from_csv(given_headers.get('X-Delete-At-Device')))
|
|
hosts += (
|
|
list_from_csv(given_headers.get('X-Delete-At-Host')))
|
|
|
|
# same as in test_container_update_backend_requests
|
|
n_can_fail = self.replicas(policy) - self.quorum(policy)
|
|
n_expected_updates = (
|
|
n_can_fail + utils.quorum_size(num_del_at_nodes))
|
|
|
|
n_expected_hosts = max(
|
|
n_expected_updates, num_del_at_nodes)
|
|
|
|
self.assertEqual(len(hosts), n_expected_hosts)
|
|
self.assertEqual(len(devices), n_expected_hosts)
|
|
|
|
# parts don't get doubled up, maximum is count of obj requests
|
|
n_expected_parts = min(
|
|
n_expected_hosts, self.replicas(policy))
|
|
self.assertEqual(part, n_expected_parts)
|
|
self.assertEqual(ctr, n_expected_parts)
|
|
|
|
# check that hosts are correct
|
|
self.assertEqual(
|
|
set(hosts),
|
|
set('%s:%s' % (h['ip'], h['port']) for h in del_at_nodes))
|
|
self.assertEqual(set(devices), set(('sdb',)))
|
|
|
|
def test_smooth_distributed_backend_requests(self):
|
|
t = str(int(time.time() + 100))
|
|
for policy in POLICIES:
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT',
|
|
headers={'Content-Length': '0',
|
|
'X-Backend-Storage-Policy-Index': int(policy),
|
|
'X-Delete-At': t})
|
|
controller = self.controller_cls(self.app, 'a', 'c', 'o')
|
|
|
|
for num_containers in range(1, 16):
|
|
containers = [
|
|
{'ip': '2.0.0.%s' % i, 'port': '70%s' % str(i).zfill(2),
|
|
'device': 'sdc'} for i in range(num_containers)]
|
|
del_at_nodes = [
|
|
{'ip': '1.0.0.%s' % i, 'port': '60%s' % str(i).zfill(2),
|
|
'device': 'sdb'} for i in range(num_containers)]
|
|
|
|
backend_headers = controller._backend_requests(
|
|
req, self.replicas(policy), 1, containers,
|
|
delete_at_container='dac', delete_at_partition=2,
|
|
delete_at_nodes=del_at_nodes)
|
|
|
|
# caculate no of expected updates, see
|
|
# test_container_update_backend_requests for explanation
|
|
n_expected_updates = min(max(
|
|
self.replicas(policy) - self.quorum(policy) +
|
|
utils.quorum_size(num_containers), num_containers),
|
|
self.replicas(policy))
|
|
|
|
# the first n_expected_updates servers should have received
|
|
# a container update
|
|
self.assertTrue(
|
|
all([h.get('X-Container-Partition')
|
|
for h in backend_headers[:n_expected_updates]]))
|
|
# the last n_expected_updates servers should have received
|
|
# the x-delete-at* headers
|
|
self.assertTrue(
|
|
all([h.get('X-Delete-At-Container')
|
|
for h in backend_headers[-n_expected_updates:]]))
|
|
|
|
def _check_write_affinity(
|
|
self, conf, policy_conf, policy, affinity_regions, affinity_count):
|
|
conf['policy_config'] = policy_conf
|
|
app = PatchedObjControllerApp(
|
|
conf, account_ring=FakeRing(),
|
|
container_ring=FakeRing(), logger=self.logger)
|
|
|
|
controller = self.controller_cls(app, 'a', 'c', 'o')
|
|
|
|
object_ring = app.get_object_ring(int(policy))
|
|
# make our fake ring have plenty of nodes, and not get limited
|
|
# artificially by the proxy max request node count
|
|
object_ring.max_more_nodes = 100
|
|
|
|
all_nodes = object_ring.get_part_nodes(1)
|
|
all_nodes.extend(object_ring.get_more_nodes(1))
|
|
|
|
# make sure we have enough local nodes (sanity)
|
|
all_local_nodes = [n for n in all_nodes if
|
|
n['region'] in affinity_regions]
|
|
self.assertGreaterEqual(len(all_local_nodes), affinity_count)
|
|
|
|
# finally, create the local_first_nodes iter and flatten it out
|
|
local_first_nodes = list(controller.iter_nodes_local_first(
|
|
object_ring, 1, Request.blank(''), policy))
|
|
|
|
# check that the required number of local nodes were moved up the order
|
|
node_regions = [node['region'] for node in local_first_nodes]
|
|
self.assertTrue(
|
|
all(r in affinity_regions for r in node_regions[:affinity_count]),
|
|
'Unexpected region found in local nodes, expected %s but got %s' %
|
|
(affinity_regions, node_regions))
|
|
return app
|
|
|
|
def test_write_affinity_not_configured(self):
|
|
# default is no write affinity so expect both regions 0 and 1
|
|
self._check_write_affinity({}, {}, POLICIES[0], [0, 1],
|
|
2 * self.replicas(POLICIES[0]))
|
|
self._check_write_affinity({}, {}, POLICIES[1], [0, 1],
|
|
2 * self.replicas(POLICIES[1]))
|
|
|
|
def test_write_affinity_proxy_server_config(self):
|
|
# without overrides policies use proxy-server config section options
|
|
conf = {'write_affinity_node_count': '1 * replicas',
|
|
'write_affinity': 'r0'}
|
|
self._check_write_affinity(conf, {}, POLICIES[0], [0],
|
|
self.replicas(POLICIES[0]))
|
|
self._check_write_affinity(conf, {}, POLICIES[1], [0],
|
|
self.replicas(POLICIES[1]))
|
|
|
|
def test_write_affinity_per_policy_config(self):
|
|
# check only per-policy configuration is sufficient
|
|
conf = {}
|
|
policy_conf = {'0': {'write_affinity_node_count': '1 * replicas',
|
|
'write_affinity': 'r1'},
|
|
'1': {'write_affinity_node_count': '5',
|
|
'write_affinity': 'r0'}}
|
|
self._check_write_affinity(conf, policy_conf, POLICIES[0], [1],
|
|
self.replicas(POLICIES[0]))
|
|
self._check_write_affinity(conf, policy_conf, POLICIES[1], [0], 5)
|
|
|
|
def test_write_affinity_per_policy_config_overrides_and_inherits(self):
|
|
# check per-policy config is preferred over proxy-server section config
|
|
conf = {'write_affinity_node_count': '1 * replicas',
|
|
'write_affinity': 'r0'}
|
|
policy_conf = {'0': {'write_affinity': 'r1'},
|
|
'1': {'write_affinity_node_count': '3 * replicas'}}
|
|
# policy 0 inherits default node count, override affinity to r1
|
|
self._check_write_affinity(conf, policy_conf, POLICIES[0], [1],
|
|
self.replicas(POLICIES[0]))
|
|
# policy 1 inherits default affinity to r0, overrides node count
|
|
self._check_write_affinity(conf, policy_conf, POLICIES[1], [0],
|
|
3 * self.replicas(POLICIES[1]))
|
|
|
|
# end of CommonObjectControllerMixin
|
|
|
|
|
|
@patch_policies()
|
|
class TestReplicatedObjController(CommonObjectControllerMixin,
|
|
unittest.TestCase):
|
|
|
|
controller_cls = obj.ReplicatedObjectController
|
|
|
|
def test_PUT_simple(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT')
|
|
req.headers['content-length'] = '0'
|
|
with set_http_connect(201, 201, 201):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_PUT_error_with_footers(self):
|
|
footers_callback = make_footers_callback(b'')
|
|
env = {'swift.callback.update_footers': footers_callback}
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
environ=env)
|
|
req.headers['content-length'] = '0'
|
|
codes = [503] * self.replicas()
|
|
expect_headers = {
|
|
'X-Obj-Metadata-Footer': 'yes'
|
|
}
|
|
|
|
with set_http_connect(*codes, expect_headers=expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
def _test_PUT_with_no_footers(self, test_body=b'', chunked=False):
|
|
# verify that when no footers are required then the PUT uses a regular
|
|
# single part body
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=test_body)
|
|
if chunked:
|
|
req.headers['Transfer-Encoding'] = 'chunked'
|
|
etag = md5(test_body, usedforsecurity=False).hexdigest()
|
|
req.headers['Etag'] = etag
|
|
|
|
put_requests = defaultdict(
|
|
lambda: {'headers': None, 'chunks': [], 'connection': None})
|
|
|
|
def capture_body(conn, chunk):
|
|
put_requests[conn.connection_id]['chunks'].append(chunk)
|
|
put_requests[conn.connection_id]['connection'] = conn
|
|
|
|
def capture_headers(ip, port, device, part, method, path, headers,
|
|
**kwargs):
|
|
conn_id = kwargs['connection_id']
|
|
put_requests[conn_id]['headers'] = headers
|
|
|
|
codes = [201] * self.replicas()
|
|
expect_headers = {'X-Obj-Metadata-Footer': 'yes'}
|
|
resp_headers = {
|
|
'Some-Header': 'Four',
|
|
'Etag': '"%s"' % etag,
|
|
}
|
|
with set_http_connect(*codes, expect_headers=expect_headers,
|
|
give_send=capture_body,
|
|
give_connect=capture_headers,
|
|
headers=resp_headers):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
timestamps = {captured_req['headers']['x-timestamp']
|
|
for captured_req in put_requests.values()}
|
|
self.assertEqual(1, len(timestamps), timestamps)
|
|
self.assertEqual(dict(resp.headers), {
|
|
'Content-Type': 'text/html; charset=UTF-8',
|
|
'Content-Length': '0',
|
|
'Etag': etag,
|
|
'Last-Modified': time.strftime(
|
|
"%a, %d %b %Y %H:%M:%S GMT",
|
|
time.gmtime(math.ceil(float(timestamps.pop())))),
|
|
})
|
|
for connection_id, info in put_requests.items():
|
|
body = b''.join(info['chunks'])
|
|
headers = info['headers']
|
|
if chunked or not test_body:
|
|
body = unchunk_body(body)
|
|
self.assertEqual('100-continue', headers['Expect'])
|
|
self.assertEqual('chunked', headers['Transfer-Encoding'])
|
|
self.assertNotIn('Content-Length', headers)
|
|
else:
|
|
self.assertNotIn('Transfer-Encoding', headers)
|
|
if body or not test_body:
|
|
self.assertEqual('100-continue', headers['Expect'])
|
|
else:
|
|
self.assertNotIn('Expect', headers)
|
|
self.assertNotIn('X-Backend-Obj-Multipart-Mime-Boundary', headers)
|
|
self.assertNotIn('X-Backend-Obj-Metadata-Footer', headers)
|
|
self.assertNotIn('X-Backend-Obj-Multiphase-Commit', headers)
|
|
self.assertEqual(etag, headers['Etag'])
|
|
|
|
self.assertEqual(test_body, body)
|
|
self.assertTrue(info['connection'].closed)
|
|
|
|
def test_PUT_with_chunked_body_and_no_footers(self):
|
|
self._test_PUT_with_no_footers(test_body=b'asdf', chunked=True)
|
|
|
|
def test_PUT_with_body_and_no_footers(self):
|
|
self._test_PUT_with_no_footers(test_body=b'asdf', chunked=False)
|
|
|
|
def test_PUT_with_no_body_and_no_footers(self):
|
|
self._test_PUT_with_no_footers(test_body=b'', chunked=False)
|
|
|
|
def test_txn_id_logging_on_PUT(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT')
|
|
self.app.logger.txn_id = req.environ['swift.trans_id'] = 'test-txn-id'
|
|
req.headers['content-length'] = '0'
|
|
# we capture stdout since the debug log formatter prints the formatted
|
|
# message to stdout
|
|
stdout = StringIO()
|
|
with set_http_connect((100, Timeout()), 503, 503), \
|
|
mock.patch('sys.stdout', stdout):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
for line in stdout.getvalue().splitlines():
|
|
self.assertIn('test-txn-id', line)
|
|
self.assertIn('Trying to get final status of PUT to',
|
|
stdout.getvalue())
|
|
|
|
def test_PUT_empty_bad_etag(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT')
|
|
req.headers['Content-Length'] = '0'
|
|
req.headers['Etag'] = '"catbus"'
|
|
|
|
# The 2-tuple here makes getexpect() return 422, not 100. For objects
|
|
# that are >0 bytes, you get a 100 Continue and then a 422
|
|
# Unprocessable Entity after sending the body. For zero-byte objects,
|
|
# though, you get the 422 right away because no Expect header is sent
|
|
# with zero-byte PUT. The second status in the tuple should not be
|
|
# consumed, it's just there to make the FakeStatus treat the first as
|
|
# an expect status, but we'll make it something other than a 422 so
|
|
# that if it is consumed then the test should fail.
|
|
codes = [FakeStatus((422, 200))
|
|
for _junk in range(self.replicas())]
|
|
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 422)
|
|
|
|
def test_PUT_if_none_match(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT')
|
|
req.headers['if-none-match'] = '*'
|
|
req.headers['content-length'] = '0'
|
|
with set_http_connect(201, 201, 201):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_PUT_if_none_match_denied(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT')
|
|
req.headers['if-none-match'] = '*'
|
|
req.headers['content-length'] = '0'
|
|
with set_http_connect(201, 412, 201):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 412)
|
|
|
|
def test_PUT_if_none_match_not_star(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT')
|
|
req.headers['if-none-match'] = 'somethingelse'
|
|
req.headers['content-length'] = '0'
|
|
with set_http_connect():
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 400)
|
|
|
|
def test_PUT_connect_exceptions(self):
|
|
object_ring = self.app.get_object_ring(None)
|
|
self.app.sort_nodes = lambda n, *args, **kwargs: n # disable shuffle
|
|
|
|
def test_status_map(statuses, expected):
|
|
self.app.error_limiter.stats.clear()
|
|
req = swob.Request.blank('/v1/a/c/o.jpg', method='PUT',
|
|
body=b'test body')
|
|
with set_http_connect(*statuses):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, expected)
|
|
|
|
base_status = [201] * 3
|
|
# test happy path
|
|
test_status_map(list(base_status), 201)
|
|
for i in range(3):
|
|
self.assertEqual(node_error_count(
|
|
self.app, object_ring.devs[i]), 0)
|
|
# single node errors and test isolation
|
|
for i in range(3):
|
|
status_list = list(base_status)
|
|
status_list[i] = 503
|
|
test_status_map(status_list, 201)
|
|
for j in range(3):
|
|
self.assertEqual(node_error_count(
|
|
self.app, object_ring.devs[j]), 1 if j == i else 0)
|
|
# connect errors
|
|
test_status_map((201, Timeout(), 201, 201), 201)
|
|
self.assertEqual(node_error_count(
|
|
self.app, object_ring.devs[1]), 1)
|
|
test_status_map((Exception('kaboom!'), 201, 201, 201), 201)
|
|
self.assertEqual(node_error_count(
|
|
self.app, object_ring.devs[0]), 1)
|
|
# expect errors
|
|
test_status_map((201, 201, (503, None), 201), 201)
|
|
self.assertEqual(node_error_count(
|
|
self.app, object_ring.devs[2]), 1)
|
|
test_status_map(((507, None), 201, 201, 201), 201)
|
|
self.assertEqual(
|
|
node_error_count(self.app, object_ring.devs[0]),
|
|
self.app.error_limiter.suppression_limit + 1)
|
|
# response errors
|
|
test_status_map(((100, Timeout()), 201, 201), 201)
|
|
self.assertEqual(
|
|
node_error_count(self.app, object_ring.devs[0]), 1)
|
|
test_status_map((201, 201, (100, Exception())), 201)
|
|
self.assertEqual(
|
|
node_error_count(self.app, object_ring.devs[2]), 1)
|
|
test_status_map((201, (100, 507), 201), 201)
|
|
self.assertEqual(
|
|
node_error_count(self.app, object_ring.devs[1]),
|
|
self.app.error_limiter.suppression_limit + 1)
|
|
|
|
def test_PUT_connect_exception_with_unicode_path(self):
|
|
expected = 201
|
|
statuses = (
|
|
Exception('Connection refused: Please insert ten dollars'),
|
|
201, 201, 201)
|
|
|
|
req = swob.Request.blank('/v1/AUTH_kilroy/%ED%88%8E/%E9%90%89',
|
|
method='PUT',
|
|
body=b'life is utf-gr8')
|
|
self.app.logger.clear()
|
|
with set_http_connect(*statuses):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, expected)
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertFalse(log_lines[1:])
|
|
self.assertIn('ERROR with Object server', log_lines[0])
|
|
self.assertIn(quote(req.swift_entity_path), log_lines[0])
|
|
self.assertIn('re: Expect: 100-continue', log_lines[0])
|
|
|
|
def test_PUT_get_expect_errors_with_unicode_path(self):
|
|
def do_test(statuses):
|
|
req = swob.Request.blank('/v1/AUTH_kilroy/%ED%88%8E/%E9%90%89',
|
|
method='PUT',
|
|
body=b'life is utf-gr8')
|
|
self.app.logger.clear()
|
|
with set_http_connect(*statuses):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertFalse(log_lines[1:])
|
|
return log_lines
|
|
|
|
log_lines = do_test((201, (507, None), 201, 201))
|
|
self.assertIn('ERROR Insufficient Storage', log_lines[0])
|
|
|
|
log_lines = do_test((201, (503, None), 201, 201))
|
|
self.assertIn('ERROR 503 Expect: 100-continue From Object Server',
|
|
log_lines[0])
|
|
|
|
def test_PUT_send_exception_with_unicode_path(self):
|
|
def do_test(exc):
|
|
conns = set()
|
|
|
|
def capture_send(conn, data):
|
|
conns.add(conn)
|
|
if len(conns) == 2:
|
|
raise exc
|
|
|
|
req = swob.Request.blank('/v1/AUTH_kilroy/%ED%88%8E/%E9%90%89',
|
|
method='PUT',
|
|
body=b'life is utf-gr8')
|
|
self.app.logger.clear()
|
|
with set_http_connect(201, 201, 201, give_send=capture_send):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertFalse(log_lines[1:])
|
|
self.assertIn('ERROR with Object server', log_lines[0])
|
|
self.assertIn(quote(req.swift_entity_path), log_lines[0])
|
|
self.assertIn('Trying to write to', log_lines[0])
|
|
|
|
do_test(Exception('Exception while sending data on connection'))
|
|
do_test(ChunkWriteTimeout())
|
|
|
|
def test_PUT_final_response_errors_with_unicode_path(self):
|
|
def do_test(statuses):
|
|
req = swob.Request.blank('/v1/AUTH_kilroy/%ED%88%8E/%E9%90%89',
|
|
method='PUT',
|
|
body=b'life is utf-gr8')
|
|
self.app.logger.clear()
|
|
with set_http_connect(*statuses):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertFalse(log_lines[1:])
|
|
return req, log_lines
|
|
|
|
req, log_lines = do_test((201, (100, Exception('boom')), 201))
|
|
self.assertIn('ERROR with Object server', log_lines[0])
|
|
if six.PY3:
|
|
self.assertIn(req.path, log_lines[0])
|
|
else:
|
|
self.assertIn(req.path.decode('utf-8'), log_lines[0])
|
|
self.assertIn('Trying to get final status of PUT', log_lines[0])
|
|
|
|
req, log_lines = do_test((201, (100, Timeout()), 201))
|
|
self.assertIn('ERROR with Object server', log_lines[0])
|
|
if six.PY3:
|
|
self.assertIn(req.path, log_lines[0])
|
|
else:
|
|
self.assertIn(req.path.decode('utf-8'), log_lines[0])
|
|
self.assertIn('Trying to get final status of PUT', log_lines[0])
|
|
|
|
req, log_lines = do_test((201, (100, 507), 201))
|
|
self.assertIn('ERROR Insufficient Storage', log_lines[0])
|
|
|
|
req, log_lines = do_test((201, (100, 500), 201))
|
|
if six.PY3:
|
|
# We allow the b'' in logs because we want to see bad characters.
|
|
self.assertIn(
|
|
"ERROR 500 b'' Trying to PUT /v1/AUTH_kilroy/%ED%88%8E/"
|
|
"%E9%90%89 From Object Server", log_lines[0])
|
|
self.assertIn(req.path, log_lines[0])
|
|
else:
|
|
self.assertIn(
|
|
'ERROR 500 Trying to PUT /v1/AUTH_kilroy/%ED%88%8E/%E9%90%89 '
|
|
'From Object Server', log_lines[0])
|
|
self.assertIn(req.path.decode('utf-8'), log_lines[0])
|
|
|
|
def test_DELETE_errors(self):
|
|
# verify logged errors with and without non-ascii characters in path
|
|
def do_test(path, statuses):
|
|
|
|
req = swob.Request.blank('/v1' + path,
|
|
method='DELETE',
|
|
body=b'life is utf-gr8')
|
|
self.app.logger.clear()
|
|
with set_http_connect(*statuses):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertFalse(log_lines[1:])
|
|
return req, log_lines
|
|
|
|
req, log_lines = do_test('/AUTH_kilroy/ascii/ascii',
|
|
(201, 500, 201, 201))
|
|
self.assertIn('Trying to DELETE', log_lines[0])
|
|
if six.PY3:
|
|
self.assertIn(req.swift_entity_path, log_lines[0])
|
|
else:
|
|
self.assertIn(req.swift_entity_path.decode('utf-8'), log_lines[0])
|
|
self.assertIn(' From Object Server', log_lines[0])
|
|
|
|
req, log_lines = do_test('/AUTH_kilroy/%ED%88%8E/%E9%90%89',
|
|
(201, 500, 201, 201))
|
|
self.assertIn('Trying to DELETE', log_lines[0])
|
|
if six.PY3:
|
|
self.assertIn(req.swift_entity_path, log_lines[0])
|
|
else:
|
|
self.assertIn(req.swift_entity_path.decode('utf-8'), log_lines[0])
|
|
self.assertIn(' From Object Server', log_lines[0])
|
|
|
|
req, log_lines = do_test('/AUTH_kilroy/ascii/ascii',
|
|
(201, 507, 201, 201))
|
|
self.assertIn('ERROR Insufficient Storage', log_lines[0])
|
|
|
|
req, log_lines = do_test('/AUTH_kilroy/%ED%88%8E/%E9%90%89',
|
|
(201, 507, 201, 201))
|
|
self.assertIn('ERROR Insufficient Storage', log_lines[0])
|
|
|
|
req, log_lines = do_test('/AUTH_kilroy/ascii/ascii',
|
|
(201, Exception(), 201, 201))
|
|
self.assertIn('Trying to DELETE', log_lines[0])
|
|
if six.PY3:
|
|
self.assertIn(req.swift_entity_path, log_lines[0])
|
|
else:
|
|
self.assertIn(req.swift_entity_path.decode('utf-8'), log_lines[0])
|
|
self.assertIn('ERROR with Object server', log_lines[0])
|
|
|
|
req, log_lines = do_test('/AUTH_kilroy/%ED%88%8E/%E9%90%89',
|
|
(201, Exception(), 201, 201))
|
|
self.assertIn('Trying to DELETE', log_lines[0])
|
|
if six.PY3:
|
|
self.assertIn(req.swift_entity_path, log_lines[0])
|
|
else:
|
|
self.assertIn(req.swift_entity_path.decode('utf-8'), log_lines[0])
|
|
self.assertIn('ERROR with Object server', log_lines[0])
|
|
|
|
def test_DELETE_with_write_affinity(self):
|
|
policy_conf = self.app.get_policy_options(self.policy)
|
|
policy_conf.write_affinity_handoff_delete_count = self.replicas() // 2
|
|
policy_conf.write_affinity_is_local_fn = (
|
|
lambda node: node['region'] == 1)
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
|
|
codes = [204, 204, 404, 204]
|
|
with mocked_http_conn(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
codes = [204, 404, 404, 204]
|
|
with mocked_http_conn(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
policy_conf.write_affinity_handoff_delete_count = 2
|
|
|
|
codes = [204, 204, 404, 204, 404]
|
|
with mocked_http_conn(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
codes = [204, 404, 404, 204, 204]
|
|
with mocked_http_conn(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
def test_PUT_error_during_transfer_data(self):
|
|
class FakeReader(object):
|
|
def read(self, size):
|
|
raise IOError('error message')
|
|
|
|
req = swob.Request.blank('/v1/a/c/o.jpg', method='PUT',
|
|
body=b'test body')
|
|
|
|
req.environ['wsgi.input'] = FakeReader()
|
|
req.headers['content-length'] = '6'
|
|
with set_http_connect(201, 201, 201):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 499)
|
|
|
|
def test_PUT_chunkreadtimeout_during_transfer_data(self):
|
|
class FakeReader(object):
|
|
def read(self, size):
|
|
raise exceptions.ChunkReadTimeout()
|
|
|
|
req = swob.Request.blank('/v1/a/c/o.jpg', method='PUT',
|
|
body=b'test body')
|
|
|
|
req.environ['wsgi.input'] = FakeReader()
|
|
req.headers['content-length'] = '6'
|
|
with set_http_connect(201, 201, 201):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 408)
|
|
|
|
def test_PUT_timeout_during_transfer_data(self):
|
|
class FakeReader(object):
|
|
def read(self, size):
|
|
raise Timeout()
|
|
conns = []
|
|
|
|
def capture_expect(conn):
|
|
# stash connections so that we can verify they all get closed
|
|
conns.append(conn)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o.jpg', method='PUT',
|
|
body=b'test body')
|
|
|
|
req.environ['wsgi.input'] = FakeReader()
|
|
req.headers['content-length'] = '6'
|
|
with set_http_connect(201, 201, 201, give_expect=capture_expect):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 499)
|
|
self.assertEqual(self.replicas(), len(conns))
|
|
for conn in conns:
|
|
self.assertTrue(conn.closed)
|
|
|
|
def test_PUT_insufficient_data_from_client(self):
|
|
class FakeReader(object):
|
|
def read(self, size):
|
|
raise Timeout()
|
|
conns = []
|
|
|
|
def capture_expect(conn):
|
|
# stash connections so that we can verify they all get closed
|
|
conns.append(conn)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o.jpg', method='PUT',
|
|
body='7 bytes')
|
|
req.headers['content-length'] = '99'
|
|
with set_http_connect(201, 201, 201, give_expect=capture_expect):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 499)
|
|
warning_lines = self.app.logger.get_lines_for_level('warning')
|
|
self.assertEqual(1, len(warning_lines))
|
|
self.assertIn('Client disconnected without sending enough data',
|
|
warning_lines[0])
|
|
self.assertEqual(self.replicas(), len(conns))
|
|
for conn in conns:
|
|
self.assertTrue(conn.closed)
|
|
|
|
def test_PUT_exception_during_transfer_data(self):
|
|
class FakeReader(object):
|
|
def read(self, size):
|
|
raise Exception('exception message')
|
|
|
|
req = swob.Request.blank('/v1/a/c/o.jpg', method='PUT',
|
|
body=b'test body')
|
|
|
|
req.environ['wsgi.input'] = FakeReader()
|
|
req.headers['content-length'] = '6'
|
|
with set_http_connect(201, 201, 201):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 500)
|
|
|
|
def test_GET_simple(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
with set_http_connect(200, headers={'Connection': 'close'}):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertIn('Accept-Ranges', resp.headers)
|
|
self.assertNotIn('Connection', resp.headers)
|
|
|
|
def test_GET_slow_read(self):
|
|
self.app.recoverable_node_timeout = 0.01
|
|
self.app.client_timeout = 0.1
|
|
self.app.object_chunk_size = 10
|
|
body = b'test'
|
|
etag = md5(body, usedforsecurity=False).hexdigest()
|
|
headers = {
|
|
'Etag': etag,
|
|
'Content-Length': len(body),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
responses = [(200, body, headers)] * 2
|
|
status_codes, body_iter, headers = zip(*responses)
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
# make the first response slow...
|
|
read_sleeps = [0.1, 0]
|
|
with mocked_http_conn(*status_codes, body_iter=body_iter,
|
|
headers=headers, slow=read_sleeps) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
body = resp.body
|
|
self.assertEqual(b'test', body)
|
|
self.assertEqual(len(log.requests), 2)
|
|
|
|
def make_key(r):
|
|
r['device'] = r['path'].split('/')[1]
|
|
return '%(ip)s:%(port)s/%(device)s' % r
|
|
# the first node got errors incr'd
|
|
expected_error_limiting = {
|
|
make_key(log.requests[0]): {
|
|
'errors': 1,
|
|
'last_error': mock.ANY,
|
|
}
|
|
}
|
|
actual = {}
|
|
for n in self.app.get_object_ring(int(self.policy)).devs:
|
|
node_key = self.app.error_limiter.node_key(n)
|
|
stats = self.app.error_limiter.stats.get(node_key) or {}
|
|
if stats:
|
|
actual[self.app.error_limiter.node_key(n)] = stats
|
|
self.assertEqual(actual, expected_error_limiting)
|
|
for read_line in self.app.logger.get_lines_for_level('error'):
|
|
self.assertIn("Trying to read object during GET (retrying)",
|
|
read_line)
|
|
self.assertEqual(
|
|
len(self.logger.logger.records['ERROR']), 1,
|
|
'Expected 1 ERROR lines, got %r' % (
|
|
self.logger.logger.records['ERROR'], ))
|
|
|
|
def _do_test_GET_with_multirange_slow_body_resumes(
|
|
self, slowdown_after=0, resume_bytes=0):
|
|
self.app.logger.clear()
|
|
self.app.recoverable_node_timeout = 0.01
|
|
self.app.object_chunk_size = 10
|
|
obj_data = b''.join([b'testing%03d' % i for i in range(100)])
|
|
etag = md5(obj_data, usedforsecurity=False).hexdigest()
|
|
boundary1 = b'81eb9c110b32ced5fe'
|
|
resp_body1 = b'\r\n'.join([
|
|
b'--' + boundary1,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 0-49/700',
|
|
b'',
|
|
obj_data[0:50],
|
|
b'--' + boundary1,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 100-104/700',
|
|
b'',
|
|
obj_data[100:105],
|
|
b'--' + boundary1 + b'--',
|
|
])
|
|
boundary2 = b'aaeb9c110b32ced5fe'
|
|
resp_body2 = b'\r\n'.join([
|
|
b'--' + boundary2,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes %d-49/700' % resume_bytes,
|
|
b'',
|
|
obj_data[resume_bytes:50],
|
|
b'--' + boundary2,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 100-104/700',
|
|
b'',
|
|
obj_data[100:105],
|
|
b'--' + boundary2 + b'--',
|
|
])
|
|
|
|
headers1 = {
|
|
'Etag': etag,
|
|
'Content-Type': b'multipart/byteranges;boundary=' + boundary1,
|
|
'Content-Length': len(resp_body1),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
headers2 = {
|
|
'Etag': etag,
|
|
'Content-Type': b'multipart/byteranges;boundary=' + boundary2,
|
|
'Content-Length': len(resp_body2),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
responses = [
|
|
StubResponse(206, resp_body1, headers1, slowdown=0.1,
|
|
slowdown_after=slowdown_after),
|
|
StubResponse(206, resp_body2, headers2)
|
|
]
|
|
req_range_hdrs = []
|
|
|
|
def get_response(req):
|
|
req_range_hdrs.append(req['headers'].get('Range'))
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=0-49,100-104'})
|
|
with capture_http_requests(get_response) as captured_requests:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 206)
|
|
actual_body = resp.body
|
|
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(2, len(captured_requests))
|
|
self.assertEqual([1] + [0] * (self.replicas() - 1),
|
|
node_error_counts(self.app, self.obj_ring.devs))
|
|
# note: client response uses boundary from first backend response
|
|
self.assertEqual(resp_body1, actual_body)
|
|
return req_range_hdrs
|
|
|
|
def test_GET_with_multirange_slow_body_resumes(self):
|
|
req_range_hdrs = self._do_test_GET_with_multirange_slow_body_resumes(
|
|
slowdown_after=0)
|
|
self.assertEqual(['bytes=0-49,100-104'] * 2, req_range_hdrs)
|
|
error_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(1, len(error_lines))
|
|
self.assertIn('Trying to read next part of object multi-part GET '
|
|
'(retrying)', error_lines[0])
|
|
|
|
def test_GET_with_multirange_slow_body_resumes_before_body_started(self):
|
|
# First response times out while first part boundary/headers are being
|
|
# read. No part body has been yielded to the client so range header is
|
|
# not adjusted for the second backend request.
|
|
req_range_hdrs = self._do_test_GET_with_multirange_slow_body_resumes(
|
|
slowdown_after=40, resume_bytes=0)
|
|
self.assertEqual(['bytes=0-49,100-104'] * 2, req_range_hdrs)
|
|
error_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(1, len(error_lines))
|
|
self.assertIn('Trying to read next part of object multi-part GET '
|
|
'(retrying)', error_lines[0])
|
|
|
|
def test_GET_with_multirange_slow_body_resumes_after_body_started(self):
|
|
# First response times out after first part boundary/headers have been
|
|
# read. Some part body has been yielded to the client so range header
|
|
# is adjusted for the second backend request.
|
|
# 140 bytes before timeout is sufficient for the part boundary, headers
|
|
# and approx 50 body bytes to be read, but _MultipartMimeFileLikeObject
|
|
# buffers bytes from the backend response such that only 20 bytes are
|
|
# actually yielded to the client.
|
|
req_range_hdrs = self._do_test_GET_with_multirange_slow_body_resumes(
|
|
slowdown_after=140, resume_bytes=20)
|
|
self.assertEqual(['bytes=0-49,100-104', 'bytes=20-49,100-104'],
|
|
req_range_hdrs)
|
|
error_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(1, len(error_lines))
|
|
self.assertIn('Trying to read object during GET (retrying) ',
|
|
error_lines[0])
|
|
|
|
def test_GET_with_multirange_slow_body_unable_to_resume(self):
|
|
self.app.recoverable_node_timeout = 0.01
|
|
obj_data = b'testing' * 100
|
|
etag = md5(obj_data, usedforsecurity=False).hexdigest()
|
|
boundary = b'81eb9c110b32ced5fe'
|
|
|
|
resp_body = b'\r\n'.join([
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 0-49/700',
|
|
b'',
|
|
obj_data[0:50],
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 100-104/700',
|
|
b'',
|
|
obj_data[100:105],
|
|
b'--' + boundary + b'--',
|
|
])
|
|
|
|
headers = {
|
|
'Etag': etag,
|
|
'Content-Type': b'multipart/byteranges;boundary=' + boundary,
|
|
'Content-Length': len(resp_body),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
responses = [
|
|
StubResponse(206, resp_body, headers, slowdown=0.1),
|
|
StubResponse(206, resp_body, headers, slowdown=0.1),
|
|
StubResponse(206, resp_body, headers, slowdown=0.1),
|
|
]
|
|
|
|
def get_response(req):
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=0-49,100-104'})
|
|
response_chunks = []
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
with self.assertRaises(ChunkReadTimeout):
|
|
# note: the error is raised while the resp_iter is read...
|
|
for chunk in resp.app_iter:
|
|
response_chunks.append(chunk)
|
|
self.assertEqual(response_chunks, [])
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual([1, 1, 1],
|
|
node_error_counts(self.app, self.obj_ring.devs))
|
|
self.assertEqual(6, len(log))
|
|
error_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(3, len(error_lines))
|
|
for line in error_lines:
|
|
self.assertIn('Trying to read next part of object multi-part GET '
|
|
'(retrying)', line)
|
|
|
|
def test_GET_unable_to_resume(self):
|
|
self.app.recoverable_node_timeout = 0.01
|
|
self.app.client_timeout = 0.1
|
|
self.app.object_chunk_size = 10
|
|
resp_body = b'length 8'
|
|
etag = md5(resp_body, usedforsecurity=False).hexdigest()
|
|
headers = {
|
|
'Etag': etag,
|
|
'Content-Type': b'plain/text',
|
|
'Content-Length': len(resp_body),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
|
|
# make all responses slow...
|
|
responses = [
|
|
StubResponse(200, resp_body, headers, slowdown=0.1),
|
|
StubResponse(200, resp_body, headers, slowdown=0.1),
|
|
StubResponse(200, resp_body, headers, slowdown=0.1),
|
|
]
|
|
|
|
def get_response(req):
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(get_response):
|
|
resp = req.get_response(self.app)
|
|
with self.assertRaises(ChunkReadTimeout):
|
|
_ = resp.body
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(etag, resp.headers.get('ETag'))
|
|
self.assertEqual([1] * self.replicas(),
|
|
node_error_counts(self.app, self.obj_ring.devs))
|
|
|
|
error_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(3, len(error_lines))
|
|
for line in error_lines[:3]:
|
|
self.assertIn('Trying to read object during GET', line)
|
|
|
|
def test_GET_newest_will_not_resume(self):
|
|
# verify that request with x-newest use node_timeout and don't resume
|
|
self.app.node_timeout = 0.01
|
|
# set recoverable_node_timeout crazy high to verify that this is not
|
|
# the timeout value that is used
|
|
self.app.recoverable_node_timeout = 1000
|
|
self.app.client_timeout = 0.1
|
|
self.app.object_chunk_size = 10
|
|
resp_body = b'length 8'
|
|
etag = md5(resp_body, usedforsecurity=False).hexdigest()
|
|
headers = {
|
|
'Etag': etag,
|
|
'Content-Type': b'plain/text',
|
|
'Content-Length': len(resp_body),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
|
|
# make all responses slow...
|
|
responses = [
|
|
StubResponse(200, resp_body, headers, slowdown=0.1),
|
|
]
|
|
|
|
def get_response(req):
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={'X-Newest': 'true'})
|
|
with capture_http_requests(get_response):
|
|
resp = req.get_response(self.app)
|
|
with self.assertRaises(ChunkReadTimeout):
|
|
_ = resp.body
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(etag, resp.headers.get('ETag'))
|
|
error_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(0, len(error_lines))
|
|
|
|
def test_GET_resuming_ignores_416(self):
|
|
# verify that a resuming getter will not try to use the content of a
|
|
# 416 response (because it's etag will mismatch that from the first
|
|
# response)
|
|
self.app.recoverable_node_timeout = 0.01
|
|
self.app.client_timeout = 0.1
|
|
self.app.object_chunk_size = 10
|
|
body = b'length 8'
|
|
body_short = b'four'
|
|
body_416 = b'<html><h1>Requested Range Not Satisfiable</h1>' \
|
|
b'<p>The Range requested is not available.</p></html>'
|
|
etag = md5(body, usedforsecurity=False).hexdigest()
|
|
etag_short = md5(body_short, usedforsecurity=False).hexdigest()
|
|
headers_206 = {
|
|
'Etag': etag,
|
|
'Content-Length': len(body),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
'Content-Range': 'bytes 7-8/8'
|
|
}
|
|
headers_416 = {
|
|
# note: 416 when applying the same range implies different object
|
|
# length and therefore different etag
|
|
'Etag': etag_short,
|
|
'Content-Length': len(body_416),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
'Content-Range': 'bytes */4'
|
|
}
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', headers={'Range': 'bytes=7-8'})
|
|
# make the first response slow...
|
|
read_sleeps = [0.1, 0]
|
|
with mocked_http_conn(206, 416, 206, body_iter=[body, body_416, body],
|
|
headers=[headers_206, headers_416, headers_206],
|
|
slow=read_sleeps) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 206)
|
|
resp_body = resp.body
|
|
self.assertEqual(b'length 8', resp_body)
|
|
self.assertEqual(len(log.requests), 3)
|
|
self.assertEqual('bytes=7-8', log.requests[0]['headers']['Range'])
|
|
self.assertEqual('bytes=7-8', log.requests[1]['headers']['Range'])
|
|
self.assertEqual('bytes=7-8', log.requests[2]['headers']['Range'])
|
|
|
|
def test_GET_resuming(self):
|
|
self.app.recoverable_node_timeout = 0.01
|
|
self.app.client_timeout = 0.1
|
|
self.app.object_chunk_size = 10
|
|
body = b'length 8'
|
|
etag = md5(body, usedforsecurity=False).hexdigest()
|
|
headers_200 = {
|
|
'Etag': etag,
|
|
'Content-Length': len(body),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
headers_206 = {
|
|
# note: use of 'X-Backend-Ignore-Range-If-Metadata-Present' in
|
|
# request means that 200 response did not evaluate the Range and
|
|
# the proxy modifies requested backend range accordingly
|
|
'Etag': etag,
|
|
'Content-Length': len(body),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
'Content-Range': 'bytes 0-7/8'
|
|
}
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o',
|
|
headers={'Range': 'bytes=9-10, 20-30',
|
|
'X-Backend-Ignore-Range-If-Metadata-Present':
|
|
'X-Static-Large-Object'})
|
|
# make the first 2 responses slow...
|
|
read_sleeps = [0.1, 0.1, 0]
|
|
with mocked_http_conn(200, 206, 206, body_iter=[body, body, body],
|
|
headers=[headers_200, headers_206, headers_206],
|
|
slow=read_sleeps) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
resp_body = resp.body
|
|
self.assertEqual(b'length 8', resp_body)
|
|
self.assertEqual(len(log.requests), 3)
|
|
# NB: original range is not satisfiable but is ignored
|
|
self.assertEqual('bytes=9-10, 20-30',
|
|
log.requests[0]['headers']['Range'])
|
|
self.assertIn('X-Backend-Ignore-Range-If-Metadata-Present',
|
|
log.requests[0]['headers'])
|
|
# backend Range is updated to something that is satisfiable
|
|
self.assertEqual('bytes=0-7,20-30',
|
|
log.requests[1]['headers']['Range'])
|
|
self.assertNotIn('X-Backend-Ignore-Range-If-Metadata-Present',
|
|
log.requests[1]['headers'])
|
|
self.assertEqual('bytes=0-7,20-30',
|
|
log.requests[2]['headers']['Range'])
|
|
self.assertNotIn('X-Backend-Ignore-Range-If-Metadata-Present',
|
|
log.requests[2]['headers'])
|
|
self.assertEqual([1, 1] + [0] * (self.replicas() - 2),
|
|
node_error_counts(self.app, self.obj_ring.devs))
|
|
|
|
def test_GET_transfer_encoding_chunked(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
with set_http_connect(200, headers={'transfer-encoding': 'chunked'}):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['Transfer-Encoding'], 'chunked')
|
|
|
|
def _test_removes_swift_bytes(self, method):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method=method)
|
|
with set_http_connect(
|
|
200, headers={'content-type': 'image/jpeg; swift_bytes=99'}):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['Content-Type'], 'image/jpeg')
|
|
|
|
def test_GET_removes_swift_bytes(self):
|
|
self._test_removes_swift_bytes('GET')
|
|
|
|
def test_HEAD_removes_swift_bytes(self):
|
|
self._test_removes_swift_bytes('HEAD')
|
|
|
|
def test_GET_error(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
self.app.logger.txn_id = req.environ['swift.trans_id'] = 'my-txn-id'
|
|
stdout = StringIO()
|
|
with set_http_connect(503, 200), \
|
|
mock.patch('sys.stdout', stdout):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
for line in stdout.getvalue().splitlines():
|
|
self.assertIn('my-txn-id', line)
|
|
self.assertIn('From Object Server', stdout.getvalue())
|
|
|
|
def test_GET_handoff(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
codes = [503] * self.obj_ring.replicas + [200]
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
def test_GET_not_found(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
codes = [404] * (self.obj_ring.replicas +
|
|
self.obj_ring.max_more_nodes)
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
def test_GET_primaries_explode(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
codes = [Exception('kaboom!')] * self.obj_ring.replicas + (
|
|
[404] * self.obj_ring.max_more_nodes)
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
def test_GET_primaries_timeout(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
codes = [Timeout()] * self.obj_ring.replicas + (
|
|
[404] * self.obj_ring.max_more_nodes)
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
self.assertEqual([1] * self.replicas(),
|
|
node_error_counts(self.app, self.obj_ring.devs))
|
|
|
|
def test_HEAD_error_limit_supression_count(self):
|
|
def do_test(primary_codes, expected, clear_stats=True):
|
|
if clear_stats:
|
|
self.app.error_limiter.stats.clear()
|
|
random.shuffle(primary_codes)
|
|
handoff_codes = [404] * self.obj_ring.max_more_nodes
|
|
with set_http_connect(*primary_codes + handoff_codes):
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='HEAD')
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, expected)
|
|
|
|
policy_opts = self.app.get_policy_options(None)
|
|
policy_opts.rebalance_missing_suppression_count = 1
|
|
# even with disks unmounted you can run with suppression_count = 1
|
|
do_test([507, 404, 404], 404)
|
|
# error limiting can make things wonky
|
|
do_test([404, 404], 404, clear_stats=False)
|
|
# and it gets a little dicy rebooting nodes
|
|
do_test([Timeout(), 404], 503, clear_stats=False)
|
|
do_test([507, Timeout(), 404], 503)
|
|
# unless you turn it off
|
|
policy_opts.rebalance_missing_suppression_count = 0
|
|
do_test([507, Timeout(), 404], 404)
|
|
|
|
def test_GET_primaries_error_during_rebalance(self):
|
|
def do_test(primary_codes, expected, include_timestamp=False):
|
|
random.shuffle(primary_codes)
|
|
handoff_codes = [404] * self.obj_ring.max_more_nodes
|
|
headers = None
|
|
if include_timestamp:
|
|
headers = [{'X-Backend-Timestamp': '123.456'}] * 3
|
|
headers.extend({} for _ in handoff_codes)
|
|
with set_http_connect(*primary_codes + handoff_codes,
|
|
headers=headers):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, expected)
|
|
|
|
# with two of out three backend errors a client should retry
|
|
do_test([Timeout(), Exception('kaboom!'), 404], 503)
|
|
# unless there's a timestamp associated
|
|
do_test([Timeout(), Exception('kaboom!'), 404], 404,
|
|
include_timestamp=True)
|
|
# when there's more 404s, we trust it more
|
|
do_test([Timeout(), 404, 404], 404)
|
|
# unless we explicitly *don't* want to trust it
|
|
policy_opts = self.app.get_policy_options(None)
|
|
policy_opts.rebalance_missing_suppression_count = 2
|
|
do_test([Timeout(), 404, 404], 503)
|
|
|
|
# overloaded primary after double rebalance
|
|
# ... opts should increase rebalance_missing_suppression_count
|
|
policy_opts.rebalance_missing_suppression_count = 2
|
|
do_test([Timeout(), 404, 404], 503)
|
|
|
|
# two primaries out, but no rebalance
|
|
# ... default is fine for tombstones
|
|
policy_opts.rebalance_missing_suppression_count = 1
|
|
do_test([Timeout(), Exception('kaboom!'), 404], 404,
|
|
include_timestamp=True)
|
|
# ... but maybe not ideal for missing names
|
|
# (N.B. 503 isn't really a BAD response here)
|
|
do_test([Timeout(), Exception('kaboom!'), 404], 503)
|
|
# still ... ops might think they should tune it down
|
|
policy_opts.rebalance_missing_suppression_count = 0
|
|
do_test([Timeout(), Exception('kaboom!'), 404], 404)
|
|
# and we could maybe leave it like this for the next rebalance
|
|
do_test([Timeout(), 404, 404], 404)
|
|
# ... but it gets bad when faced with timeouts, b/c we can't trust a
|
|
# single primary 404 response during rebalance
|
|
do_test([Timeout(), Timeout(), 404], 404)
|
|
# ops needs to fix configs to get the 503
|
|
policy_opts.rebalance_missing_suppression_count = 1
|
|
do_test([Timeout(), Timeout(), 404], 503)
|
|
|
|
def test_GET_primaries_mixed_explode_and_timeout(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
primaries = []
|
|
for i in range(self.obj_ring.replicas):
|
|
if i % 2:
|
|
primaries.append(Timeout())
|
|
else:
|
|
primaries.append(Exception('kaboom!'))
|
|
codes = primaries + [404] * self.obj_ring.max_more_nodes
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
def test_primary_returns_some_nonsense_timestamp(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
# an un-handled ValueError in _make_node_request should just continue
|
|
# to the next node rather than hang the request
|
|
headers = [{'X-Backend-Timestamp': 'not-a-timestamp'}, {}]
|
|
codes = [200, 200]
|
|
with quiet_eventlet_exceptions(), set_http_connect(
|
|
*codes, headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
def test_GET_not_found_when_404_newer(self):
|
|
# if proxy receives a 404, it keeps waiting for other connections until
|
|
# max number of nodes in hopes of finding an object, but if 404 is
|
|
# more recent than a 200, then it should ignore 200 and return 404
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
codes = [404] * self.obj_ring.replicas + \
|
|
[200] * self.obj_ring.max_more_nodes
|
|
ts_iter = iter([2] * self.obj_ring.replicas +
|
|
[1] * self.obj_ring.max_more_nodes)
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
def test_GET_x_newest_not_found_when_404_newer(self):
|
|
# if proxy receives a 404, it keeps waiting for other connections until
|
|
# max number of nodes in hopes of finding an object, but if 404 is
|
|
# more recent than a 200, then it should ignore 200 and return 404
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o',
|
|
headers={'X-Newest': 'true'})
|
|
codes = ([200] +
|
|
[404] * self.obj_ring.replicas +
|
|
[200] * (self.obj_ring.max_more_nodes - 1))
|
|
ts_iter = iter([1] +
|
|
[2] * self.obj_ring.replicas +
|
|
[1] * (self.obj_ring.max_more_nodes - 1))
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
def test_PUT_delete_at(self):
|
|
t = str(int(time.time() + 100))
|
|
req = swob.Request.blank('/v1/a/c/o', method='PUT', body=b'',
|
|
headers={'Content-Type': 'foo/bar',
|
|
'X-Delete-At': t})
|
|
put_headers = []
|
|
|
|
def capture_headers(ip, port, device, part, method, path, headers,
|
|
**kwargs):
|
|
if method == 'PUT':
|
|
put_headers.append(headers)
|
|
codes = [201] * self.obj_ring.replicas
|
|
with set_http_connect(*codes, give_connect=capture_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
for given_headers in put_headers:
|
|
self.assertEqual(given_headers.get('X-Delete-At'), t)
|
|
self.assertIn('X-Delete-At-Host', given_headers)
|
|
self.assertIn('X-Delete-At-Device', given_headers)
|
|
self.assertIn('X-Delete-At-Partition', given_headers)
|
|
self.assertIn('X-Delete-At-Container', given_headers)
|
|
|
|
def test_POST_delete_at_with_x_open_expired(self):
|
|
t_delete = str(int(time.time() + 30))
|
|
|
|
def capture_headers(ip, port, device, part, method, path, headers,
|
|
**kwargs):
|
|
if method == 'POST':
|
|
post_headers.append(headers)
|
|
|
|
def do_post(extra_headers):
|
|
headers = {'Content-Type': 'foo/bar',
|
|
'X-Delete-At': t_delete}
|
|
headers.update(extra_headers)
|
|
req_post = swob.Request.blank('/v1/a/c/o', method='POST', body=b'',
|
|
headers=headers)
|
|
|
|
post_codes = [202] * self.obj_ring.replicas
|
|
with set_http_connect(*post_codes, give_connect=capture_headers):
|
|
resp = req_post.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 202)
|
|
self.assertEqual(len(post_headers), self.obj_ring.replicas)
|
|
for given_headers in post_headers:
|
|
self.assertEqual(given_headers.get('X-Delete-At'), t_delete)
|
|
self.assertIn('X-Delete-At-Host', given_headers)
|
|
self.assertIn('X-Delete-At-Device', given_headers)
|
|
self.assertIn('X-Delete-At-Partition', given_headers)
|
|
self.assertIn('X-Delete-At-Container', given_headers)
|
|
|
|
# Check when allow_open_expired config is set to true
|
|
conf = {'allow_open_expired': 'true'}
|
|
self.app = PatchedObjControllerApp(
|
|
conf, account_ring=FakeRing(),
|
|
container_ring=FakeRing(), logger=None)
|
|
self.app.container_info = dict(self.container_info)
|
|
self.obj_ring = self.app.get_object_ring(int(self.policy))
|
|
|
|
post_headers = []
|
|
do_post({})
|
|
for given_headers in post_headers:
|
|
self.assertNotIn('X-Backend-Open-Expired', given_headers)
|
|
|
|
post_headers = []
|
|
do_post({'X-Open-Expired': 'false'})
|
|
for given_headers in post_headers:
|
|
self.assertNotIn('X-Backend-Open-Expired', given_headers)
|
|
|
|
post_headers = []
|
|
do_post({'X-Open-Expired': 'true'})
|
|
for given_headers in post_headers:
|
|
self.assertEqual(given_headers.get('X-Backend-Open-Expired'),
|
|
'true')
|
|
|
|
# Check when allow_open_expired config is set to false
|
|
conf = {'allow_open_expired': 'false'}
|
|
self.app = PatchedObjControllerApp(
|
|
conf, account_ring=FakeRing(),
|
|
container_ring=FakeRing(), logger=None)
|
|
self.app.container_info = dict(self.container_info)
|
|
self.obj_ring = self.app.get_object_ring(int(self.policy))
|
|
|
|
post_headers = []
|
|
do_post({})
|
|
for given_headers in post_headers:
|
|
self.assertNotIn('X-Backend-Open-Expired', given_headers)
|
|
|
|
post_headers = []
|
|
do_post({'X-Open-Expired': 'false'})
|
|
for given_headers in post_headers:
|
|
self.assertNotIn('X-Backend-Open-Expired', given_headers)
|
|
|
|
post_headers = []
|
|
do_post({'X-Open-Expired': 'true'})
|
|
for given_headers in post_headers:
|
|
self.assertNotIn('X-Backend-Open-Expired', given_headers)
|
|
|
|
def test_PUT_converts_delete_after_to_delete_at(self):
|
|
req = swob.Request.blank('/v1/a/c/o', method='PUT', body=b'',
|
|
headers={'Content-Type': 'foo/bar',
|
|
'X-Delete-After': '60'})
|
|
put_headers = []
|
|
|
|
def capture_headers(ip, port, device, part, method, path, headers,
|
|
**kwargs):
|
|
if method == 'PUT':
|
|
put_headers.append(headers)
|
|
codes = [201] * self.obj_ring.replicas
|
|
t = time.time()
|
|
with set_http_connect(*codes, give_connect=capture_headers):
|
|
with mock.patch('time.time', lambda: t):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
expected_delete_at = str(int(t) + 60)
|
|
for given_headers in put_headers:
|
|
self.assertEqual(given_headers.get('X-Delete-At'),
|
|
expected_delete_at)
|
|
self.assertIn('X-Delete-At-Host', given_headers)
|
|
self.assertIn('X-Delete-At-Device', given_headers)
|
|
self.assertIn('X-Delete-At-Partition', given_headers)
|
|
self.assertIn('X-Delete-At-Container', given_headers)
|
|
|
|
def test_container_sync_put_x_timestamp_not_found(self):
|
|
test_indexes = [None] + [int(p) for p in POLICIES]
|
|
for policy_index in test_indexes:
|
|
self.app.container_info['storage_policy'] = policy_index
|
|
put_timestamp = utils.Timestamp.now().normal
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': 0,
|
|
'X-Timestamp': put_timestamp})
|
|
codes = [201] * self.obj_ring.replicas
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_container_sync_put_x_timestamp_match(self):
|
|
test_indexes = [None] + [int(p) for p in POLICIES]
|
|
for policy_index in test_indexes:
|
|
self.app.container_info['storage_policy'] = policy_index
|
|
put_timestamp = utils.Timestamp.now().normal
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': 0,
|
|
'X-Timestamp': put_timestamp})
|
|
ts_iter = itertools.repeat(put_timestamp)
|
|
codes = [409] * self.obj_ring.replicas
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 202)
|
|
|
|
def test_container_sync_put_x_timestamp_older(self):
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
test_indexes = [None] + [int(p) for p in POLICIES]
|
|
for policy_index in test_indexes:
|
|
self.app.container_info['storage_policy'] = policy_index
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': 0,
|
|
'X-Timestamp': next(ts).internal})
|
|
ts_iter = itertools.repeat(next(ts).internal)
|
|
codes = [409] * self.obj_ring.replicas
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 202)
|
|
|
|
def test_container_sync_put_x_timestamp_newer(self):
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
test_indexes = [None] + [int(p) for p in POLICIES]
|
|
for policy_index in test_indexes:
|
|
orig_timestamp = next(ts).internal
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': 0,
|
|
'X-Timestamp': next(ts).internal})
|
|
ts_iter = itertools.repeat(orig_timestamp)
|
|
codes = [201] * self.obj_ring.replicas
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_put_x_timestamp_conflict(self):
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': 0,
|
|
'X-Timestamp': next(ts).internal})
|
|
ts_iter = iter([next(ts).internal, None, None])
|
|
codes = [409] + [201] * (self.obj_ring.replicas - 1)
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 202)
|
|
|
|
def test_put_x_timestamp_conflict_with_missing_backend_timestamp(self):
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': 0,
|
|
'X-Timestamp': next(ts).internal})
|
|
ts_iter = iter([None, None, None])
|
|
codes = [409] * self.obj_ring.replicas
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 202)
|
|
|
|
def test_put_x_timestamp_conflict_with_other_weird_success_response(self):
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': 0,
|
|
'X-Timestamp': next(ts).internal})
|
|
ts_iter = iter([next(ts).internal, None, None])
|
|
codes = [409] + [(201, 'notused')] * (self.obj_ring.replicas - 1)
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 202)
|
|
|
|
def test_put_x_timestamp_conflict_with_if_none_match(self):
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': 0,
|
|
'If-None-Match': '*',
|
|
'X-Timestamp': next(ts).internal})
|
|
ts_iter = iter([next(ts).internal, None, None])
|
|
codes = [409] + [(412, 'notused')] * (self.obj_ring.replicas - 1)
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 412)
|
|
|
|
def test_container_sync_put_x_timestamp_race(self):
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
test_indexes = [None] + [int(p) for p in POLICIES]
|
|
for policy_index in test_indexes:
|
|
put_timestamp = next(ts).internal
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': 0,
|
|
'X-Timestamp': put_timestamp})
|
|
|
|
# object nodes they respond 409 because another in-flight request
|
|
# finished and now the on disk timestamp is equal to the request.
|
|
put_ts = [put_timestamp] * self.obj_ring.replicas
|
|
codes = [409] * self.obj_ring.replicas
|
|
|
|
ts_iter = iter(put_ts)
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 202)
|
|
|
|
def test_container_sync_put_x_timestamp_unsynced_race(self):
|
|
ts = (utils.Timestamp(t) for t in itertools.count(int(time.time())))
|
|
test_indexes = [None] + [int(p) for p in POLICIES]
|
|
for policy_index in test_indexes:
|
|
put_timestamp = next(ts).internal
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers={
|
|
'Content-Length': 0,
|
|
'X-Timestamp': put_timestamp})
|
|
|
|
# only one in-flight request finished
|
|
put_ts = [None] * (self.obj_ring.replicas - 1)
|
|
put_resp = [201] * (self.obj_ring.replicas - 1)
|
|
put_ts += [put_timestamp]
|
|
put_resp += [409]
|
|
|
|
ts_iter = iter(put_ts)
|
|
codes = put_resp
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 202)
|
|
|
|
def test_x_timestamp_not_overridden(self):
|
|
def do_test(method, base_headers, resp_code):
|
|
# no given x-timestamp
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method=method, headers=base_headers)
|
|
codes = [resp_code] * self.replicas()
|
|
with mocked_http_conn(*codes) as fake_conn:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, resp_code)
|
|
self.assertEqual(self.replicas(), len(fake_conn.requests))
|
|
for req in fake_conn.requests:
|
|
self.assertIn('X-Timestamp', req['headers'])
|
|
# check value can be parsed as valid timestamp
|
|
Timestamp(req['headers']['X-Timestamp'])
|
|
|
|
# given x-timestamp is retained
|
|
def do_check(ts):
|
|
headers = dict(base_headers)
|
|
headers['X-Timestamp'] = ts.internal
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method=method, headers=headers)
|
|
codes = [resp_code] * self.replicas()
|
|
with mocked_http_conn(*codes) as fake_conn:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, resp_code)
|
|
self.assertEqual(self.replicas(), len(fake_conn.requests))
|
|
for req in fake_conn.requests:
|
|
self.assertEqual(ts.internal,
|
|
req['headers']['X-Timestamp'])
|
|
|
|
do_check(Timestamp.now())
|
|
do_check(Timestamp.now(offset=123))
|
|
|
|
# given x-timestamp gets sanity checked
|
|
headers = dict(base_headers)
|
|
headers['X-Timestamp'] = 'bad timestamp'
|
|
req = swob.Request.blank(
|
|
'/v1/a/c/o', method=method, headers=headers)
|
|
with mocked_http_conn() as fake_conn:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 400)
|
|
self.assertIn(b'X-Timestamp should be a UNIX timestamp ',
|
|
resp.body)
|
|
|
|
do_test('PUT', {'Content-Length': 0}, 200)
|
|
do_test('DELETE', {}, 204)
|
|
|
|
|
|
@patch_policies(
|
|
[StoragePolicy(0, '1-replica', True),
|
|
StoragePolicy(1, '4-replica', False),
|
|
StoragePolicy(2, '8-replica', False),
|
|
StoragePolicy(3, '15-replica', False)],
|
|
fake_ring_args=[
|
|
{'replicas': 1}, {'replicas': 4}, {'replicas': 8}, {'replicas': 15}])
|
|
class TestReplicatedObjControllerVariousReplicas(CommonObjectControllerMixin,
|
|
unittest.TestCase):
|
|
controller_cls = obj.ReplicatedObjectController
|
|
|
|
def test_DELETE_with_write_affinity(self):
|
|
policy_index = 1
|
|
self.policy = POLICIES[policy_index]
|
|
policy_conf = self.app.get_policy_options(self.policy)
|
|
self.app.container_info['storage_policy'] = policy_index
|
|
policy_conf.write_affinity_handoff_delete_count = \
|
|
self.replicas(self.policy) // 2
|
|
policy_conf.write_affinity_is_local_fn = (
|
|
lambda node: node['region'] == 1)
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='DELETE')
|
|
|
|
codes = [204, 204, 404, 404, 204, 204]
|
|
with mocked_http_conn(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
policy_conf.write_affinity_handoff_delete_count = 1
|
|
|
|
codes = [204, 204, 404, 404, 204]
|
|
with mocked_http_conn(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
|
|
@patch_policies()
|
|
class TestReplicatedObjControllerMimePutter(BaseObjectControllerMixin,
|
|
unittest.TestCase):
|
|
# tests specific to PUTs using a MimePutter
|
|
expect_headers = {
|
|
'X-Obj-Metadata-Footer': 'yes'
|
|
}
|
|
|
|
def setUp(self):
|
|
super(TestReplicatedObjControllerMimePutter, self).setUp()
|
|
# force use of a MimePutter
|
|
self.app.use_put_v1 = False
|
|
|
|
def test_PUT_error(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
codes = [503] * self.replicas()
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
def _test_PUT_with_footers(self, test_body=b''):
|
|
# verify that when footers are required the PUT body is multipart
|
|
# and the footers are appended
|
|
footers_callback = make_footers_callback(test_body)
|
|
env = {'swift.callback.update_footers': footers_callback}
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
environ=env)
|
|
req.body = test_body
|
|
# send bogus Etag header to differentiate from footer value
|
|
req.headers['Etag'] = 'header_etag'
|
|
codes = [201] * self.replicas()
|
|
|
|
put_requests = defaultdict(
|
|
lambda: {'headers': None, 'chunks': [], 'connection': None})
|
|
|
|
def capture_body(conn, chunk):
|
|
put_requests[conn.connection_id]['chunks'].append(chunk)
|
|
put_requests[conn.connection_id]['connection'] = conn
|
|
|
|
def capture_headers(ip, port, device, part, method, path, headers,
|
|
**kwargs):
|
|
conn_id = kwargs['connection_id']
|
|
put_requests[conn_id]['headers'] = headers
|
|
|
|
resp_headers = {
|
|
'Etag': '"resp_etag"',
|
|
# NB: ignored!
|
|
'Some-Header': 'Four',
|
|
}
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers,
|
|
give_send=capture_body,
|
|
give_connect=capture_headers,
|
|
headers=resp_headers):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
timestamps = {captured_req['headers']['x-timestamp']
|
|
for captured_req in put_requests.values()}
|
|
self.assertEqual(1, len(timestamps), timestamps)
|
|
self.assertEqual(dict(resp.headers), {
|
|
'Content-Type': 'text/html; charset=UTF-8',
|
|
'Content-Length': '0',
|
|
'Etag': 'resp_etag',
|
|
'Last-Modified': time.strftime(
|
|
"%a, %d %b %Y %H:%M:%S GMT",
|
|
time.gmtime(math.ceil(float(timestamps.pop())))),
|
|
})
|
|
for connection_id, info in put_requests.items():
|
|
body = unchunk_body(b''.join(info['chunks']))
|
|
headers = info['headers']
|
|
boundary = headers['X-Backend-Obj-Multipart-Mime-Boundary']
|
|
self.assertTrue(boundary is not None,
|
|
"didn't get boundary for conn %r" % (
|
|
connection_id,))
|
|
self.assertEqual('chunked', headers['Transfer-Encoding'])
|
|
self.assertEqual('100-continue', headers['Expect'])
|
|
self.assertEqual('yes', headers['X-Backend-Obj-Metadata-Footer'])
|
|
self.assertNotIn('X-Backend-Obj-Multiphase-Commit', headers)
|
|
self.assertEqual('header_etag', headers['Etag'])
|
|
|
|
# email.parser.FeedParser doesn't know how to take a multipart
|
|
# message and boundary together and parse it; it only knows how
|
|
# to take a string, parse the headers, and figure out the
|
|
# boundary on its own.
|
|
parser = EmailFeedParser()
|
|
parser.feed(
|
|
("Content-Type: multipart/nobodycares; boundary=%s\r\n\r\n" %
|
|
boundary).encode('ascii'))
|
|
parser.feed(body)
|
|
message = parser.close()
|
|
|
|
self.assertTrue(message.is_multipart()) # sanity check
|
|
mime_parts = message.get_payload()
|
|
# notice, no commit confirmation
|
|
self.assertEqual(len(mime_parts), 2)
|
|
obj_part, footer_part = mime_parts
|
|
|
|
self.assertEqual(obj_part['X-Document'], 'object body')
|
|
self.assertEqual(test_body, obj_part.get_payload(decode=True))
|
|
|
|
# validate footer metadata
|
|
self.assertEqual(footer_part['X-Document'], 'object metadata')
|
|
footer_metadata = json.loads(footer_part.get_payload())
|
|
self.assertTrue(footer_metadata)
|
|
expected = {}
|
|
footers_callback(expected)
|
|
self.assertDictEqual(expected, footer_metadata)
|
|
|
|
self.assertTrue(info['connection'].closed)
|
|
|
|
def test_PUT_with_body_and_footers(self):
|
|
self._test_PUT_with_footers(test_body=b'asdf')
|
|
|
|
def test_PUT_with_no_body_and_footers(self):
|
|
self._test_PUT_with_footers()
|
|
|
|
|
|
@contextmanager
|
|
def capture_http_requests(get_response):
|
|
|
|
class FakeConn(object):
|
|
|
|
def __init__(self, req):
|
|
self.req = req
|
|
self.resp = None
|
|
self.path = "/"
|
|
self.closed = False
|
|
|
|
def getresponse(self):
|
|
self.resp = get_response(self.req)
|
|
return self.resp
|
|
|
|
def putrequest(self, method, path, **kwargs):
|
|
pass
|
|
|
|
def putheader(self, k, v):
|
|
pass
|
|
|
|
def endheaders(self):
|
|
pass
|
|
|
|
def close(self):
|
|
self.closed = True
|
|
|
|
class ConnectionLog(object):
|
|
|
|
def __init__(self):
|
|
self.connections = []
|
|
|
|
def __len__(self):
|
|
return len(self.connections)
|
|
|
|
def __getitem__(self, i):
|
|
return self.connections[i]
|
|
|
|
def __iter__(self):
|
|
return iter(self.connections)
|
|
|
|
def __call__(self, ip, port, method, path, headers, qs, ssl):
|
|
req = {
|
|
'ip': ip,
|
|
'port': port,
|
|
'method': method,
|
|
'path': path,
|
|
'headers': headers,
|
|
'qs': qs,
|
|
'ssl': ssl,
|
|
}
|
|
conn = FakeConn(req)
|
|
self.connections.append(conn)
|
|
return conn
|
|
|
|
fake_conn = ConnectionLog()
|
|
|
|
with mock.patch('swift.common.bufferedhttp.http_connect_raw',
|
|
new=fake_conn):
|
|
yield fake_conn
|
|
|
|
|
|
class ECObjectControllerMixin(CommonObjectControllerMixin):
|
|
# Add a few helper methods for EC tests.
|
|
def _make_ec_archive_bodies(self, test_body, policy=None):
|
|
policy = policy or self.policy
|
|
return encode_frag_archive_bodies(policy, test_body)
|
|
|
|
def _make_ec_object_stub(self, pattern='test', policy=None,
|
|
timestamp=None):
|
|
policy = policy or self.policy
|
|
if isinstance(pattern, six.text_type):
|
|
pattern = pattern.encode('utf-8')
|
|
test_body = pattern * policy.ec_segment_size
|
|
test_body = test_body[:-random.randint(1, 1000)]
|
|
return make_ec_object_stub(test_body, policy, timestamp)
|
|
|
|
def _fake_ec_node_response(self, node_frags):
|
|
return fake_ec_node_response(node_frags, self.policy)
|
|
|
|
def test_GET_with_duplicate_but_sufficient_frag_indexes(self):
|
|
obj1 = self._make_ec_object_stub()
|
|
# proxy should ignore duplicated frag indexes and continue search for
|
|
# a set of unique indexes, finding last one on a handoff
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0},
|
|
{'obj': obj1, 'frag': 0}, # duplicate frag
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj1, 'frag': 1}, # duplicate frag
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj1, 'frag': 2}, # duplicate frag
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj1, 'frag': 3}, # duplicate frag
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj1, 'frag': 4}, # duplicate frag
|
|
{'obj': obj1, 'frag': 10},
|
|
{'obj': obj1, 'frag': 11},
|
|
{'obj': obj1, 'frag': 12},
|
|
{'obj': obj1, 'frag': 13},
|
|
] * self.policy.ec_duplication_factor
|
|
node_frags.append({'obj': obj1, 'frag': 5}) # first handoff
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj1['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj1['etag'])
|
|
|
|
# expect a request to all primaries plus one handoff
|
|
self.assertEqual(self.replicas() + 1, len(log))
|
|
collected_indexes = defaultdict(list)
|
|
for conn in log:
|
|
fi = conn.resp.headers.get('X-Object-Sysmeta-Ec-Frag-Index')
|
|
if fi is not None:
|
|
collected_indexes[fi].append(conn)
|
|
self.assertEqual(len(collected_indexes), self.policy.ec_ndata)
|
|
|
|
def test_GET_with_duplicate_but_insufficient_frag_indexes(self):
|
|
obj1 = self._make_ec_object_stub()
|
|
# proxy should ignore duplicated frag indexes and continue search for
|
|
# a set of unique indexes, but fails to find one
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0},
|
|
{'obj': obj1, 'frag': 0}, # duplicate frag
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj1, 'frag': 1}, # duplicate frag
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj1, 'frag': 2}, # duplicate frag
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj1, 'frag': 3}, # duplicate frag
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj1, 'frag': 4}, # duplicate frag
|
|
{'obj': obj1, 'frag': 10},
|
|
{'obj': obj1, 'frag': 11},
|
|
{'obj': obj1, 'frag': 12},
|
|
{'obj': obj1, 'frag': 13},
|
|
]
|
|
|
|
# ... and the rest are 404s which is limited by request_count
|
|
# (2 * replicas in default) rather than max_extra_requests limitation
|
|
# because the retries will be in ResumingGetter if the responses
|
|
# are 404s
|
|
node_frags += [[]] * (self.replicas() * 2 - len(node_frags))
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
# expect a request to all nodes
|
|
self.assertEqual(2 * self.replicas(), len(log))
|
|
collected_indexes = defaultdict(list)
|
|
for conn in log:
|
|
fi = conn.resp.headers.get('X-Object-Sysmeta-Ec-Frag-Index')
|
|
if fi is not None:
|
|
collected_indexes[fi].append(conn)
|
|
self.assertEqual(len(collected_indexes), self.policy.ec_ndata - 1)
|
|
|
|
|
|
@patch_policies(with_ec_default=True)
|
|
class TestECObjController(ECObjectControllerMixin, unittest.TestCase):
|
|
container_info = {
|
|
'status': 200,
|
|
'read_acl': None,
|
|
'write_acl': None,
|
|
'sync_key': None,
|
|
'versions': None,
|
|
'storage_policy': '0',
|
|
}
|
|
|
|
controller_cls = obj.ECObjectController
|
|
|
|
def _add_frag_index(self, index, headers):
|
|
# helper method to add a frag index header to an existing header dict
|
|
hdr_name = 'X-Object-Sysmeta-Ec-Frag-Index'
|
|
return dict(list(headers.items()) + [(hdr_name, index)])
|
|
|
|
def test_determine_chunk_destinations(self):
|
|
class FakePutter(object):
|
|
def __init__(self, index):
|
|
self.node_index = index
|
|
|
|
controller = self.controller_cls(
|
|
self.app, 'a', 'c', 'o')
|
|
|
|
# create a dummy list of putters, check no handoffs
|
|
putters = []
|
|
expected = {}
|
|
for index in range(self.policy.object_ring.replica_count):
|
|
p = FakePutter(index)
|
|
putters.append(p)
|
|
expected[p] = self.policy.get_backend_index(index)
|
|
got = controller._determine_chunk_destinations(putters, self.policy)
|
|
self.assertEqual(got, expected)
|
|
|
|
def _test_one_handoff(index):
|
|
with mock.patch.object(putters[index], 'node_index', None):
|
|
got = controller._determine_chunk_destinations(
|
|
putters, self.policy)
|
|
self.assertEqual(got, expected)
|
|
# Check that we don't mutate the putter
|
|
self.assertEqual([p.node_index for p in putters],
|
|
[None if i == index else i
|
|
for i, _ in enumerate(putters)])
|
|
|
|
# now lets make a handoff at the end
|
|
_test_one_handoff(self.policy.object_ring.replica_count - 1)
|
|
# now lets make a handoff at the start
|
|
_test_one_handoff(0)
|
|
# now lets make a handoff in the middle
|
|
_test_one_handoff(2)
|
|
|
|
# now lets make all of them handoffs
|
|
for index in range(self.policy.object_ring.replica_count):
|
|
putters[index].node_index = None
|
|
got = controller._determine_chunk_destinations(putters, self.policy)
|
|
self.assertEqual(sorted(got, key=lambda p: id(p)),
|
|
sorted(expected, key=lambda p: id(p)))
|
|
|
|
def test_GET_simple(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
get_statuses = [200] * self.policy.ec_ndata
|
|
get_hdrs = [{
|
|
'Connection': 'close',
|
|
'X-Object-Sysmeta-Ec-Scheme': self.policy.ec_scheme_description,
|
|
}] * self.policy.ec_ndata
|
|
with set_http_connect(*get_statuses, headers=get_hdrs):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertIn('Accept-Ranges', resp.headers)
|
|
self.assertNotIn('Connection', resp.headers)
|
|
self.assertFalse([h for h in resp.headers
|
|
if h.lower().startswith('x-object-sysmeta-ec-')])
|
|
|
|
def test_GET_disconnect(self):
|
|
self.app.recoverable_node_timeout = 0.01
|
|
self.app.client_timeout = 0.1
|
|
# Before, we used the default 64k chunk size, so the entire ~16k test
|
|
# data would come in the first chunk, and the generator should
|
|
# cleanly exit by the time we reiterate() the response.
|
|
self.app.object_chunk_size = 10
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-743]
|
|
etag = md5(test_data, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
headers = {
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(test_data),
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
num_slow = 4
|
|
responses = [
|
|
(200, SlowBody(body, 0.1 if i < num_slow else 0.0),
|
|
self._add_frag_index(i, headers))
|
|
for i, body in enumerate(ec_archive_bodies)
|
|
] * self.policy.ec_duplication_factor
|
|
|
|
status_codes, body_iter, headers = zip(*responses)
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
with mocked_http_conn(*status_codes, body_iter=body_iter,
|
|
headers=headers) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
resp.app_iter.close()
|
|
self.assertEqual(len(log.requests),
|
|
self.policy.ec_ndata + num_slow)
|
|
|
|
def make_key(r):
|
|
r['device'] = r['path'].split('/')[1]
|
|
return '%(ip)s:%(port)s/%(device)s' % r
|
|
# the first four slow nodes get errors incr'd
|
|
expected_error_limiting = {
|
|
make_key(r): {
|
|
'errors': 1,
|
|
'last_error': mock.ANY,
|
|
}
|
|
for r in log.requests[:4]
|
|
}
|
|
actual = {}
|
|
for n in self.app.get_object_ring(int(self.policy)).devs:
|
|
node_key = self.app.error_limiter.node_key(n)
|
|
stats = self.app.error_limiter.stats.get(node_key) or {}
|
|
if stats:
|
|
actual[self.app.error_limiter.node_key(n)] = stats
|
|
self.assertEqual(actual, expected_error_limiting)
|
|
expected = ["Client disconnected on read of EC frag '/a/c/o'"] * 10
|
|
self.assertEqual(
|
|
self.app.logger.get_lines_for_level('warning'),
|
|
expected)
|
|
for read_line in self.app.logger.get_lines_for_level('error'):
|
|
self.assertIn("Trying to read EC fragment during GET (retrying)",
|
|
read_line)
|
|
self.assertEqual(
|
|
len(self.logger.logger.records['ERROR']), 4,
|
|
'Expected 4 ERROR lines, got %r' % (
|
|
self.logger.logger.records['ERROR'], ))
|
|
|
|
def test_GET_not_found_when_404_newer(self):
|
|
# if proxy receives a 404, it keeps waiting for other connections until
|
|
# max number of nodes in hopes of finding an object, but if 404 is
|
|
# more recent than a 200, then it should ignore 200 and return 404
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
rest = 2 * self.policy.object_ring.replica_count - 2
|
|
codes = [200, 404] + [200] * rest
|
|
ts_iter = iter([1, 2] + [1] * rest)
|
|
with set_http_connect(*codes, timestamps=ts_iter):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
def test_GET_primaries_error_during_rebalance(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
codes = [404] * (2 * self.policy.object_ring.replica_count)
|
|
with mocked_http_conn(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
for i in range(self.policy.object_ring.replica_count - 2):
|
|
codes[i] = Timeout()
|
|
with mocked_http_conn(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
self.app.error_limiter.stats.clear() # Reset error limiting
|
|
|
|
# one more timeout is past the tipping point
|
|
codes[self.policy.object_ring.replica_count - 2] = Timeout()
|
|
with mocked_http_conn(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
self.app.error_limiter.stats.clear() # Reset error limiting
|
|
|
|
# unless we have tombstones
|
|
with mocked_http_conn(*codes, headers={'X-Backend-Timestamp': '1'}):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
def _test_if_match(self, method):
|
|
num_responses = self.policy.ec_ndata if method == 'GET' else 1
|
|
|
|
def _do_test(match_value, backend_status,
|
|
etag_is_at='X-Object-Sysmeta-Does-Not-Exist'):
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method=method,
|
|
headers={'If-Match': match_value,
|
|
'X-Backend-Etag-Is-At': etag_is_at})
|
|
get_resp = [backend_status] * num_responses
|
|
resp_headers = {'Etag': 'frag_etag',
|
|
'X-Object-Sysmeta-Ec-Etag': 'data_etag',
|
|
'X-Object-Sysmeta-Alternate-Etag': 'alt_etag'}
|
|
with set_http_connect(*get_resp, headers=resp_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual('data_etag', resp.headers['Etag'])
|
|
return resp
|
|
|
|
# wildcard
|
|
resp = _do_test('*', 200)
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
# match
|
|
resp = _do_test('"data_etag"', 200)
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
# no match
|
|
resp = _do_test('"frag_etag"', 412)
|
|
self.assertEqual(resp.status_int, 412)
|
|
|
|
# match wildcard against an alternate etag
|
|
resp = _do_test('*', 200,
|
|
etag_is_at='X-Object-Sysmeta-Alternate-Etag')
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
# match against an alternate etag
|
|
resp = _do_test('"alt_etag"', 200,
|
|
etag_is_at='X-Object-Sysmeta-Alternate-Etag')
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
# no match against an alternate etag
|
|
resp = _do_test('"data_etag"', 412,
|
|
etag_is_at='X-Object-Sysmeta-Alternate-Etag')
|
|
self.assertEqual(resp.status_int, 412)
|
|
|
|
def test_GET_if_match(self):
|
|
self._test_if_match('GET')
|
|
|
|
def test_HEAD_if_match(self):
|
|
self._test_if_match('HEAD')
|
|
|
|
def _test_if_none_match(self, method):
|
|
num_responses = self.policy.ec_ndata if method == 'GET' else 1
|
|
|
|
def _do_test(match_value, backend_status,
|
|
etag_is_at='X-Object-Sysmeta-Does-Not-Exist'):
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method=method,
|
|
headers={'If-None-Match': match_value,
|
|
'X-Backend-Etag-Is-At': etag_is_at})
|
|
get_resp = [backend_status] * num_responses
|
|
resp_headers = {'Etag': 'frag_etag',
|
|
'X-Object-Sysmeta-Ec-Etag': 'data_etag',
|
|
'X-Object-Sysmeta-Alternate-Etag': 'alt_etag'}
|
|
with set_http_connect(*get_resp, headers=resp_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual('data_etag', resp.headers['Etag'])
|
|
return resp
|
|
|
|
# wildcard
|
|
resp = _do_test('*', 304)
|
|
self.assertEqual(resp.status_int, 304)
|
|
|
|
# match
|
|
resp = _do_test('"data_etag"', 304)
|
|
self.assertEqual(resp.status_int, 304)
|
|
|
|
# no match
|
|
resp = _do_test('"frag_etag"', 200)
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
# match wildcard against an alternate etag
|
|
resp = _do_test('*', 304,
|
|
etag_is_at='X-Object-Sysmeta-Alternate-Etag')
|
|
self.assertEqual(resp.status_int, 304)
|
|
|
|
# match against an alternate etag
|
|
resp = _do_test('"alt_etag"', 304,
|
|
etag_is_at='X-Object-Sysmeta-Alternate-Etag')
|
|
self.assertEqual(resp.status_int, 304)
|
|
|
|
# no match against an alternate etag
|
|
resp = _do_test('"data_etag"', 200,
|
|
etag_is_at='X-Object-Sysmeta-Alternate-Etag')
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
def test_GET_if_none_match(self):
|
|
self._test_if_none_match('GET')
|
|
|
|
def test_HEAD_if_none_match(self):
|
|
self._test_if_none_match('HEAD')
|
|
|
|
def test_GET_simple_x_newest(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o',
|
|
headers={'X-Newest': 'true'})
|
|
codes = [200] * self.policy.ec_ndata
|
|
with set_http_connect(*codes):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
def test_GET_error(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
get_resp = [503] + [200] * self.policy.ec_ndata
|
|
with set_http_connect(*get_resp):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
def test_GET_no_response_error(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
with set_http_connect():
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
def test_feed_remaining_primaries(self):
|
|
controller = self.controller_cls(
|
|
self.app, 'a', 'c', 'o')
|
|
safe_iter = utils.GreenthreadSafeIterator(NodeIter(
|
|
'object', self.app, self.policy.object_ring, 0, self.logger,
|
|
policy=self.policy, request=Request.blank('')))
|
|
controller._fragment_GET_request = lambda *a, **k: next(safe_iter)
|
|
pile = utils.GreenAsyncPile(self.policy.ec_ndata)
|
|
for i in range(self.policy.ec_ndata):
|
|
pile.spawn(controller._fragment_GET_request)
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
|
|
feeder_q = mock.MagicMock()
|
|
|
|
def feeder_timeout(*a, **kw):
|
|
# simulate trampoline
|
|
sleep()
|
|
# timeout immediately
|
|
raise Empty
|
|
feeder_q.get.side_effect = feeder_timeout
|
|
controller.feed_remaining_primaries(
|
|
safe_iter, pile, req, 0, self.policy,
|
|
mock.MagicMock(), feeder_q, mock.MagicMock())
|
|
expected_timeout = self.app.get_policy_options(
|
|
self.policy).concurrency_timeout
|
|
expected_call = mock.call(timeout=expected_timeout)
|
|
expected_num_calls = self.policy.ec_nparity + 1
|
|
self.assertEqual(feeder_q.get.call_args_list,
|
|
[expected_call] * expected_num_calls)
|
|
|
|
def test_GET_timeout(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
self.app.recoverable_node_timeout = 0.01
|
|
codes = [FakeStatus(404, response_sleep=1.0)] * 2 + \
|
|
[200] * (self.policy.ec_ndata)
|
|
with mocked_http_conn(*codes) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(self.policy.ec_ndata + 2, len(log.requests))
|
|
self.assertEqual(
|
|
len(self.logger.logger.records['ERROR']), 2,
|
|
'Expected 2 ERROR lines, got %r' % (
|
|
self.logger.logger.records['ERROR'], ))
|
|
for retry_line in self.logger.logger.records['ERROR']:
|
|
self.assertIn('ERROR with Object server', retry_line)
|
|
self.assertIn('Trying to GET', retry_line)
|
|
self.assertIn('Timeout (0.01s)', retry_line)
|
|
self.assertIn(req.headers['x-trans-id'], retry_line)
|
|
|
|
def test_GET_with_slow_primaries(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-743]
|
|
etag = md5(test_data, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
ts = self.ts()
|
|
headers = []
|
|
for i, body in enumerate(ec_archive_bodies):
|
|
headers.append({
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(body),
|
|
'X-Object-Sysmeta-Ec-Frag-Index':
|
|
self.policy.get_backend_index(i),
|
|
'X-Backend-Timestamp': ts.internal,
|
|
'X-Timestamp': ts.normal,
|
|
'X-Backend-Durable-Timestamp': ts.internal,
|
|
'X-Backend-Data-Timestamp': ts.internal,
|
|
})
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
|
|
policy_opts = self.app.get_policy_options(self.policy)
|
|
policy_opts.concurrent_gets = True
|
|
policy_opts.concurrency_timeout = 0.1
|
|
|
|
status_codes = ([
|
|
FakeStatus(200, response_sleep=2.0),
|
|
] * self.policy.ec_nparity) + ([
|
|
FakeStatus(200),
|
|
] * self.policy.ec_ndata)
|
|
self.assertEqual(len(status_codes), len(ec_archive_bodies))
|
|
with mocked_http_conn(*status_codes, body_iter=ec_archive_bodies,
|
|
headers=headers) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(len(log.requests),
|
|
self.policy.ec_n_unique_fragments)
|
|
|
|
def test_GET_with_some_slow_primaries(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-289]
|
|
etag = md5(test_data, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
ts = self.ts()
|
|
headers = []
|
|
for i, body in enumerate(ec_archive_bodies):
|
|
headers.append({
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(body),
|
|
'X-Object-Sysmeta-Ec-Frag-Index':
|
|
self.policy.get_backend_index(i),
|
|
'X-Backend-Timestamp': ts.internal,
|
|
'X-Timestamp': ts.normal,
|
|
'X-Backend-Durable-Timestamp': ts.internal,
|
|
'X-Backend-Data-Timestamp': ts.internal,
|
|
})
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
|
|
policy_opts = self.app.get_policy_options(self.policy)
|
|
policy_opts.concurrent_gets = True
|
|
policy_opts.concurrency_timeout = 0.1
|
|
|
|
slow_count = self.policy.ec_nparity
|
|
status_codes = ([
|
|
FakeStatus(200, response_sleep=2.0),
|
|
] * slow_count) + ([
|
|
FakeStatus(200),
|
|
] * (self.policy.ec_ndata - slow_count))
|
|
random.shuffle(status_codes)
|
|
status_codes.extend([
|
|
FakeStatus(200),
|
|
] * slow_count)
|
|
self.assertEqual(len(status_codes), len(ec_archive_bodies))
|
|
with mocked_http_conn(*status_codes, body_iter=ec_archive_bodies,
|
|
headers=headers) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(len(log.requests),
|
|
self.policy.ec_n_unique_fragments)
|
|
|
|
def test_ec_concurrent_GET_with_slow_leaders(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-289]
|
|
etag = md5(test_data).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
ts = self.ts()
|
|
headers = []
|
|
for i, body in enumerate(ec_archive_bodies):
|
|
headers.append({
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(body),
|
|
'X-Object-Sysmeta-Ec-Frag-Index':
|
|
self.policy.get_backend_index(i),
|
|
'X-Backend-Timestamp': ts.internal,
|
|
'X-Timestamp': ts.normal,
|
|
'X-Backend-Durable-Timestamp': ts.internal,
|
|
'X-Backend-Data-Timestamp': ts.internal,
|
|
})
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
|
|
policy_opts = self.app.get_policy_options(self.policy)
|
|
policy_opts.concurrent_gets = True
|
|
policy_opts.concurrency_timeout = 0.0
|
|
|
|
slow_count = 4
|
|
status_codes = ([
|
|
FakeStatus(200, response_sleep=0.2),
|
|
] * slow_count) + ([
|
|
FakeStatus(200, response_sleep=0.1),
|
|
] * (self.policy.ec_n_unique_fragments - slow_count))
|
|
for i in range(slow_count):
|
|
# poison the super slow requests
|
|
ec_archive_bodies[i] = ''
|
|
with mocked_http_conn(*status_codes, body_iter=ec_archive_bodies,
|
|
headers=headers) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.body, test_data, '%r != %r' % (
|
|
resp.body if len(resp.body) < 60 else '%s...' % resp.body[:60],
|
|
test_data if len(test_data) < 60 else '%s...' % test_data[:60],
|
|
))
|
|
self.assertEqual(len(log.requests), self.policy.ec_n_unique_fragments)
|
|
|
|
def test_GET_with_slow_nodes_and_failures(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-289]
|
|
etag = md5(test_data, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
ts = self.ts()
|
|
headers = []
|
|
for i, body in enumerate(ec_archive_bodies):
|
|
headers.append({
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(body),
|
|
'X-Object-Sysmeta-Ec-Frag-Index':
|
|
self.policy.get_backend_index(i),
|
|
'X-Backend-Timestamp': ts.internal,
|
|
'X-Timestamp': ts.normal,
|
|
'X-Backend-Durable-Timestamp': ts.internal,
|
|
'X-Backend-Data-Timestamp': ts.internal,
|
|
})
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
|
|
policy_opts = self.app.get_policy_options(self.policy)
|
|
policy_opts.concurrent_gets = True
|
|
policy_opts.concurrency_timeout = 0.1
|
|
|
|
unused_resp = [
|
|
FakeStatus(200, response_sleep=2.0),
|
|
FakeStatus(200, response_sleep=2.0),
|
|
500,
|
|
416,
|
|
]
|
|
self.assertEqual(len(unused_resp), self.policy.ec_nparity)
|
|
status_codes = (
|
|
[200] * (self.policy.ec_ndata - 4)) + unused_resp
|
|
self.assertEqual(len(status_codes), self.policy.ec_ndata)
|
|
# random.shuffle(status_codes)
|
|
# make up for the failures
|
|
status_codes.extend([200] * self.policy.ec_nparity)
|
|
self.assertEqual(len(status_codes), len(ec_archive_bodies))
|
|
bodies_with_errors = []
|
|
for code, body in zip(status_codes, ec_archive_bodies):
|
|
if code == 500:
|
|
bodies_with_errors.append('Kaboom')
|
|
elif code == 416:
|
|
bodies_with_errors.append('That Range is no.')
|
|
else:
|
|
bodies_with_errors.append(body)
|
|
with mocked_http_conn(*status_codes, body_iter=bodies_with_errors,
|
|
headers=headers) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(len(log.requests),
|
|
self.policy.ec_n_unique_fragments)
|
|
|
|
def test_GET_with_one_slow_frag_lane(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-454]
|
|
etag = md5(test_data, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
ts = self.ts()
|
|
headers = []
|
|
for i, body in enumerate(ec_archive_bodies):
|
|
headers.append({
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(body),
|
|
'X-Object-Sysmeta-Ec-Frag-Index':
|
|
self.policy.get_backend_index(i),
|
|
'X-Backend-Timestamp': ts.internal,
|
|
'X-Timestamp': ts.normal,
|
|
'X-Backend-Durable-Timestamp': ts.internal,
|
|
'X-Backend-Data-Timestamp': ts.internal,
|
|
})
|
|
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
|
|
policy_opts = self.app.get_policy_options(self.policy)
|
|
policy_opts.concurrent_gets = True
|
|
policy_opts.concurrency_timeout = 0.1
|
|
|
|
status_codes = [
|
|
FakeStatus(200, response_sleep=2.0),
|
|
] + ([
|
|
FakeStatus(200),
|
|
] * (self.policy.ec_ndata - 1))
|
|
random.shuffle(status_codes)
|
|
status_codes.extend([
|
|
FakeStatus(200, response_sleep=2.0),
|
|
FakeStatus(200, response_sleep=2.0),
|
|
FakeStatus(200, response_sleep=2.0),
|
|
FakeStatus(200),
|
|
])
|
|
self.assertEqual(len(status_codes), len(ec_archive_bodies))
|
|
with mocked_http_conn(*status_codes, body_iter=ec_archive_bodies,
|
|
headers=headers) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(len(log.requests),
|
|
self.policy.ec_n_unique_fragments)
|
|
|
|
def test_GET_with_concurrent_ec_extra_requests(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-454]
|
|
etag = md5(test_data, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
ts = self.ts()
|
|
headers = []
|
|
for i, body in enumerate(ec_archive_bodies):
|
|
headers.append({
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(body),
|
|
'X-Object-Sysmeta-Ec-Frag-Index':
|
|
self.policy.get_backend_index(i),
|
|
'X-Backend-Timestamp': ts.internal,
|
|
'X-Timestamp': ts.normal,
|
|
'X-Backend-Durable-Timestamp': ts.internal,
|
|
'X-Backend-Data-Timestamp': ts.internal,
|
|
})
|
|
policy_opts = self.app.get_policy_options(self.policy)
|
|
policy_opts.concurrent_ec_extra_requests = self.policy.ec_nparity - 1
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
# w/o concurrent_gets ec_extra_requests has no effect
|
|
status_codes = [200] * self.policy.ec_ndata
|
|
with mocked_http_conn(*status_codes, body_iter=ec_archive_bodies,
|
|
headers=headers) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(len(log.requests), self.policy.ec_ndata)
|
|
self.assertEqual(resp.body, test_data)
|
|
|
|
policy_opts.concurrent_gets = True
|
|
status_codes = [200] * (self.policy.object_ring.replicas - 1)
|
|
with mocked_http_conn(*status_codes, body_iter=ec_archive_bodies,
|
|
headers=headers) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(len(log.requests),
|
|
self.policy.object_ring.replicas - 1)
|
|
self.assertEqual(resp.body, test_data)
|
|
|
|
def test_GET_with_body(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o')
|
|
# turn a real body into fragments
|
|
segment_size = self.policy.ec_segment_size
|
|
real_body = (b'asdf' * segment_size)[:-10]
|
|
# split it up into chunks
|
|
chunks = [real_body[x:x + segment_size]
|
|
for x in range(0, len(real_body), segment_size)]
|
|
fragment_payloads = []
|
|
for chunk in chunks:
|
|
fragments = self.policy.pyeclib_driver.encode(chunk)
|
|
if not fragments:
|
|
break
|
|
fragment_payloads.append(
|
|
fragments * self.policy.ec_duplication_factor)
|
|
# sanity
|
|
sanity_body = b''
|
|
for fragment_payload in fragment_payloads:
|
|
sanity_body += self.policy.pyeclib_driver.decode(
|
|
fragment_payload)
|
|
self.assertEqual(len(real_body), len(sanity_body))
|
|
self.assertEqual(real_body, sanity_body)
|
|
|
|
# list(zip(...)) for py3 compatibility (zip is lazy there)
|
|
node_fragments = list(zip(*fragment_payloads))
|
|
self.assertEqual(len(node_fragments), self.replicas()) # sanity
|
|
headers = {'X-Object-Sysmeta-Ec-Content-Length': str(len(real_body))}
|
|
responses = [(200, b''.join(node_fragments[i]), headers)
|
|
for i in range(POLICIES.default.ec_ndata)]
|
|
status_codes, body_iter, headers = zip(*responses)
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(len(real_body), len(resp.body))
|
|
self.assertEqual(real_body, resp.body)
|
|
|
|
def test_GET_with_frags_swapped_around(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-657]
|
|
etag = md5(test_data, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
|
|
_part, primary_nodes = self.obj_ring.get_nodes('a', 'c', 'o')
|
|
|
|
node_key = lambda n: (n['ip'], n['port'])
|
|
backend_index = self.policy.get_backend_index
|
|
ts = self.ts()
|
|
|
|
response_map = {
|
|
node_key(n): StubResponse(
|
|
200, ec_archive_bodies[backend_index(i)], {
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(test_data),
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Frag-Index': backend_index(i),
|
|
'X-Timestamp': ts.normal,
|
|
'X-Backend-Timestamp': ts.internal
|
|
}) for i, n in enumerate(primary_nodes)
|
|
}
|
|
|
|
# swap a parity response into a data node
|
|
data_node = random.choice(primary_nodes[:self.policy.ec_ndata])
|
|
parity_node = random.choice(
|
|
primary_nodes[
|
|
self.policy.ec_ndata:self.policy.ec_n_unique_fragments])
|
|
(response_map[node_key(data_node)],
|
|
response_map[node_key(parity_node)]) = \
|
|
(response_map[node_key(parity_node)],
|
|
response_map[node_key(data_node)])
|
|
|
|
def get_response(req):
|
|
req_key = (req['ip'], req['port'])
|
|
return response_map.pop(req_key)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(len(log), self.policy.ec_ndata)
|
|
self.assertEqual(len(response_map),
|
|
len(primary_nodes) - self.policy.ec_ndata)
|
|
|
|
def test_GET_with_no_success(self):
|
|
node_frags = [[]] * 28 # no frags on any node
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 404)
|
|
self.assertEqual(len(log), 2 * self.replicas())
|
|
|
|
def test_GET_with_only_handoffs(self):
|
|
obj1 = self._make_ec_object_stub()
|
|
|
|
node_frags = [[]] * self.replicas() # all primaries missing
|
|
node_frags = node_frags + [ # handoffs
|
|
{'obj': obj1, 'frag': 0},
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj1, 'frag': 5},
|
|
{'obj': obj1, 'frag': 6},
|
|
{'obj': obj1, 'frag': 7},
|
|
{'obj': obj1, 'frag': 8},
|
|
{'obj': obj1, 'frag': 9},
|
|
{'obj': obj1, 'frag': 10}, # parity
|
|
{'obj': obj1, 'frag': 11}, # parity
|
|
{'obj': obj1, 'frag': 12}, # parity
|
|
{'obj': obj1, 'frag': 13}, # parity
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj1['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj1['etag'])
|
|
|
|
collected_responses = defaultdict(list)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
index = conn.resp.headers['X-Object-Sysmeta-Ec-Frag-Index']
|
|
collected_responses[etag].append(index)
|
|
|
|
# GETS would be required to all primaries and then ndata handoffs
|
|
self.assertEqual(len(log), self.replicas() + self.policy.ec_ndata)
|
|
self.assertEqual(2, len(collected_responses))
|
|
# 404s
|
|
self.assertEqual(self.replicas(), len(collected_responses[None]))
|
|
self.assertEqual(self.policy.ec_ndata,
|
|
len(collected_responses[obj1['etag']]))
|
|
|
|
def test_GET_with_single_missed_overwrite_does_not_need_handoff(self):
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
|
|
node_frags = [
|
|
{'obj': obj2, 'frag': 0},
|
|
{'obj': obj2, 'frag': 1},
|
|
{'obj': obj1, 'frag': 2}, # missed over write
|
|
{'obj': obj2, 'frag': 3},
|
|
{'obj': obj2, 'frag': 4},
|
|
{'obj': obj2, 'frag': 5},
|
|
{'obj': obj2, 'frag': 6},
|
|
{'obj': obj2, 'frag': 7},
|
|
{'obj': obj2, 'frag': 8},
|
|
{'obj': obj2, 'frag': 9},
|
|
{'obj': obj2, 'frag': 10}, # parity
|
|
{'obj': obj2, 'frag': 11}, # parity
|
|
{'obj': obj2, 'frag': 12}, # parity
|
|
{'obj': obj2, 'frag': 13}, # parity
|
|
# {'obj': obj2, 'frag': 2}, # handoff (not used in this test)
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with mock.patch('swift.proxy.server.shuffle', lambda n: n), \
|
|
capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj2['etag'])
|
|
closed_conn = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
closed_conn[etag].add(conn.closed)
|
|
self.assertEqual({
|
|
obj1['etag']: {True},
|
|
obj2['etag']: {False},
|
|
}, closed_conn)
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj2['etag'])
|
|
self.assertEqual({True}, {conn.closed for conn in log})
|
|
|
|
collected_responses = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
index = conn.resp.headers['X-Object-Sysmeta-Ec-Frag-Index']
|
|
collected_responses[etag].add(index)
|
|
|
|
self.assertEqual(len(log), self.policy.ec_ndata + 1)
|
|
expected = {
|
|
obj1['etag']: 1,
|
|
obj2['etag']: self.policy.ec_ndata,
|
|
}
|
|
self.assertEqual(expected, {
|
|
e: len(f) for e, f in collected_responses.items()})
|
|
|
|
def test_GET_with_many_missed_overwrite_will_need_handoff(self):
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
|
|
node_frags = [
|
|
{'obj': obj2, 'frag': 0},
|
|
{'obj': obj2, 'frag': 1},
|
|
{'obj': obj1, 'frag': 2}, # missed
|
|
{'obj': obj2, 'frag': 3},
|
|
{'obj': obj2, 'frag': 4},
|
|
{'obj': obj2, 'frag': 5},
|
|
{'obj': obj1, 'frag': 6}, # missed
|
|
{'obj': obj2, 'frag': 7},
|
|
{'obj': obj2, 'frag': 8},
|
|
{'obj': obj1, 'frag': 9}, # missed
|
|
{'obj': obj1, 'frag': 10}, # missed
|
|
{'obj': obj1, 'frag': 11}, # missed
|
|
{'obj': obj2, 'frag': 12},
|
|
{'obj': obj2, 'frag': 13},
|
|
{'obj': obj2, 'frag': 6}, # handoff
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj2['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj2['etag'])
|
|
|
|
collected_responses = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
index = conn.resp.headers['X-Object-Sysmeta-Ec-Frag-Index']
|
|
collected_responses[etag].add(index)
|
|
|
|
# there's not enough of the obj2 etag on the primaries, we would
|
|
# have collected responses for both etags, and would have made
|
|
# one more request to the handoff node
|
|
self.assertEqual(len(log), self.replicas() + 1)
|
|
self.assertEqual(len(collected_responses), 2)
|
|
|
|
# ... regardless we should never need to fetch more than ec_ndata
|
|
# frags for any given etag
|
|
for etag, frags in collected_responses.items():
|
|
self.assertLessEqual(len(frags), self.policy.ec_ndata,
|
|
'collected %s frags for etag %s' % (
|
|
len(frags), etag))
|
|
|
|
def test_GET_with_missing_and_mixed_frags_will_dig_deep_but_succeed(self):
|
|
obj1 = self._make_ec_object_stub(pattern='obj1', timestamp=self.ts())
|
|
obj2 = self._make_ec_object_stub(pattern='obj2', timestamp=self.ts())
|
|
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0},
|
|
{'obj': obj2, 'frag': 0},
|
|
[],
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj2, 'frag': 1},
|
|
[],
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj2, 'frag': 2},
|
|
[],
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj2, 'frag': 3},
|
|
[],
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj2, 'frag': 4},
|
|
[],
|
|
{'obj': obj1, 'frag': 5},
|
|
{'obj': obj2, 'frag': 5},
|
|
[],
|
|
{'obj': obj1, 'frag': 6},
|
|
{'obj': obj2, 'frag': 6},
|
|
[],
|
|
{'obj': obj1, 'frag': 7},
|
|
{'obj': obj2, 'frag': 7},
|
|
[],
|
|
{'obj': obj1, 'frag': 8},
|
|
{'obj': obj2, 'frag': 8},
|
|
[],
|
|
{'obj': obj2, 'frag': 9},
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj2['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj2['etag'])
|
|
|
|
collected_responses = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
index = conn.resp.headers['X-Object-Sysmeta-Ec-Frag-Index']
|
|
collected_responses[etag].add(index)
|
|
|
|
# we go exactly as long as we have to, finding two different
|
|
# etags and some 404's (i.e. collected_responses[None])
|
|
self.assertEqual(len(log), len(node_frags))
|
|
self.assertEqual(len(collected_responses), 3)
|
|
|
|
# ... regardless we should never need to fetch more than ec_ndata
|
|
# frags for any given etag
|
|
for etag, frags in collected_responses.items():
|
|
self.assertLessEqual(len(frags), self.policy.ec_ndata,
|
|
'collected %s frags for etag %s' % (
|
|
len(frags), etag))
|
|
|
|
def test_GET_with_missing_and_mixed_frags_will_dig_deep_but_stop(self):
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0},
|
|
{'obj': obj2, 'frag': 0},
|
|
[],
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj2, 'frag': 1},
|
|
[],
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj2, 'frag': 2},
|
|
[],
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj2, 'frag': 3},
|
|
[],
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj2, 'frag': 4},
|
|
[],
|
|
{'obj': obj1, 'frag': 5},
|
|
{'obj': obj2, 'frag': 5},
|
|
[],
|
|
{'obj': obj1, 'frag': 6},
|
|
{'obj': obj2, 'frag': 6},
|
|
[],
|
|
{'obj': obj1, 'frag': 7},
|
|
{'obj': obj2, 'frag': 7},
|
|
[],
|
|
{'obj': obj1, 'frag': 8},
|
|
{'obj': obj2, 'frag': 8},
|
|
[],
|
|
# handoffs are iter'd in order so proxy will see 404 from this
|
|
# final handoff
|
|
[],
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
closed_conn = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers.get('X-Object-Sysmeta-Ec-Etag')
|
|
closed_conn[etag].add(conn.closed)
|
|
self.assertEqual({
|
|
obj1['etag']: {True},
|
|
obj2['etag']: {True},
|
|
None: {True},
|
|
}, dict(closed_conn))
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
collected_responses = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
index = conn.resp.headers['X-Object-Sysmeta-Ec-Frag-Index']
|
|
collected_responses[etag].add(index)
|
|
|
|
# default node_iter will exhaust at 2 * replicas
|
|
self.assertEqual(len(log), 2 * self.replicas())
|
|
self.assertEqual(len(collected_responses), 3)
|
|
|
|
# ... regardless we should never need to fetch more than ec_ndata
|
|
# frags for any given etag
|
|
for etag, frags in collected_responses.items():
|
|
self.assertLessEqual(len(frags), self.policy.ec_ndata,
|
|
'collected %s frags for etag %s' % (
|
|
len(frags), etag))
|
|
|
|
def test_GET_with_duplicate_and_hidden_frag_indexes(self):
|
|
obj1 = self._make_ec_object_stub()
|
|
# proxy should ignore duplicated frag indexes and continue search for
|
|
# a set of unique indexes, finding last one on a handoff
|
|
node_frags = [
|
|
[{'obj': obj1, 'frag': 0}, {'obj': obj1, 'frag': 5}],
|
|
{'obj': obj1, 'frag': 0}, # duplicate frag
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj1, 'frag': 1}, # duplicate frag
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj1, 'frag': 2}, # duplicate frag
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj1, 'frag': 3}, # duplicate frag
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj1, 'frag': 4}, # duplicate frag
|
|
{'obj': obj1, 'frag': 10},
|
|
{'obj': obj1, 'frag': 11},
|
|
{'obj': obj1, 'frag': 12},
|
|
{'obj': obj1, 'frag': 13},
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj1['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj1['etag'])
|
|
|
|
# Expect a maximum of one request to each primary plus one extra
|
|
# request to node 1. Actual value could be less if the extra request
|
|
# occurs and quorum is reached before requests to nodes with a
|
|
# duplicate frag.
|
|
self.assertLessEqual(len(log), self.replicas() + 1)
|
|
collected_indexes = defaultdict(list)
|
|
for conn in log:
|
|
fi = conn.resp.headers.get('X-Object-Sysmeta-Ec-Frag-Index')
|
|
if fi is not None:
|
|
collected_indexes[fi].append(conn)
|
|
self.assertEqual(len(collected_indexes), self.policy.ec_ndata)
|
|
|
|
def test_GET_with_missing_and_mixed_frags_may_503(self):
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
# we get a 503 when all the handoffs return 200
|
|
node_frags = [[]] * self.replicas() # primaries have no frags
|
|
node_frags = node_frags + [ # handoffs all have frags
|
|
{'obj': obj1, 'frag': 0},
|
|
{'obj': obj2, 'frag': 0},
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj2, 'frag': 1},
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj2, 'frag': 2},
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj2, 'frag': 3},
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj2, 'frag': 4},
|
|
{'obj': obj1, 'frag': 5},
|
|
{'obj': obj2, 'frag': 5},
|
|
{'obj': obj1, 'frag': 6},
|
|
{'obj': obj2, 'frag': 6},
|
|
]
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 503)
|
|
# never get a quorum so all nodes are searched
|
|
self.assertEqual(len(log), 2 * self.replicas())
|
|
collected_indexes = defaultdict(list)
|
|
for conn in log:
|
|
fi = conn.resp.headers.get('X-Object-Sysmeta-Ec-Frag-Index')
|
|
if fi is not None:
|
|
collected_indexes[fi].append(conn)
|
|
self.assertEqual(len(collected_indexes), 7)
|
|
|
|
def test_GET_with_mixed_nondurable_frags_and_will_404(self):
|
|
# all nodes have a frag but there is no one set that reaches quorum,
|
|
# which means there is no backend 404 response, but proxy should still
|
|
# return 404 rather than 503
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
obj3 = self._make_ec_object_stub(pattern='obj3')
|
|
obj4 = self._make_ec_object_stub(pattern='obj4')
|
|
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0, 'durable': False},
|
|
{'obj': obj2, 'frag': 0, 'durable': False},
|
|
{'obj': obj3, 'frag': 0, 'durable': False},
|
|
{'obj': obj1, 'frag': 1, 'durable': False},
|
|
{'obj': obj2, 'frag': 1, 'durable': False},
|
|
{'obj': obj3, 'frag': 1, 'durable': False},
|
|
{'obj': obj1, 'frag': 2, 'durable': False},
|
|
{'obj': obj2, 'frag': 2, 'durable': False},
|
|
{'obj': obj3, 'frag': 2, 'durable': False},
|
|
{'obj': obj1, 'frag': 3, 'durable': False},
|
|
{'obj': obj2, 'frag': 3, 'durable': False},
|
|
{'obj': obj3, 'frag': 3, 'durable': False},
|
|
{'obj': obj1, 'frag': 4, 'durable': False},
|
|
{'obj': obj2, 'frag': 4, 'durable': False},
|
|
{'obj': obj3, 'frag': 4, 'durable': False},
|
|
{'obj': obj1, 'frag': 5, 'durable': False},
|
|
{'obj': obj2, 'frag': 5, 'durable': False},
|
|
{'obj': obj3, 'frag': 5, 'durable': False},
|
|
{'obj': obj1, 'frag': 6, 'durable': False},
|
|
{'obj': obj2, 'frag': 6, 'durable': False},
|
|
{'obj': obj3, 'frag': 6, 'durable': False},
|
|
{'obj': obj1, 'frag': 7, 'durable': False},
|
|
{'obj': obj2, 'frag': 7, 'durable': False},
|
|
{'obj': obj3, 'frag': 7, 'durable': False},
|
|
{'obj': obj1, 'frag': 8, 'durable': False},
|
|
{'obj': obj2, 'frag': 8, 'durable': False},
|
|
{'obj': obj3, 'frag': 8, 'durable': False},
|
|
{'obj': obj4, 'frag': 8, 'durable': False},
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
collected_etags = set()
|
|
collected_status = set()
|
|
closed_conn = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
collected_etags.add(etag)
|
|
collected_status.add(conn.resp.status)
|
|
closed_conn[etag].add(conn.closed)
|
|
|
|
# default node_iter will exhaust at 2 * replicas
|
|
self.assertEqual(len(log), 2 * self.replicas())
|
|
self.assertEqual(
|
|
{obj1['etag'], obj2['etag'], obj3['etag'], obj4['etag']},
|
|
collected_etags)
|
|
self.assertEqual({200}, collected_status)
|
|
self.assertEqual({
|
|
obj1['etag']: {True},
|
|
obj2['etag']: {True},
|
|
obj3['etag']: {True},
|
|
obj4['etag']: {True},
|
|
}, closed_conn)
|
|
|
|
def test_GET_with_mixed_durable_and_nondurable_frags_will_503(self):
|
|
# all nodes have a frag but there is no one set that reaches quorum,
|
|
# but since one is marked durable we *should* be able to reconstruct,
|
|
# so proxy should 503
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
obj3 = self._make_ec_object_stub(pattern='obj3')
|
|
obj4 = self._make_ec_object_stub(pattern='obj4')
|
|
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0, 'durable': False},
|
|
{'obj': obj2, 'frag': 0, 'durable': False},
|
|
{'obj': obj3, 'frag': 0, 'durable': False},
|
|
{'obj': obj1, 'frag': 1, 'durable': False},
|
|
{'obj': obj2, 'frag': 1, 'durable': False},
|
|
{'obj': obj3, 'frag': 1, 'durable': False},
|
|
{'obj': obj1, 'frag': 2, 'durable': False},
|
|
{'obj': obj2, 'frag': 2, 'durable': False},
|
|
{'obj': obj3, 'frag': 2, 'durable': False},
|
|
{'obj': obj1, 'frag': 3, 'durable': False},
|
|
{'obj': obj2, 'frag': 3, 'durable': False},
|
|
{'obj': obj3, 'frag': 3, 'durable': False},
|
|
{'obj': obj1, 'frag': 4, 'durable': False},
|
|
{'obj': obj2, 'frag': 4, 'durable': False},
|
|
{'obj': obj3, 'frag': 4, 'durable': False},
|
|
{'obj': obj1, 'frag': 5, 'durable': False},
|
|
{'obj': obj2, 'frag': 5, 'durable': False},
|
|
{'obj': obj3, 'frag': 5, 'durable': False},
|
|
{'obj': obj1, 'frag': 6, 'durable': False},
|
|
{'obj': obj2, 'frag': 6, 'durable': False},
|
|
{'obj': obj3, 'frag': 6, 'durable': False},
|
|
{'obj': obj1, 'frag': 7, 'durable': False},
|
|
{'obj': obj2, 'frag': 7, 'durable': False},
|
|
{'obj': obj3, 'frag': 7},
|
|
{'obj': obj1, 'frag': 8, 'durable': False},
|
|
{'obj': obj2, 'frag': 8, 'durable': False},
|
|
{'obj': obj3, 'frag': 8, 'durable': False},
|
|
{'obj': obj4, 'frag': 8, 'durable': False},
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
closed_conn = defaultdict(set)
|
|
collected_etags = set()
|
|
collected_status = set()
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
collected_etags.add(etag)
|
|
collected_status.add(conn.resp.status)
|
|
closed_conn[etag].add(conn.closed)
|
|
|
|
# default node_iter will exhaust at 2 * replicas
|
|
self.assertEqual(len(log), 2 * self.replicas())
|
|
self.assertEqual(
|
|
{obj1['etag'], obj2['etag'], obj3['etag'], obj4['etag']},
|
|
collected_etags)
|
|
self.assertEqual({200}, collected_status)
|
|
self.assertEqual({
|
|
obj1['etag']: {True},
|
|
obj2['etag']: {True},
|
|
obj3['etag']: {True},
|
|
obj4['etag']: {True},
|
|
}, closed_conn)
|
|
|
|
def test_GET_with_mixed_durable_frags_and_no_quorum_will_503(self):
|
|
# all nodes have a frag but there is no one set that reaches quorum,
|
|
# and since at least one is marked durable we *should* be able to
|
|
# reconstruct, so proxy will 503
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
obj3 = self._make_ec_object_stub(pattern='obj3')
|
|
obj4 = self._make_ec_object_stub(pattern='obj4')
|
|
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0},
|
|
{'obj': obj2, 'frag': 0},
|
|
{'obj': obj3, 'frag': 0},
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj2, 'frag': 1},
|
|
{'obj': obj3, 'frag': 1},
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj2, 'frag': 2},
|
|
{'obj': obj3, 'frag': 2},
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj2, 'frag': 3},
|
|
{'obj': obj3, 'frag': 3},
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj2, 'frag': 4},
|
|
{'obj': obj3, 'frag': 4},
|
|
{'obj': obj1, 'frag': 5},
|
|
{'obj': obj2, 'frag': 5},
|
|
{'obj': obj3, 'frag': 5},
|
|
{'obj': obj1, 'frag': 6},
|
|
{'obj': obj2, 'frag': 6},
|
|
{'obj': obj3, 'frag': 6},
|
|
{'obj': obj1, 'frag': 7},
|
|
{'obj': obj2, 'frag': 7},
|
|
{'obj': obj3, 'frag': 7},
|
|
{'obj': obj1, 'frag': 8},
|
|
{'obj': obj2, 'frag': 8},
|
|
{'obj': obj3, 'frag': 8},
|
|
{'obj': obj4, 'frag': 8},
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
for conn in log:
|
|
etag = conn.resp.headers.get('X-Object-Sysmeta-Ec-Etag')
|
|
|
|
collected_etags = set()
|
|
collected_status = set()
|
|
closed_conn = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
collected_etags.add(etag)
|
|
collected_status.add(conn.resp.status)
|
|
closed_conn[etag].add(conn.closed)
|
|
|
|
# default node_iter will exhaust at 2 * replicas
|
|
self.assertEqual(len(log), 2 * self.replicas())
|
|
self.assertEqual(
|
|
{obj1['etag'], obj2['etag'], obj3['etag'], obj4['etag']},
|
|
collected_etags)
|
|
self.assertEqual({200}, collected_status)
|
|
self.assertEqual({
|
|
obj1['etag']: {True},
|
|
obj2['etag']: {True},
|
|
obj3['etag']: {True},
|
|
obj4['etag']: {True},
|
|
}, closed_conn)
|
|
|
|
def test_GET_with_quorum_durable_files(self):
|
|
# verify that only (ec_nparity + 1) nodes need to be durable for a GET
|
|
# to be completed with ec_ndata requests.
|
|
obj1 = self._make_ec_object_stub()
|
|
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0, 'durable': True}, # durable
|
|
{'obj': obj1, 'frag': 1, 'durable': True}, # durable
|
|
{'obj': obj1, 'frag': 2, 'durable': True}, # durable
|
|
{'obj': obj1, 'frag': 3, 'durable': True}, # durable
|
|
{'obj': obj1, 'frag': 4, 'durable': True}, # durable
|
|
{'obj': obj1, 'frag': 5, 'durable': False},
|
|
{'obj': obj1, 'frag': 6, 'durable': False},
|
|
{'obj': obj1, 'frag': 7, 'durable': False},
|
|
{'obj': obj1, 'frag': 8, 'durable': False},
|
|
{'obj': obj1, 'frag': 9, 'durable': False},
|
|
{'obj': obj1, 'frag': 10, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 11, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 12, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 13, 'durable': False}, # parity
|
|
] + [[]] * self.replicas() # handoffs all 404
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj1['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj1['etag'])
|
|
|
|
self.assertGreaterEqual(len(log), self.policy.ec_ndata)
|
|
collected_durables = []
|
|
for conn in log:
|
|
if not conn.resp.headers.get('X-Backend-Data-Timestamp'):
|
|
continue
|
|
if (conn.resp.headers.get('X-Backend-Durable-Timestamp')
|
|
== conn.resp.headers.get('X-Backend-Data-Timestamp')):
|
|
collected_durables.append(conn)
|
|
# because nodes are shuffled we can't be sure how many durables are
|
|
# returned but it must be at least 1 and cannot exceed 5
|
|
self.assertLessEqual(len(collected_durables), 5)
|
|
self.assertGreaterEqual(len(collected_durables), 1)
|
|
|
|
def test_GET_with_single_durable_file(self):
|
|
# verify that a single durable is sufficient for a GET
|
|
# to be completed with ec_ndata requests.
|
|
obj1 = self._make_ec_object_stub()
|
|
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0, 'durable': True}, # durable
|
|
{'obj': obj1, 'frag': 1, 'durable': False},
|
|
{'obj': obj1, 'frag': 2, 'durable': False},
|
|
{'obj': obj1, 'frag': 3, 'durable': False},
|
|
{'obj': obj1, 'frag': 4, 'durable': False},
|
|
{'obj': obj1, 'frag': 5, 'durable': False},
|
|
{'obj': obj1, 'frag': 6, 'durable': False},
|
|
{'obj': obj1, 'frag': 7, 'durable': False},
|
|
{'obj': obj1, 'frag': 8, 'durable': False},
|
|
{'obj': obj1, 'frag': 9, 'durable': False},
|
|
{'obj': obj1, 'frag': 10, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 11, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 12, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 13, 'durable': False}, # parity
|
|
] + [[]] * self.replicas() # handoffs all 404
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj1['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj1['etag'])
|
|
|
|
collected_durables = []
|
|
for conn in log:
|
|
if not conn.resp.headers.get('X-Backend-Data-Timestamp'):
|
|
continue
|
|
if (conn.resp.headers.get('X-Backend-Durable-Timestamp')
|
|
== conn.resp.headers.get('X-Backend-Data-Timestamp')):
|
|
collected_durables.append(conn)
|
|
# because nodes are shuffled we can't be sure how many non-durables
|
|
# are returned before the durable, but we do expect a single durable
|
|
self.assertEqual(1, len(collected_durables))
|
|
|
|
def test_GET_with_no_durable_files(self):
|
|
# verify that at least one durable is necessary for a successful GET
|
|
obj1 = self._make_ec_object_stub()
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0, 'durable': False},
|
|
{'obj': obj1, 'frag': 1, 'durable': False},
|
|
{'obj': obj1, 'frag': 2, 'durable': False},
|
|
{'obj': obj1, 'frag': 3, 'durable': False},
|
|
{'obj': obj1, 'frag': 4, 'durable': False},
|
|
{'obj': obj1, 'frag': 5, 'durable': False},
|
|
{'obj': obj1, 'frag': 6, 'durable': False},
|
|
{'obj': obj1, 'frag': 7, 'durable': False},
|
|
{'obj': obj1, 'frag': 8, 'durable': False},
|
|
{'obj': obj1, 'frag': 9, 'durable': False},
|
|
{'obj': obj1, 'frag': 10, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 11, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 12, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 13, 'durable': False}, # parity
|
|
] + [[]] * self.replicas() # handoffs
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 404)
|
|
# all 28 nodes tried with an optimistic get, none are durable and none
|
|
# report having a durable timestamp
|
|
self.assertEqual(28, len(log))
|
|
self.assertEqual([], self.app.logger.get_lines_for_level('error'))
|
|
|
|
def test_GET_nondurable_when_node_iter_runs_out_of_nodes(self):
|
|
policy_opts = self.app.get_policy_options(self.policy)
|
|
policy_opts.concurrent_gets = True
|
|
policy_opts.concurrent_ec_extra_requests = 1
|
|
policy_opts.concurrency_timeout = 0.1
|
|
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': i, 'durable': False}
|
|
for i in range(self.policy.ec_ndata + 1)
|
|
]
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
# limit remaining nodes
|
|
obj_ring = self.app.get_object_ring(int(self.policy))
|
|
part, nodes = obj_ring.get_nodes('a', 'c', 'o')
|
|
nodes.extend(list(obj_ring.get_more_nodes(part)))
|
|
for dev in nodes[self.policy.ec_ndata + 1:]:
|
|
self.app.error_limiter.limit(dev)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
# non-durable responds 404
|
|
self.assertEqual(resp.status_int, 404)
|
|
self.assertEqual(len(log), self.policy.ec_ndata + 1)
|
|
self.assertEqual([], self.app.logger.get_lines_for_level('error'))
|
|
|
|
def test_GET_with_missing_durable_files_and_mixed_etags(self):
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
|
|
# non-quorate durables for another object won't stop us finding the
|
|
# quorate object
|
|
node_frags = [
|
|
# ec_ndata - 1 frags of obj2 are available and durable
|
|
{'obj': obj2, 'frag': 0, 'durable': True},
|
|
{'obj': obj2, 'frag': 1, 'durable': True},
|
|
{'obj': obj2, 'frag': 2, 'durable': True},
|
|
{'obj': obj2, 'frag': 3, 'durable': True},
|
|
{'obj': obj2, 'frag': 4, 'durable': True},
|
|
{'obj': obj2, 'frag': 5, 'durable': True},
|
|
{'obj': obj2, 'frag': 6, 'durable': True},
|
|
{'obj': obj2, 'frag': 7, 'durable': True},
|
|
{'obj': obj2, 'frag': 8, 'durable': True},
|
|
# ec_ndata frags of obj1 are available and one is durable
|
|
{'obj': obj1, 'frag': 0, 'durable': False},
|
|
{'obj': obj1, 'frag': 1, 'durable': False},
|
|
{'obj': obj1, 'frag': 2, 'durable': False},
|
|
{'obj': obj1, 'frag': 3, 'durable': False},
|
|
{'obj': obj1, 'frag': 4, 'durable': False},
|
|
{'obj': obj1, 'frag': 5, 'durable': False},
|
|
{'obj': obj1, 'frag': 6, 'durable': False},
|
|
{'obj': obj1, 'frag': 7, 'durable': False},
|
|
{'obj': obj1, 'frag': 8, 'durable': False},
|
|
{'obj': obj1, 'frag': 9, 'durable': True},
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
closed_conn = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers.get('X-Object-Sysmeta-Ec-Etag')
|
|
closed_conn[etag].add(conn.closed)
|
|
self.assertEqual({
|
|
obj1['etag']: {False},
|
|
obj2['etag']: {True},
|
|
}, closed_conn)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj1['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj1['etag'])
|
|
|
|
# Quorum of non-durables for a different object won't
|
|
# prevent us hunting down the durable object
|
|
node_frags = [
|
|
# primaries
|
|
{'obj': obj2, 'frag': 0, 'durable': False},
|
|
{'obj': obj2, 'frag': 1, 'durable': False},
|
|
{'obj': obj2, 'frag': 2, 'durable': False},
|
|
{'obj': obj2, 'frag': 3, 'durable': False},
|
|
{'obj': obj2, 'frag': 4, 'durable': False},
|
|
{'obj': obj2, 'frag': 5, 'durable': False},
|
|
{'obj': obj2, 'frag': 6, 'durable': False},
|
|
{'obj': obj2, 'frag': 7, 'durable': False},
|
|
{'obj': obj2, 'frag': 8, 'durable': False},
|
|
{'obj': obj2, 'frag': 9, 'durable': False},
|
|
{'obj': obj2, 'frag': 10, 'durable': False},
|
|
{'obj': obj2, 'frag': 11, 'durable': False},
|
|
{'obj': obj2, 'frag': 12, 'durable': False},
|
|
{'obj': obj2, 'frag': 13, 'durable': False},
|
|
# handoffs
|
|
{'obj': obj1, 'frag': 0, 'durable': False},
|
|
{'obj': obj1, 'frag': 1, 'durable': False},
|
|
{'obj': obj1, 'frag': 2, 'durable': False},
|
|
{'obj': obj1, 'frag': 3, 'durable': False},
|
|
{'obj': obj1, 'frag': 4, 'durable': False},
|
|
{'obj': obj1, 'frag': 5, 'durable': False},
|
|
{'obj': obj1, 'frag': 6, 'durable': False},
|
|
{'obj': obj1, 'frag': 7, 'durable': False},
|
|
{'obj': obj1, 'frag': 8, 'durable': False},
|
|
{'obj': obj1, 'frag': 9, 'durable': False},
|
|
{'obj': obj1, 'frag': 10, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 11, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 12, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 13, 'durable': True}, # parity
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj1['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj1['etag'])
|
|
|
|
def test_GET_with_missing_durables_and_older_durables(self):
|
|
# scenario: non-durable frags of newer obj1 obscure all durable frags
|
|
# of older obj2, so first 14 requests result in a non-durable set.
|
|
# At that point (or before) the proxy knows that a durable set of
|
|
# frags for obj2 exists so will fetch them, requiring another 10
|
|
# directed requests.
|
|
obj2 = self._make_ec_object_stub(pattern='obj2', timestamp=self.ts())
|
|
obj1 = self._make_ec_object_stub(pattern='obj1', timestamp=self.ts())
|
|
|
|
node_frags = [
|
|
[{'obj': obj1, 'frag': 0, 'durable': False}], # obj2 missing
|
|
[{'obj': obj1, 'frag': 1, 'durable': False},
|
|
{'obj': obj2, 'frag': 1, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 2, 'durable': False}], # obj2 missing
|
|
[{'obj': obj1, 'frag': 3, 'durable': False},
|
|
{'obj': obj2, 'frag': 3, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 4, 'durable': False},
|
|
{'obj': obj2, 'frag': 4, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 5, 'durable': False},
|
|
{'obj': obj2, 'frag': 5, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 6, 'durable': False},
|
|
{'obj': obj2, 'frag': 6, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 7, 'durable': False},
|
|
{'obj': obj2, 'frag': 7, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 8, 'durable': False},
|
|
{'obj': obj2, 'frag': 8, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 9, 'durable': False}], # obj2 missing
|
|
[{'obj': obj1, 'frag': 10, 'durable': False},
|
|
{'obj': obj2, 'frag': 10, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 11, 'durable': False},
|
|
{'obj': obj2, 'frag': 11, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 12, 'durable': False}], # obj2 missing
|
|
[{'obj': obj1, 'frag': 13, 'durable': False},
|
|
{'obj': obj2, 'frag': 13, 'durable': True}],
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
closed_conn = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers.get('X-Object-Sysmeta-Ec-Etag')
|
|
closed_conn[etag].add(conn.closed)
|
|
self.assertEqual({
|
|
obj1['etag']: {True},
|
|
obj2['etag']: {False},
|
|
}, closed_conn)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj2['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj2['etag'])
|
|
# max: proxy will GET all non-durable obj1 frags and then 10 obj frags
|
|
self.assertLessEqual(len(log), self.replicas() + self.policy.ec_ndata)
|
|
# min: proxy will GET 10 non-durable obj1 frags and then 10 obj frags
|
|
self.assertGreaterEqual(len(log), 2 * self.policy.ec_ndata)
|
|
|
|
def test_GET_with_missing_durables_and_older_obscured_durables(self):
|
|
# scenario: obj3 has 14 frags but only 2 are durable and these are
|
|
# obscured by two non-durable frags of obj1. There is also a single
|
|
# non-durable frag of obj2. The proxy will need to do at least 10
|
|
# GETs to see all the obj3 frags plus 1 more to GET a durable frag.
|
|
# The proxy may also do one more GET if the obj2 frag is found.
|
|
# i.e. 10 + 1 durable for obj3, 2 for obj1 and 1 more if obj2 found
|
|
obj2 = self._make_ec_object_stub(pattern='obj2', timestamp=self.ts())
|
|
obj3 = self._make_ec_object_stub(pattern='obj3', timestamp=self.ts())
|
|
obj1 = self._make_ec_object_stub(pattern='obj1', timestamp=self.ts())
|
|
|
|
node_frags = [
|
|
[{'obj': obj1, 'frag': 0, 'durable': False}, # obj1 frag
|
|
{'obj': obj3, 'frag': 0, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 1, 'durable': False}, # obj1 frag
|
|
{'obj': obj3, 'frag': 1, 'durable': True}],
|
|
[{'obj': obj2, 'frag': 2, 'durable': False}, # obj2 frag
|
|
{'obj': obj3, 'frag': 2, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 3, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 4, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 5, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 6, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 7, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 8, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 9, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 10, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 11, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 12, 'durable': False}],
|
|
[{'obj': obj3, 'frag': 13, 'durable': False}],
|
|
] + [[]] * self.replicas() # handoffs 404
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj3['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj3['etag'])
|
|
self.assertGreaterEqual(len(log), self.policy.ec_ndata + 1)
|
|
self.assertLessEqual(len(log), (self.policy.ec_ndata * 2) + 1)
|
|
|
|
def test_GET_with_missing_durables_and_older_non_durables(self):
|
|
# scenario: non-durable frags of newer obj1 obscure all frags
|
|
# of older obj2, so first 28 requests result in a non-durable set.
|
|
# There are only 10 frags for obj2 and one is not durable.
|
|
obj2 = self._make_ec_object_stub(pattern='obj2', timestamp=self.ts())
|
|
obj1 = self._make_ec_object_stub(pattern='obj1', timestamp=self.ts())
|
|
|
|
node_frags = [
|
|
[{'obj': obj1, 'frag': 0, 'durable': False}], # obj2 missing
|
|
[{'obj': obj1, 'frag': 1, 'durable': False},
|
|
{'obj': obj2, 'frag': 1, 'durable': False}], # obj2 non-durable
|
|
[{'obj': obj1, 'frag': 2, 'durable': False}], # obj2 missing
|
|
[{'obj': obj1, 'frag': 3, 'durable': False},
|
|
{'obj': obj2, 'frag': 3, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 4, 'durable': False},
|
|
{'obj': obj2, 'frag': 4, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 5, 'durable': False},
|
|
{'obj': obj2, 'frag': 5, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 6, 'durable': False},
|
|
{'obj': obj2, 'frag': 6, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 7, 'durable': False},
|
|
{'obj': obj2, 'frag': 7, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 8, 'durable': False},
|
|
{'obj': obj2, 'frag': 8, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 9, 'durable': False}], # obj2 missing
|
|
[{'obj': obj1, 'frag': 10, 'durable': False},
|
|
{'obj': obj2, 'frag': 10, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 11, 'durable': False},
|
|
{'obj': obj2, 'frag': 11, 'durable': True}],
|
|
[{'obj': obj1, 'frag': 12, 'durable': False}], # obj2 missing
|
|
[{'obj': obj1, 'frag': 13, 'durable': False},
|
|
{'obj': obj2, 'frag': 13, 'durable': True}],
|
|
[], # 1 empty primary
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj2['etag'])
|
|
self.assertEqual(md5(
|
|
resp.body, usedforsecurity=False).hexdigest(), obj2['etag'])
|
|
# max: proxy will GET all non-durable obj1 frags and then 10 obj2 frags
|
|
self.assertLessEqual(len(log), self.replicas() + self.policy.ec_ndata)
|
|
# min: proxy will GET 10 non-durable obj1 frags and then 10 obj2 frags
|
|
self.assertGreaterEqual(len(log), 2 * self.policy.ec_ndata)
|
|
|
|
def test_GET_with_mixed_etags_at_same_timestamp(self):
|
|
# this scenario should never occur but if there are somehow
|
|
# fragments for different content at the same timestamp then the
|
|
# object controller should handle it gracefully
|
|
ts = self.ts() # force equal timestamps for two objects
|
|
obj1 = self._make_ec_object_stub(timestamp=ts, pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(timestamp=ts, pattern='obj2')
|
|
self.assertNotEqual(obj1['etag'], obj2['etag']) # sanity
|
|
|
|
node_frags = [
|
|
# 7 frags of obj2 are available and durable
|
|
{'obj': obj2, 'frag': 0, 'durable': True},
|
|
{'obj': obj2, 'frag': 1, 'durable': True},
|
|
{'obj': obj2, 'frag': 2, 'durable': True},
|
|
{'obj': obj2, 'frag': 3, 'durable': True},
|
|
{'obj': obj2, 'frag': 4, 'durable': True},
|
|
{'obj': obj2, 'frag': 5, 'durable': True},
|
|
{'obj': obj2, 'frag': 6, 'durable': True},
|
|
# 7 frags of obj1 are available and durable
|
|
{'obj': obj1, 'frag': 7, 'durable': True},
|
|
{'obj': obj1, 'frag': 8, 'durable': True},
|
|
{'obj': obj1, 'frag': 9, 'durable': True},
|
|
{'obj': obj1, 'frag': 10, 'durable': True},
|
|
{'obj': obj1, 'frag': 11, 'durable': True},
|
|
{'obj': obj1, 'frag': 12, 'durable': True},
|
|
{'obj': obj1, 'frag': 13, 'durable': True},
|
|
] + [[]] * self.replicas() # handoffs
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
# read body to provoke any EC decode errors
|
|
self.assertTrue(resp.body)
|
|
|
|
self.assertEqual(resp.status_int, 503)
|
|
self.assertEqual(len(log), self.replicas() * 2)
|
|
collected_etags = set()
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
collected_etags.add(etag) # will be None from handoffs
|
|
self.assertEqual({obj1['etag'], obj2['etag'], None}, collected_etags)
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(log_lines,
|
|
['Problem with fragment response: ETag mismatch'] * 7
|
|
+ ['Object returning 503 for []'])
|
|
# Note the empty list above -- that log line comes out of
|
|
# best_response but we've already thrown out the "good" responses :-/
|
|
|
|
def test_GET_mixed_success_with_range(self):
|
|
fragment_size = self.policy.fragment_size
|
|
|
|
ec_stub = self._make_ec_object_stub()
|
|
frag_archives = ec_stub['frags']
|
|
frag_archive_size = len(ec_stub['frags'][0])
|
|
|
|
headers = {
|
|
'Content-Type': 'text/plain',
|
|
'Content-Length': fragment_size,
|
|
'Content-Range': 'bytes 0-%s/%s' % (fragment_size - 1,
|
|
frag_archive_size),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(ec_stub['body']),
|
|
'X-Object-Sysmeta-Ec-Etag': ec_stub['etag'],
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
responses = [
|
|
StubResponse(206, frag_archives[0][:fragment_size], headers, 0),
|
|
StubResponse(206, frag_archives[1][:fragment_size], headers, 1),
|
|
StubResponse(206, frag_archives[2][:fragment_size], headers, 2),
|
|
StubResponse(206, frag_archives[3][:fragment_size], headers, 3),
|
|
StubResponse(206, frag_archives[4][:fragment_size], headers, 4),
|
|
# data nodes with old frag
|
|
StubResponse(416, frag_index=5),
|
|
StubResponse(416, frag_index=6),
|
|
StubResponse(206, frag_archives[7][:fragment_size], headers, 7),
|
|
StubResponse(206, frag_archives[8][:fragment_size], headers, 8),
|
|
StubResponse(206, frag_archives[9][:fragment_size], headers, 9),
|
|
# hopefully we ask for two more
|
|
StubResponse(206, frag_archives[10][:fragment_size], headers, 10),
|
|
StubResponse(206, frag_archives[11][:fragment_size], headers, 11),
|
|
]
|
|
|
|
def get_response(req):
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={'Range': 'bytes=0-3'})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(resp.body, b'test')
|
|
self.assertEqual(len(log), self.policy.ec_ndata + 2)
|
|
|
|
# verify that even when last responses to be collected are 416's
|
|
# the shortfall of 2xx responses still triggers extra spawned requests
|
|
responses = [
|
|
StubResponse(206, frag_archives[0][:fragment_size], headers, 0),
|
|
StubResponse(206, frag_archives[1][:fragment_size], headers, 1),
|
|
StubResponse(206, frag_archives[2][:fragment_size], headers, 2),
|
|
StubResponse(206, frag_archives[3][:fragment_size], headers, 3),
|
|
StubResponse(206, frag_archives[4][:fragment_size], headers, 4),
|
|
StubResponse(206, frag_archives[7][:fragment_size], headers, 7),
|
|
StubResponse(206, frag_archives[8][:fragment_size], headers, 8),
|
|
StubResponse(206, frag_archives[9][:fragment_size], headers, 9),
|
|
StubResponse(206, frag_archives[10][:fragment_size], headers, 10),
|
|
# data nodes with old frag
|
|
StubResponse(416, frag_index=5),
|
|
# hopefully we ask for one more
|
|
StubResponse(416, frag_index=6),
|
|
# and hopefully we ask for another
|
|
StubResponse(206, frag_archives[11][:fragment_size], headers, 11),
|
|
]
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={'Range': 'bytes=0-3'})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(resp.body, b'test')
|
|
self.assertEqual(len(log), self.policy.ec_ndata + 2)
|
|
|
|
def test_GET_with_range_unsatisfiable_mixed_success(self):
|
|
responses = [
|
|
StubResponse(416, frag_index=0),
|
|
StubResponse(416, frag_index=1),
|
|
StubResponse(416, frag_index=2),
|
|
StubResponse(416, frag_index=3),
|
|
StubResponse(416, frag_index=4),
|
|
StubResponse(416, frag_index=5),
|
|
StubResponse(416, frag_index=6),
|
|
# sneak a couple bogus extra responses
|
|
StubResponse(404),
|
|
StubResponse(206, frag_index=8),
|
|
# and then just "enough" more 416's
|
|
StubResponse(416, frag_index=9),
|
|
StubResponse(416, frag_index=10),
|
|
StubResponse(416, frag_index=11),
|
|
]
|
|
|
|
def get_response(req):
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=%s-' % 100000000000000})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 416)
|
|
# we're going to engage ndata primaries, plus the bogus extra
|
|
# self.assertEqual(len(log), self.policy.ec_ndata + 2)
|
|
self.assertEqual([c.resp.status for c in log],
|
|
([416] * 7) + [404, 206] + ([416] * 3))
|
|
|
|
def test_GET_with_missing_and_range_unsatisifiable(self):
|
|
responses = [ # not quite ec_ndata frags on primaries
|
|
StubResponse(416, frag_index=0),
|
|
StubResponse(416, frag_index=1),
|
|
StubResponse(416, frag_index=2),
|
|
StubResponse(416, frag_index=3),
|
|
StubResponse(416, frag_index=4),
|
|
StubResponse(416, frag_index=5),
|
|
StubResponse(416, frag_index=6),
|
|
StubResponse(416, frag_index=7),
|
|
StubResponse(416, frag_index=8),
|
|
]
|
|
|
|
def get_response(req):
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=%s-' % 100000000000000})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
# TODO: does 416 make sense without a quorum, or should this be a 404?
|
|
# a non-range GET of same object would return 404
|
|
self.assertEqual(resp.status_int, 416)
|
|
self.assertEqual(len(log), 2 * self.replicas())
|
|
|
|
@patch_policies(
|
|
[ECStoragePolicy(0, name='ec', is_default=True,
|
|
ec_type=DEFAULT_TEST_EC_TYPE, ec_ndata=4,
|
|
ec_nparity=4, ec_segment_size=4096)],
|
|
fake_ring_args=[{'replicas': 8}]
|
|
)
|
|
def test_GET_ndata_equals_nparity_with_missing_and_errors(self):
|
|
# when ec_ndata == ec_nparity it is possible for the shortfall of a bad
|
|
# bucket (412's) to equal ec_ndata; verify that the 412 bucket is still
|
|
# chosen ahead of the initial 'dummy' bad bucket
|
|
POLICIES.default.object_ring.max_more_nodes = 8
|
|
responses = [
|
|
StubResponse(412, frag_index=0),
|
|
StubResponse(412, frag_index=1),
|
|
]
|
|
|
|
def get_response(req):
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=%s-' % 100000000000000})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 412)
|
|
self.assertEqual(len(log), 2 * 8)
|
|
|
|
def test_GET_with_multirange(self):
|
|
self.app.object_chunk_size = 256
|
|
test_body = b'test' * self.policy.ec_segment_size
|
|
ec_stub = make_ec_object_stub(test_body, self.policy, None)
|
|
frag_archives = ec_stub['frags']
|
|
self.assertEqual(len(frag_archives[0]), 1960)
|
|
boundary = b'81eb9c110b32ced5fe'
|
|
|
|
def make_mime_body(frag_archive):
|
|
return b'\r\n'.join([
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 0-489/1960',
|
|
b'',
|
|
frag_archive[0:490],
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 1470-1959/1960',
|
|
b'',
|
|
frag_archive[1470:],
|
|
b'--' + boundary + b'--',
|
|
])
|
|
|
|
obj_resp_bodies = [make_mime_body(fa) for fa
|
|
in ec_stub['frags'][:self.policy.ec_ndata]]
|
|
|
|
headers = {
|
|
'Content-Type': b'multipart/byteranges;boundary=' + boundary,
|
|
'Content-Length': len(obj_resp_bodies[0]),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(ec_stub['body']),
|
|
'X-Object-Sysmeta-Ec-Etag': ec_stub['etag'],
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
|
|
responses = [
|
|
StubResponse(206, body, headers, i)
|
|
for i, body in enumerate(obj_resp_bodies)
|
|
]
|
|
|
|
def get_response(req):
|
|
# there's some math going on here I don't quite understand, the
|
|
# fragment_size is 490 and there's like 4 of them because ec_body
|
|
# is 'test' * segment_size
|
|
self.assertEqual(req['headers']['Range'], 'bytes=0-489,1470-1959')
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=1000-2000,14000-15000'})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(len(log), self.policy.ec_ndata)
|
|
resp_boundary = resp.headers['content-type'].rsplit('=', 1)[1].encode()
|
|
expected = b'\r\n'.join([
|
|
b'--' + resp_boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 1000-2000/16384',
|
|
b'',
|
|
ec_stub['body'][1000:2001],
|
|
b'--' + resp_boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 14000-15000/16384',
|
|
b'',
|
|
ec_stub['body'][14000:15001],
|
|
b'--' + resp_boundary + b'--',
|
|
])
|
|
self.assertEqual(resp.body, expected)
|
|
|
|
def test_GET_with_multirange_slow_body(self):
|
|
self.app.object_chunk_size = 256
|
|
self.app.recoverable_node_timeout = 0.01
|
|
test_body = b'test' * self.policy.ec_segment_size
|
|
ec_stub = make_ec_object_stub(test_body, self.policy, None)
|
|
frag_archives = ec_stub['frags']
|
|
self.assertEqual(len(frag_archives[0]), 1960)
|
|
boundary = b'81eb9c110b32ced5fe'
|
|
|
|
def make_mime_body(frag_archive):
|
|
return b'\r\n'.join([
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 0-489/1960',
|
|
b'',
|
|
frag_archive[0:490],
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 1470-1959/1960',
|
|
b'',
|
|
frag_archive[1470:],
|
|
b'--' + boundary + b'--',
|
|
])
|
|
|
|
obj_resp_bodies = [make_mime_body(fa) for fa
|
|
in ec_stub['frags'][:self.policy.ec_ndata + 1]]
|
|
|
|
headers = {
|
|
'Content-Type': b'multipart/byteranges;boundary=' + boundary,
|
|
'Content-Length': len(obj_resp_bodies[0]),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(ec_stub['body']),
|
|
'X-Object-Sysmeta-Ec-Etag': ec_stub['etag'],
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
|
|
responses = [
|
|
StubResponse(206, body, headers, i,
|
|
# make the first one slow
|
|
slowdown=0.1 if i == 0 else None)
|
|
for i, body in enumerate(obj_resp_bodies)
|
|
]
|
|
|
|
def get_response(req):
|
|
# there's some math going on here I don't quite understand, the
|
|
# fragment_size is 490 and there's like 4 of them because ec_body
|
|
# is 'test' * segment_size
|
|
self.assertEqual(req['headers']['Range'], 'bytes=0-489,1470-1959')
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=1000-2000,14000-15000'})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(len(log), self.policy.ec_ndata + 1)
|
|
resp_boundary = resp.headers['content-type'].rsplit('=', 1)[1].encode()
|
|
expected = b'\r\n'.join([
|
|
b'--' + resp_boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 1000-2000/16384',
|
|
b'',
|
|
ec_stub['body'][1000:2001],
|
|
b'--' + resp_boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 14000-15000/16384',
|
|
b'',
|
|
ec_stub['body'][14000:15001],
|
|
b'--' + resp_boundary + b'--',
|
|
])
|
|
self.assertEqual(resp.body, expected)
|
|
|
|
def test_GET_with_multirange_unable_to_resume(self):
|
|
self.app.object_chunk_size = 256
|
|
self.app.recoverable_node_timeout = 0.01
|
|
test_body = b'test' * self.policy.ec_segment_size
|
|
ec_stub = make_ec_object_stub(test_body, self.policy, None)
|
|
frag_archives = ec_stub['frags']
|
|
self.assertEqual(len(frag_archives[0]), 1960)
|
|
boundary = b'81eb9c110b32ced5fe'
|
|
|
|
def make_mime_body(frag_archive):
|
|
return b'\r\n'.join([
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 0-489/1960',
|
|
b'',
|
|
frag_archive[0:490],
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 1470-1959/1960',
|
|
b'',
|
|
frag_archive[1470:],
|
|
b'--' + boundary + b'--',
|
|
])
|
|
|
|
obj_resp_bodies = [make_mime_body(fa) for fa
|
|
# no extra good responses
|
|
in ec_stub['frags'][:self.policy.ec_ndata]]
|
|
|
|
headers = {
|
|
'Content-Type': b'multipart/byteranges;boundary=' + boundary,
|
|
'Content-Length': len(obj_resp_bodies[0]),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(ec_stub['body']),
|
|
'X-Object-Sysmeta-Ec-Etag': ec_stub['etag'],
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
|
|
responses = [
|
|
StubResponse(206, body, headers, i,
|
|
# make the first one slow
|
|
slowdown=0.1 if i == 0 else None)
|
|
for i, body in enumerate(obj_resp_bodies)
|
|
]
|
|
|
|
def get_response(req):
|
|
# there's some math going on here I don't quite understand, the
|
|
# fragment_size is 490 and there's like 4 of them because ec_body
|
|
# is 'test' * segment_size
|
|
self.assertEqual(req['headers']['Range'], 'bytes=0-489,1470-1959')
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=1000-2000,14000-15000'})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
# note: the error is raised before the resp_iter is read
|
|
self.assertIn(b'Internal Error', resp.body)
|
|
self.assertEqual(resp.status_int, 500)
|
|
self.assertEqual(len(log), self.policy.ec_n_unique_fragments * 2)
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(3, len(log_lines), log_lines)
|
|
self.assertIn('Trying to read next part of EC fragment multi-part GET',
|
|
log_lines[0])
|
|
self.assertIn('Trying to read during GET: ChunkReadTimeout',
|
|
log_lines[1])
|
|
# not the most graceful ending
|
|
self.assertIn('Unhandled exception in request: ChunkReadTimeout',
|
|
log_lines[2])
|
|
|
|
def test_GET_with_multirange_unable_to_resume_body_started(self):
|
|
self.app.object_chunk_size = 256
|
|
self.app.recoverable_node_timeout = 0.01
|
|
test_body = b'test' * self.policy.ec_segment_size
|
|
ec_stub = make_ec_object_stub(test_body, self.policy, None)
|
|
frag_archives = ec_stub['frags']
|
|
self.assertEqual(len(frag_archives[0]), 1960)
|
|
boundary = b'81eb9c110b32ced5fe'
|
|
|
|
def make_mime_body(frag_archive):
|
|
return b'\r\n'.join([
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 0-489/1960',
|
|
b'',
|
|
frag_archive[0:490],
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 1470-1959/1960',
|
|
b'',
|
|
frag_archive[1470:],
|
|
b'--' + boundary + b'--',
|
|
])
|
|
|
|
obj_resp_bodies = [make_mime_body(fa) for fa
|
|
# no extra good responses
|
|
in ec_stub['frags'][:self.policy.ec_ndata]]
|
|
|
|
headers = {
|
|
'Content-Type': b'multipart/byteranges;boundary=' + boundary,
|
|
'Content-Length': len(obj_resp_bodies[0]),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(ec_stub['body']),
|
|
'X-Object-Sysmeta-Ec-Etag': ec_stub['etag'],
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
|
|
responses = [
|
|
StubResponse(206, body, headers, i,
|
|
# make the first one slow
|
|
slowdown=0.1 if i == 0 else None)
|
|
for i, body in enumerate(obj_resp_bodies)
|
|
]
|
|
# the first response serves some bytes before slowing down
|
|
responses[0].slowdown_after = 1000
|
|
|
|
def get_response(req):
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=1000-2000,14000-15000'})
|
|
response_chunks = []
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
with self.assertRaises(ChunkReadTimeout):
|
|
# note: the error is raised while the resp_iter is read
|
|
for chunk in resp.app_iter:
|
|
response_chunks.append(chunk)
|
|
boundary = resp.headers['Content-Type'].split('=', 1)[1]
|
|
self.assertEqual(response_chunks, [
|
|
b'\r\n'.join([
|
|
b'--' + boundary.encode('ascii'),
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 1000-2000/16384',
|
|
b'',
|
|
b'',
|
|
]),
|
|
test_body[0:1001],
|
|
b'\r\n',
|
|
])
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(len(log), self.policy.ec_n_unique_fragments * 2)
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(2, len(log_lines), log_lines)
|
|
self.assertIn('Trying to read next part of EC fragment multi-part GET',
|
|
log_lines[0])
|
|
self.assertIn('Trying to read during GET: ChunkReadTimeout',
|
|
log_lines[1])
|
|
|
|
def test_GET_with_multirange_short_resume_body(self):
|
|
self.app.object_chunk_size = 256
|
|
self.app.recoverable_node_timeout = 0.01
|
|
test_body = b'test' * self.policy.ec_segment_size
|
|
ec_stub = make_ec_object_stub(test_body, self.policy, None)
|
|
frag_archives = ec_stub['frags']
|
|
self.assertEqual(len(frag_archives[0]), 1960)
|
|
boundary = b'81eb9c110b32ced5fe'
|
|
|
|
def make_mime_body(frag_archive):
|
|
return b'\r\n'.join([
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 0-489/1960',
|
|
b'',
|
|
frag_archive[0:490],
|
|
b'--' + boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 1470-1959/1960',
|
|
b'',
|
|
frag_archive[1470:],
|
|
b'--' + boundary + b'--',
|
|
])
|
|
|
|
obj_resp_bodies = [make_mime_body(fa) for fa
|
|
# no extra good responses
|
|
in ec_stub['frags'][:self.policy.ec_ndata]]
|
|
|
|
headers = {
|
|
'Content-Type': b'multipart/byteranges;boundary=' + boundary,
|
|
'Content-Length': len(obj_resp_bodies[0]),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(ec_stub['body']),
|
|
'X-Object-Sysmeta-Ec-Etag': ec_stub['etag'],
|
|
'X-Timestamp': Timestamp(self.ts()).normal,
|
|
}
|
|
|
|
responses = [
|
|
StubResponse(206, body, headers, i,
|
|
# make the first one slow
|
|
slowdown=0.1 if i == 0 else None)
|
|
for i, body in enumerate(obj_resp_bodies)
|
|
]
|
|
# add a short read response for the resume
|
|
short_body = obj_resp_bodies[0][:512]
|
|
responses.append(StubResponse(206, short_body, headers, 0))
|
|
|
|
def get_response(req):
|
|
# there's some math going on here I don't quite understand, the
|
|
# fragment_size is 490 and there's like 4 of them because ec_body
|
|
# is 'test' * segment_size
|
|
self.assertEqual(req['headers']['Range'], 'bytes=0-489,1470-1959')
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=1000-2000,14000-15000'})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
resp_boundary = resp.headers['content-type'].rsplit(
|
|
'=', 1)[1].encode()
|
|
expected = b'\r\n'.join([
|
|
b'--' + resp_boundary,
|
|
b'Content-Type: application/octet-stream',
|
|
b'Content-Range: bytes 1000-2000/16384',
|
|
b'',
|
|
b'',
|
|
b'--' + resp_boundary + b'--',
|
|
])
|
|
self.assertEqual(expected, resp.body)
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(len(log), self.policy.ec_n_unique_fragments * 2)
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertIn("Trying to read next part of EC fragment multi-part "
|
|
"GET (retrying)", log_lines[0])
|
|
# not the most graceful ending
|
|
self.assertIn("Exception fetching fragments for '/a/c/o'",
|
|
log_lines[-1])
|
|
|
|
def test_GET_with_success_and_507_will_503(self):
|
|
responses = [ # only 9 good nodes
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
]
|
|
|
|
def get_response(req):
|
|
# bad disk on all other nodes
|
|
return responses.pop(0) if responses else StubResponse(507)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 503)
|
|
self.assertEqual(len(log), 2 * self.replicas())
|
|
|
|
def test_GET_with_success_and_404_will_404(self):
|
|
responses = [ # only 9 good nodes
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
StubResponse(200),
|
|
]
|
|
|
|
def get_response(req):
|
|
# no frags on other nodes
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 404)
|
|
self.assertEqual(len(log), 2 * self.replicas())
|
|
|
|
def test_ranged_get(self):
|
|
self.app.object_chunk_size = 256
|
|
segment_size = self.policy.ec_segment_size
|
|
frag_size = self.policy.fragment_size
|
|
data = (b'test' * segment_size)[:-492]
|
|
etag = md5(data).hexdigest()
|
|
archives = self._make_ec_archive_bodies(data)
|
|
frag_archive_size = len(archives[0])
|
|
range_size = frag_size * 2
|
|
headers = {
|
|
'Content-Type': 'text/plain',
|
|
'Content-Length': range_size,
|
|
'Content-Range': 'bytes 0-%s/%s' % (range_size - 1,
|
|
frag_archive_size),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(data),
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Backend-Timestamp': Timestamp(self.ts()).internal
|
|
}
|
|
responses = [
|
|
StubResponse(206, body[:range_size], headers, i)
|
|
for i, body in enumerate(archives[:self.policy.ec_ndata])
|
|
]
|
|
|
|
obj_req_ranges = set()
|
|
|
|
def get_response(req):
|
|
obj_req_ranges.add(req['headers']['Range'])
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=3000-5000'})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(obj_req_ranges, {'bytes=0-%s' % (range_size - 1)})
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(resp.headers['Content-Range'],
|
|
'bytes 3000-5000/%s' % len(data))
|
|
self.assertEqual(resp.body, data[3000:5001])
|
|
self.assertEqual(len(log), self.policy.ec_ndata)
|
|
|
|
def test_ranged_get_with_slow_resp(self):
|
|
self.app.object_chunk_size = 256
|
|
self.app.recoverable_node_timeout = 0.01
|
|
segment_size = self.policy.ec_segment_size
|
|
frag_size = self.policy.fragment_size
|
|
data = (b'test' * segment_size)[:-492]
|
|
etag = md5(data).hexdigest()
|
|
archives = self._make_ec_archive_bodies(data)
|
|
frag_archive_size = len(archives[0])
|
|
range_size = frag_size * 2
|
|
headers = {
|
|
'Content-Type': 'text/plain',
|
|
'Content-Length': range_size,
|
|
'Content-Range': 'bytes 0-%s/%s' % (range_size - 1,
|
|
frag_archive_size),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(data),
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Backend-Timestamp': Timestamp(self.ts()).internal
|
|
}
|
|
responses = [
|
|
StubResponse(206, body[:range_size], headers, i,
|
|
# the first body comes up slow
|
|
slowdown=0.1 if i == 0 else None)
|
|
for i, body in enumerate(archives[:self.policy.ec_ndata])
|
|
]
|
|
responses.append(StubResponse(
|
|
206, archives[self.policy.ec_ndata][:range_size],
|
|
headers, self.policy.ec_ndata))
|
|
|
|
obj_req_ranges = set()
|
|
|
|
def get_response(req):
|
|
obj_req_ranges.add(req['headers']['Range'])
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=3000-5000'})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.body, data[3000:5001])
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(obj_req_ranges, {'bytes=0-%s' % (range_size - 1)})
|
|
self.assertEqual(resp.headers['Content-Range'],
|
|
'bytes 3000-5000/%s' % len(data))
|
|
self.assertEqual(len(log), self.policy.ec_ndata + 1)
|
|
|
|
def test_ranged_get_with_short_resp(self):
|
|
self.app.object_chunk_size = 256
|
|
segment_size = self.policy.ec_segment_size
|
|
frag_size = self.policy.fragment_size
|
|
data = (b'test' * segment_size)[:-492]
|
|
etag = md5(data).hexdigest()
|
|
archives = self._make_ec_archive_bodies(data)
|
|
frag_archive_size = len(archives[0])
|
|
range_size = frag_size * 2
|
|
headers = {
|
|
'Content-Type': 'text/plain',
|
|
'Content-Length': range_size,
|
|
'Content-Range': 'bytes 0-%s/%s' % (range_size - 1,
|
|
frag_archive_size),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(data),
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Backend-Timestamp': Timestamp(self.ts()).internal
|
|
}
|
|
responses = [
|
|
StubResponse(
|
|
206,
|
|
# the first body comes up short
|
|
body[:frag_size] if i == 0 else body[:range_size],
|
|
headers, i)
|
|
for i, body in enumerate(archives[:self.policy.ec_ndata])
|
|
]
|
|
responses.append(StubResponse(
|
|
206, archives[self.policy.ec_ndata][frag_size:range_size], {
|
|
'Content-Type': 'text/plain',
|
|
'Content-Length': frag_size,
|
|
'Content-Range': 'bytes %s-%s/%s' % (
|
|
frag_size, range_size - 1, frag_archive_size),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(data),
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Backend-Timestamp': Timestamp(self.ts()).internal,
|
|
}, self.policy.ec_ndata))
|
|
|
|
obj_req_ranges = []
|
|
|
|
def get_response(req):
|
|
obj_req_ranges.append(req['headers']['Range'])
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=3000-5000'})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.body, data[3000:5001])
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(obj_req_ranges,
|
|
['bytes=0-%s' % (range_size - 1)] *
|
|
self.policy.ec_ndata +
|
|
['bytes=%s-%s' % (frag_size, range_size - 1)])
|
|
self.assertEqual(resp.headers['Content-Range'],
|
|
'bytes 3000-5000/%s' % len(data))
|
|
self.assertEqual(len(log), self.policy.ec_ndata + 1)
|
|
|
|
def test_ranged_get_with_short_resp_timeout(self):
|
|
self.app.object_chunk_size = 256
|
|
self.app.recoverable_node_timeout = 0.01
|
|
segment_size = self.policy.ec_segment_size
|
|
frag_size = self.policy.fragment_size
|
|
data = (b'test' * segment_size)[:-492]
|
|
etag = md5(data).hexdigest()
|
|
archives = self._make_ec_archive_bodies(data)
|
|
frag_archive_size = len(archives[0])
|
|
range_size = frag_size * 2
|
|
headers = {
|
|
'Content-Type': 'text/plain',
|
|
'Content-Length': range_size,
|
|
'Content-Range': 'bytes 0-%s/%s' % (range_size - 1,
|
|
frag_archive_size),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(data),
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Backend-Timestamp': Timestamp(self.ts()).internal
|
|
}
|
|
responses = [
|
|
StubResponse(
|
|
206, body[:range_size], headers, i,
|
|
# the first body slows down after awhile
|
|
slowdown=[None] * 3 + [0.1] if i == 0 else None)
|
|
for i, body in enumerate(archives[:self.policy.ec_ndata])
|
|
]
|
|
responses.append(StubResponse(
|
|
206, archives[self.policy.ec_ndata][frag_size:range_size], {
|
|
'Content-Type': 'text/plain',
|
|
'Content-Length': frag_size,
|
|
'Content-Range': 'bytes %s-%s/%s' % (
|
|
frag_size, range_size - 1, frag_archive_size),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(data),
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Backend-Timestamp': Timestamp(self.ts()).internal,
|
|
}, self.policy.ec_ndata))
|
|
|
|
obj_req_ranges = []
|
|
|
|
def get_response(req):
|
|
obj_req_ranges.append(req['headers']['Range'])
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o', headers={
|
|
'Range': 'bytes=3000-5000'})
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.body, data[3000:5001])
|
|
self.assertEqual(resp.status_int, 206)
|
|
self.assertEqual(['bytes=0-%s' % (range_size - 1)] *
|
|
self.policy.ec_ndata +
|
|
['bytes=%s-%s' % (frag_size, range_size - 1)],
|
|
obj_req_ranges)
|
|
self.assertEqual(resp.headers['Content-Range'],
|
|
'bytes 3000-5000/%s' % len(data))
|
|
self.assertEqual(len(log), self.policy.ec_ndata + 1)
|
|
|
|
def test_GET_mixed_ranged_responses_success(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
frag_size = self.policy.fragment_size
|
|
new_data = (b'test' * segment_size)[:-492]
|
|
new_etag = md5(new_data, usedforsecurity=False).hexdigest()
|
|
new_archives = self._make_ec_archive_bodies(new_data)
|
|
old_data = (b'junk' * segment_size)[:-492]
|
|
old_etag = md5(old_data, usedforsecurity=False).hexdigest()
|
|
old_archives = self._make_ec_archive_bodies(old_data)
|
|
frag_archive_size = len(new_archives[0])
|
|
|
|
# here we deliberately omit X-Backend-Data-Timestamp to check that
|
|
# proxy will tolerate responses from object server that have not been
|
|
# upgraded to send that header
|
|
old_headers = {
|
|
'Content-Type': 'text/plain',
|
|
'Content-Length': frag_size,
|
|
'Content-Range': 'bytes 0-%s/%s' % (frag_size - 1,
|
|
frag_archive_size),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(old_data),
|
|
'X-Object-Sysmeta-Ec-Etag': old_etag,
|
|
'X-Backend-Timestamp': Timestamp(self.ts()).internal
|
|
}
|
|
new_headers = {
|
|
'Content-Type': 'text/plain',
|
|
'Content-Length': frag_size,
|
|
'Content-Range': 'bytes 0-%s/%s' % (frag_size - 1,
|
|
frag_archive_size),
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(new_data),
|
|
'X-Object-Sysmeta-Ec-Etag': new_etag,
|
|
'X-Backend-Timestamp': Timestamp(self.ts()).internal
|
|
}
|
|
# 7 primaries with stale frags, 3 handoffs failed to get new frags
|
|
responses = [
|
|
StubResponse(206, old_archives[0][:frag_size], old_headers, 0),
|
|
StubResponse(206, new_archives[1][:frag_size], new_headers, 1),
|
|
StubResponse(206, old_archives[2][:frag_size], old_headers, 2),
|
|
StubResponse(206, new_archives[3][:frag_size], new_headers, 3),
|
|
StubResponse(206, old_archives[4][:frag_size], old_headers, 4),
|
|
StubResponse(206, new_archives[5][:frag_size], new_headers, 5),
|
|
StubResponse(206, old_archives[6][:frag_size], old_headers, 6),
|
|
StubResponse(206, new_archives[7][:frag_size], new_headers, 7),
|
|
StubResponse(206, old_archives[8][:frag_size], old_headers, 8),
|
|
StubResponse(206, new_archives[9][:frag_size], new_headers, 9),
|
|
StubResponse(206, old_archives[10][:frag_size], old_headers, 10),
|
|
StubResponse(206, new_archives[11][:frag_size], new_headers, 11),
|
|
StubResponse(206, old_archives[12][:frag_size], old_headers, 12),
|
|
StubResponse(206, new_archives[13][:frag_size], new_headers, 13),
|
|
StubResponse(206, new_archives[0][:frag_size], new_headers, 0),
|
|
StubResponse(404),
|
|
StubResponse(404),
|
|
StubResponse(206, new_archives[6][:frag_size], new_headers, 6),
|
|
StubResponse(404),
|
|
StubResponse(206, new_archives[10][:frag_size], new_headers, 10),
|
|
StubResponse(206, new_archives[12][:frag_size], new_headers, 12),
|
|
]
|
|
|
|
def get_response(req):
|
|
return responses.pop(0) if responses else StubResponse(404)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(get_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
closed_conn = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers.get('X-Object-Sysmeta-Ec-Etag')
|
|
closed_conn[etag].add(conn.closed)
|
|
self.assertEqual({
|
|
old_etag: {True},
|
|
new_etag: {False},
|
|
None: {True},
|
|
}, dict(closed_conn))
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.body, new_data[:segment_size])
|
|
self.assertEqual(len(log), self.policy.ec_ndata + 10)
|
|
|
|
def test_GET_mismatched_fragment_archives(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data1 = (b'test' * segment_size)[:-333]
|
|
# N.B. the object data *length* here is different
|
|
test_data2 = (b'blah1' * segment_size)[:-333]
|
|
|
|
etag1 = md5(test_data1, usedforsecurity=False).hexdigest()
|
|
etag2 = md5(test_data2, usedforsecurity=False).hexdigest()
|
|
|
|
ec_archive_bodies1 = self._make_ec_archive_bodies(test_data1)
|
|
ec_archive_bodies2 = self._make_ec_archive_bodies(test_data2)
|
|
|
|
headers1 = {'X-Object-Sysmeta-Ec-Etag': etag1,
|
|
'X-Object-Sysmeta-Ec-Content-Length': '333'}
|
|
# here we're going to *lie* and say the etag here matches
|
|
headers2 = {'X-Object-Sysmeta-Ec-Etag': etag1,
|
|
'X-Object-Sysmeta-Ec-Content-Length': '333'}
|
|
|
|
responses1 = [(200, body, self._add_frag_index(fi, headers1))
|
|
for fi, body in enumerate(ec_archive_bodies1)]
|
|
responses2 = [(200, body, self._add_frag_index(fi, headers2))
|
|
for fi, body in enumerate(ec_archive_bodies2)]
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
|
|
orig_decode = self.policy.pyeclib_driver.decode
|
|
captured_fragments = []
|
|
|
|
def mock_decode(fragments):
|
|
captured_fragments.append(fragments)
|
|
return orig_decode(fragments)
|
|
|
|
# sanity check responses1 and capture frag lengths
|
|
responses = responses1[:self.policy.ec_ndata]
|
|
status_codes, body_iter, headers = zip(*responses)
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
with mock.patch.object(
|
|
self.policy.pyeclib_driver, 'decode', mock_decode):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
# read body while decode is mocked
|
|
self.assertEqual(
|
|
md5(resp.body, usedforsecurity=False).hexdigest(),
|
|
etag1)
|
|
fragment_lengths1 = [[len(frag) for frag in frags]
|
|
for frags in captured_fragments]
|
|
self.assertEqual( # sanity check
|
|
len(ec_archive_bodies1[0]),
|
|
sum([length for length in [lengths[0]
|
|
for lengths in fragment_lengths1]]))
|
|
|
|
# sanity check responses2 and capture frag lengths
|
|
captured_fragments = []
|
|
responses = responses2[:self.policy.ec_ndata]
|
|
status_codes, body_iter, headers = zip(*responses)
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
with mock.patch.object(
|
|
self.policy.pyeclib_driver, 'decode', mock_decode):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
# read body while decode is mocked
|
|
self.assertEqual(
|
|
md5(resp.body, usedforsecurity=False).hexdigest(),
|
|
etag2)
|
|
fragment_lengths2 = [[len(frag) for frag in frags]
|
|
for frags in captured_fragments]
|
|
self.assertEqual( # sanity check
|
|
len(ec_archive_bodies2[0]),
|
|
sum([length for length in [lengths[0]
|
|
for lengths in fragment_lengths2]]))
|
|
|
|
# now mix the responses a bit
|
|
mix_index = random.randint(0, self.policy.ec_ndata - 1)
|
|
mixed_responses = responses1[:self.policy.ec_ndata]
|
|
mixed_responses[mix_index] = responses2[mix_index]
|
|
|
|
num_segments = len(fragment_lengths1)
|
|
mixed_lengths = fragment_lengths1[num_segments - 1]
|
|
mixed_lengths[mix_index] = fragment_lengths2[
|
|
num_segments - 1][mix_index]
|
|
|
|
status_codes, body_iter, headers = zip(*mixed_responses)
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
try:
|
|
resp.body
|
|
except ECDriverError:
|
|
resp._app_iter.close()
|
|
else:
|
|
self.fail('invalid ec fragment response body did not blow up!')
|
|
error_lines = self.logger.get_lines_for_level('error')
|
|
self.assertEqual(1, len(error_lines))
|
|
msg = error_lines[0]
|
|
self.assertIn('Error decoding fragments', msg)
|
|
self.assertIn('/a/c/o', msg)
|
|
self.assertIn('Segments decoded: %d' % (num_segments - 1), msg)
|
|
self.assertIn(
|
|
"[%s]" % ", ".join([str(length) for length in mixed_lengths]), msg)
|
|
self.assertIn("Invalid fragment payload in ECPyECLibDriver.decode",
|
|
msg)
|
|
|
|
def test_GET_read_timeout(self):
|
|
# verify EC GET behavior when initial batch of nodes time out then
|
|
# remaining primary nodes also time out and handoffs return 404
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-333]
|
|
etag = md5(test_data, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
headers = {'X-Object-Sysmeta-Ec-Etag': etag}
|
|
self.app.recoverable_node_timeout = 0.01
|
|
responses = [
|
|
(200, SlowBody(body, 0.1), self._add_frag_index(i, headers))
|
|
for i, body in enumerate(ec_archive_bodies)
|
|
] * self.policy.ec_duplication_factor
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
|
|
status_codes, body_iter, headers = zip(*responses + [
|
|
(404, [b''], {}) for i in range(
|
|
self.policy.object_ring.max_more_nodes)])
|
|
with mocked_http_conn(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
# do this inside the fake http context manager, it'll try to
|
|
# resume but won't be able to give us all the right bytes
|
|
self.assertNotEqual(
|
|
md5(resp.body, usedforsecurity=False).hexdigest(),
|
|
etag)
|
|
error_lines = self.logger.get_lines_for_level('error')
|
|
# all primaries timeout and get error limited
|
|
error_limit_lines = [
|
|
line for line in error_lines
|
|
if 'Trying to read EC fragment during GET (retrying)' in line]
|
|
self.assertEqual(self.policy.ec_n_unique_fragments,
|
|
len(error_limit_lines))
|
|
# all ec_ndata frag getters eventually get a read timeout
|
|
read_timeout_lines = [
|
|
line for line in error_lines if 'ChunkReadTimeout (0.01s' in line]
|
|
self.assertEqual(self.policy.ec_ndata,
|
|
len(read_timeout_lines))
|
|
for line in self.logger.logger.records['ERROR']:
|
|
self.assertIn(req.headers['x-trans-id'], line)
|
|
|
|
def test_GET_write_timeout(self):
|
|
# verify EC GET behavior when there's a timeout sending decoded frags
|
|
# via the queue.
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-333]
|
|
etag = md5(test_data, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
headers = {'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': '333'}
|
|
ndata = self.policy.ec_ndata
|
|
responses = [
|
|
(200, body, self._add_frag_index(i, headers))
|
|
for i, body in enumerate(ec_archive_bodies[:ndata])
|
|
] * self.policy.ec_duplication_factor
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
|
|
status_codes, body_iter, headers = zip(*responses)
|
|
self.app.client_timeout = 0.01
|
|
with mocked_http_conn(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
resp_body = next(resp.app_iter)
|
|
sleep(0.5) # lazy client
|
|
# remaining resp truncated
|
|
resp_body += b''.join(resp.app_iter)
|
|
# we log errors
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertTrue(log_lines)
|
|
for line in log_lines:
|
|
self.assertIn('ChunkWriteTimeout feeding fragments', line)
|
|
# client gets a short read
|
|
self.assertEqual(16051, len(test_data))
|
|
self.assertEqual(8192, len(resp_body))
|
|
self.assertNotEqual(
|
|
md5(resp_body, usedforsecurity=False).hexdigest(),
|
|
etag)
|
|
|
|
def test_GET_read_timeout_retrying_but_no_more_useful_nodes(self):
|
|
# verify EC GET behavior when initial batch of nodes time out then
|
|
# remaining nodes either return 404 or return data for different etag
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-333]
|
|
etag = md5(test_data).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
headers = {'X-Object-Sysmeta-Ec-Etag': etag}
|
|
self.app.recoverable_node_timeout = 0.01
|
|
ndata = self.policy.ec_ndata
|
|
# only ndata responses, all of which have SlowBody
|
|
responses = [
|
|
(200, SlowBody(body, 0.1), self._add_frag_index(i, headers))
|
|
for i, body in enumerate(ec_archive_bodies[:ndata])
|
|
] * self.policy.ec_duplication_factor
|
|
# 2 primaries return 404
|
|
responses += [
|
|
(404, '', {}), (404, '', {})
|
|
] * self.policy.ec_duplication_factor
|
|
# 2 primaries return different etag
|
|
headers2 = {'X-Object-Sysmeta-Ec-Etag': 'other_etag'}
|
|
responses += [
|
|
(200, body, self._add_frag_index(i, headers2))
|
|
for i, body in enumerate(ec_archive_bodies[ndata + 2:])
|
|
] * self.policy.ec_duplication_factor
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
|
|
# all other (handoff) responses are 404
|
|
status_codes, body_iter, headers = zip(*responses + [
|
|
(404, [b''], {}) for i in range(
|
|
self.policy.object_ring.max_more_nodes)])
|
|
with mocked_http_conn(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
# do this inside the fake http context manager, it'll try to
|
|
# resume but won't be able to give us all the right bytes
|
|
self.assertNotEqual(md5(resp.body).hexdigest(), etag)
|
|
error_lines = self.logger.get_lines_for_level('error')
|
|
# only ec_ndata primaries that timeout get error limited (404 or
|
|
# different etag primaries do not get error limited)
|
|
error_limit_lines = [
|
|
line for line in error_lines
|
|
if 'Trying to read EC fragment during GET (retrying)' in line]
|
|
self.assertEqual(self.policy.ec_ndata, len(error_limit_lines))
|
|
# all ec_ndata frag getters eventually get a read timeout
|
|
read_timeout_lines = [
|
|
line for line in error_lines if 'ChunkReadTimeout (0.01s' in line]
|
|
self.assertEqual(self.policy.ec_ndata,
|
|
len(read_timeout_lines))
|
|
|
|
debug_lines = self.logger.get_lines_for_level('debug')
|
|
nparity = self.policy.ec_nparity
|
|
nhandoffs = self.policy.object_ring.max_more_nodes
|
|
ignore_404 = ignore_404_handoff = 0
|
|
for line in debug_lines:
|
|
if 'Ignoring 404 from primary' in line:
|
|
ignore_404 += 1
|
|
if 'Ignoring 404 from handoff' in line:
|
|
ignore_404_handoff += 1
|
|
self.assertEqual(nparity - 2, ignore_404, debug_lines)
|
|
self.assertEqual(nhandoffs, ignore_404_handoff, debug_lines)
|
|
self.assertEqual(len(debug_lines), ignore_404_handoff + ignore_404)
|
|
self.assertEqual(self.logger.get_lines_for_level('warning'), [
|
|
'Skipping source (etag mismatch: got other_etag, '
|
|
'expected %s)' % etag] * 2)
|
|
|
|
def test_GET_read_timeout_resume(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-333]
|
|
etag = md5(test_data, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
headers = {
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(test_data),
|
|
}
|
|
self.app.recoverable_node_timeout = 0.05
|
|
# first one is slow
|
|
responses = [(200, SlowBody(ec_archive_bodies[0], 0.1),
|
|
self._add_frag_index(0, headers))]
|
|
# ... the rest are fine
|
|
responses += [(200, body, self._add_frag_index(i, headers))
|
|
for i, body in enumerate(ec_archive_bodies[1:], start=1)]
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
|
|
status_codes, body_iter, headers = zip(
|
|
*responses[:self.policy.ec_ndata + 1])
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(
|
|
md5(resp.body, usedforsecurity=False).hexdigest(),
|
|
etag)
|
|
error_lines = self.logger.get_lines_for_level('error')
|
|
self.assertEqual(1, len(error_lines))
|
|
self.assertIn('retrying', error_lines[0])
|
|
for line in self.logger.logger.records['ERROR']:
|
|
self.assertIn(req.headers['x-trans-id'], line)
|
|
|
|
def _do_test_GET_read_timeout_fast_forward_fails(self, error):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-333]
|
|
etag = md5(test_data).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
headers = {
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(test_data),
|
|
}
|
|
self.app.recoverable_node_timeout = 0.05
|
|
# first one is slow
|
|
responses = [(200, SlowBody(ec_archive_bodies[0], 0.1),
|
|
self._add_frag_index(0, headers))]
|
|
# ... the rest are fine
|
|
responses += [(200, body, self._add_frag_index(i, headers))
|
|
for i, body in enumerate(ec_archive_bodies[1:], start=1)]
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
|
|
status_codes, body_iter, headers = zip(
|
|
*responses[:self.policy.ec_ndata])
|
|
# I don't know why fast_forward would blow up, but if it does we
|
|
# re-raise the ChunkReadTimeout and still want a txn-id
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers), \
|
|
mock.patch(
|
|
'swift.proxy.controllers.obj.ECFragGetter.fast_forward',
|
|
side_effect=error):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertNotEqual(md5(resp.body).hexdigest(), etag)
|
|
|
|
for line in self.logger.logger.records['ERROR'] + \
|
|
self.logger.logger.records['WARNING']:
|
|
self.assertIn(req.headers['x-trans-id'], line)
|
|
|
|
def test_GET_read_timeout_fast_forward_fails(self):
|
|
self._do_test_GET_read_timeout_fast_forward_fails(ValueError())
|
|
|
|
error_lines = self.logger.get_lines_for_level('error')
|
|
self.assertEqual(2, len(error_lines))
|
|
self.assertIn('Unable to fast forward', error_lines[0])
|
|
self.assertIn('Timeout fetching', error_lines[1])
|
|
warning_lines = self.logger.get_lines_for_level('warning')
|
|
self.assertEqual(1, len(warning_lines))
|
|
self.assertIn(
|
|
'Un-recoverable fragment rebuild. Only received 9/10 fragments',
|
|
warning_lines[0])
|
|
|
|
def test_GET_read_timeout_fast_forward_range_complete(self):
|
|
self._do_test_GET_read_timeout_fast_forward_fails(
|
|
RangeAlreadyComplete())
|
|
|
|
error_lines = self.logger.get_lines_for_level('error')
|
|
self.assertEqual(0, len(error_lines))
|
|
# the test is a little bogus - presumably if the range was complete
|
|
# then the fragment would be ok to rebuild. But the test pretends range
|
|
# was complete without actually feeding the bytes to the getter...
|
|
warning_lines = self.logger.get_lines_for_level('warning')
|
|
self.assertEqual(1, len(warning_lines))
|
|
self.assertIn(
|
|
'Un-recoverable fragment rebuild. Only received 9/10 fragments',
|
|
warning_lines[0])
|
|
|
|
def test_GET_one_short_fragment_archive(self):
|
|
# verify that a warning is logged when one fragment archive returns
|
|
# less whole fragments than others
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-333]
|
|
etag = md5(test_data).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
|
|
def do_test(missing_length):
|
|
self.logger.clear()
|
|
headers = {
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(test_data),
|
|
}
|
|
responses = [(200, ec_archive_bodies[0][:(-1 * missing_length)],
|
|
self._add_frag_index(0, headers))]
|
|
# ... the rest are fine
|
|
responses += [
|
|
(200, body, self._add_frag_index(i, headers))
|
|
for i, body in enumerate(ec_archive_bodies[1:], start=1)]
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
|
|
status_codes, body_iter, headers = zip(
|
|
*responses[:self.policy.ec_ndata])
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertNotEqual(md5(resp.body).hexdigest(), etag)
|
|
error_lines = self.logger.get_lines_for_level('error')
|
|
self.assertEqual([], error_lines)
|
|
warning_lines = self.logger.get_lines_for_level('warning')
|
|
self.assertEqual(1, len(warning_lines))
|
|
self.assertIn(
|
|
'Un-recoverable fragment rebuild. '
|
|
'Only received 9/10 fragments', warning_lines[0])
|
|
|
|
# each fragment archive has 4 fragments of sizes [490, 490, 490, 458];
|
|
# try dropping whole fragment(s) from one archive
|
|
do_test(458)
|
|
do_test(490 + 458)
|
|
do_test(490 + 490 + 458)
|
|
do_test(490 + 490 + 490 + 458)
|
|
|
|
def test_GET_trigger_ec_metadata_check_failure(self):
|
|
# verify that a warning is logged when there are only k - 1 fragment
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data = (b'test' * segment_size)[:-333]
|
|
etag = md5(test_data).hexdigest()
|
|
ec_archive_bodies = self._make_ec_archive_bodies(test_data)
|
|
bad_bodies = [b'd' * segment_size] * (self.policy.ec_nparity + 1)
|
|
ec_archive_bodies = \
|
|
ec_archive_bodies[:self.policy.ec_ndata - 1] + bad_bodies
|
|
|
|
self.logger.clear()
|
|
headers = {
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(test_data),
|
|
}
|
|
responses = [
|
|
(200, body, self._add_frag_index(i, headers))
|
|
for i, body in enumerate(ec_archive_bodies)]
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
|
|
status_codes, body_iter, headers = zip(
|
|
*responses[:self.policy.ec_ndata])
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 500)
|
|
self.assertNotEqual(md5(resp.body).hexdigest(), etag)
|
|
error_lines = self.logger.get_lines_for_level('error')
|
|
expected_log_line = (
|
|
"Error decoding fragments for '/a/c/o'. "
|
|
"Segments decoded: 0, Lengths: "
|
|
"[490, 490, 490, 490, 490, 490, 490, 490, 490, 490]: "
|
|
"pyeclib_c_decode ERROR: Fragment integrity check failed. "
|
|
"Please inspect syslog for liberasurecode error report.")
|
|
self.assertEqual(expected_log_line, error_lines[0])
|
|
|
|
def test_GET_read_timeout_resume_mixed_etag(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_data2 = (b'blah1' * segment_size)[:-333]
|
|
test_data1 = (b'test' * segment_size)[:-333]
|
|
etag2 = md5(test_data2, usedforsecurity=False).hexdigest()
|
|
etag1 = md5(test_data1, usedforsecurity=False).hexdigest()
|
|
ec_archive_bodies2 = self._make_ec_archive_bodies(test_data2)
|
|
ec_archive_bodies1 = self._make_ec_archive_bodies(test_data1)
|
|
headers2 = {'X-Object-Sysmeta-Ec-Etag': etag2,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(test_data2),
|
|
'X-Backend-Timestamp': self.ts().internal}
|
|
headers1 = {'X-Object-Sysmeta-Ec-Etag': etag1,
|
|
'X-Object-Sysmeta-Ec-Content-Length': len(test_data1),
|
|
'X-Backend-Timestamp': self.ts().internal}
|
|
responses = [
|
|
# 404
|
|
(404, [b''], {}),
|
|
# etag1
|
|
(200, ec_archive_bodies1[1], self._add_frag_index(1, headers1)),
|
|
# 404
|
|
(404, [b''], {}),
|
|
# etag1
|
|
(200, SlowBody(ec_archive_bodies1[3], 0.1), self._add_frag_index(
|
|
3, headers1)),
|
|
# etag2
|
|
(200, ec_archive_bodies2[4], self._add_frag_index(4, headers2)),
|
|
# etag1
|
|
(200, ec_archive_bodies1[5], self._add_frag_index(5, headers1)),
|
|
# etag2
|
|
(200, ec_archive_bodies2[6], self._add_frag_index(6, headers2)),
|
|
# etag1
|
|
(200, ec_archive_bodies1[7], self._add_frag_index(7, headers1)),
|
|
# etag2
|
|
(200, ec_archive_bodies2[8], self._add_frag_index(8, headers2)),
|
|
# etag1
|
|
(200, SlowBody(ec_archive_bodies1[9], 0.1), self._add_frag_index(
|
|
9, headers1)),
|
|
# etag2
|
|
(200, ec_archive_bodies2[10], self._add_frag_index(10, headers2)),
|
|
# etag1
|
|
(200, ec_archive_bodies1[11], self._add_frag_index(11, headers1)),
|
|
# etag2
|
|
(200, ec_archive_bodies2[12], self._add_frag_index(12, headers2)),
|
|
# 404
|
|
(404, [b''], {}),
|
|
# handoffs start here
|
|
# etag2
|
|
(200, ec_archive_bodies2[0], self._add_frag_index(0, headers2)),
|
|
# 404
|
|
(404, [b''], {}),
|
|
# etag1
|
|
(200, ec_archive_bodies1[2], self._add_frag_index(2, headers1)),
|
|
# 404
|
|
(404, [b''], {}),
|
|
# etag1
|
|
(200, ec_archive_bodies1[4], self._add_frag_index(4, headers1)),
|
|
# etag2
|
|
(200, ec_archive_bodies2[1], self._add_frag_index(1, headers2)),
|
|
# etag1
|
|
(200, ec_archive_bodies1[6], self._add_frag_index(6, headers1)),
|
|
# etag2
|
|
(200, ec_archive_bodies2[7], self._add_frag_index(7, headers2)),
|
|
# etag1
|
|
(200, ec_archive_bodies1[8], self._add_frag_index(8, headers1)),
|
|
# resume requests start here
|
|
# 404
|
|
(404, [b''], {}),
|
|
# etag2
|
|
(200, ec_archive_bodies2[3], self._add_frag_index(3, headers2)),
|
|
# 404
|
|
(404, [b''], {}),
|
|
# etag1
|
|
(200, ec_archive_bodies1[10], self._add_frag_index(10, headers1)),
|
|
# etag1
|
|
(200, ec_archive_bodies1[12], self._add_frag_index(12, headers1)),
|
|
]
|
|
self.app.recoverable_node_timeout = 0.01
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
status_codes, body_iter, headers = zip(*responses)
|
|
with mocked_http_conn(*status_codes, body_iter=body_iter,
|
|
headers=headers) as log:
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(
|
|
md5(resp.body, usedforsecurity=False).hexdigest(),
|
|
etag1)
|
|
error_lines = self.logger.get_lines_for_level('error')
|
|
self.assertEqual(2, len(error_lines))
|
|
for line in error_lines:
|
|
self.assertIn('retrying', line)
|
|
for line in self.logger.logger.records['ERROR']:
|
|
self.assertIn(req.headers['x-trans-id'], line)
|
|
etag2_conns = []
|
|
for conn in log.responses:
|
|
if conn.headers.get('X-Object-Sysmeta-Ec-Etag') == etag2:
|
|
etag2_conns.append(conn)
|
|
self.assertEqual(
|
|
([True] * 8) + [False], # the resumed etag2 doesn't get closed
|
|
[conn.closed for conn in etag2_conns])
|
|
|
|
def test_fix_response_HEAD(self):
|
|
headers = {'X-Object-Sysmeta-Ec-Content-Length': '10',
|
|
'X-Object-Sysmeta-Ec-Etag': 'foo'}
|
|
|
|
# sucsessful HEAD
|
|
responses = [(200, b'', headers)]
|
|
status_codes, body_iter, headers = zip(*responses)
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='HEAD')
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.body, b'')
|
|
# 200OK shows original object content length
|
|
self.assertEqual(resp.headers['Content-Length'], '10')
|
|
self.assertEqual(resp.headers['Etag'], 'foo')
|
|
|
|
# not found HEAD
|
|
responses = [(404, b'', {})] * self.replicas() * 2
|
|
status_codes, body_iter, headers = zip(*responses)
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='HEAD')
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 404)
|
|
# 404 shows actual response body size (i.e. 0 for HEAD)
|
|
self.assertEqual(resp.headers['Content-Length'], '0')
|
|
|
|
def test_GET_with_invalid_ranges(self):
|
|
# real body size is segment_size - 10 (just 1 segment)
|
|
segment_size = self.policy.ec_segment_size
|
|
real_body = (b'a' * segment_size)[:-10]
|
|
|
|
# range is out of real body but in segment size
|
|
self._test_invalid_ranges('GET', real_body,
|
|
segment_size, '%s-' % (segment_size - 10))
|
|
# range is out of both real body and segment size
|
|
self._test_invalid_ranges('GET', real_body,
|
|
segment_size, '%s-' % (segment_size + 10))
|
|
|
|
def _test_invalid_ranges(self, method, real_body, segment_size, req_range):
|
|
# make a request with range starts from more than real size.
|
|
body_etag = md5(real_body, usedforsecurity=False).hexdigest()
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method=method,
|
|
headers={'Destination': 'c1/o',
|
|
'Range': 'bytes=%s' % (req_range)})
|
|
|
|
fragments = self.policy.pyeclib_driver.encode(real_body)
|
|
fragment_payloads = [fragments * self.policy.ec_duplication_factor]
|
|
|
|
node_fragments = list(zip(*fragment_payloads))
|
|
self.assertEqual(len(node_fragments), self.replicas()) # sanity
|
|
headers = {'X-Object-Sysmeta-Ec-Content-Length': str(len(real_body)),
|
|
'X-Object-Sysmeta-Ec-Etag': body_etag}
|
|
start = int(req_range.split('-')[0])
|
|
self.assertGreaterEqual(start, 0) # sanity
|
|
title, exp = swob.RESPONSE_REASONS[416]
|
|
range_not_satisfiable_body = \
|
|
'<html><h1>%s</h1><p>%s</p></html>' % (title, exp)
|
|
range_not_satisfiable_body = range_not_satisfiable_body.encode('utf-8')
|
|
if start >= segment_size:
|
|
responses = [(416, range_not_satisfiable_body,
|
|
self._add_frag_index(i, headers))
|
|
for i in range(POLICIES.default.ec_ndata)]
|
|
else:
|
|
responses = [(200, b''.join(node_fragments[i]),
|
|
self._add_frag_index(i, headers))
|
|
for i in range(POLICIES.default.ec_ndata)]
|
|
status_codes, body_iter, headers = zip(*responses)
|
|
expect_headers = {
|
|
'X-Obj-Metadata-Footer': 'yes',
|
|
'X-Obj-Multiphase-Commit': 'yes'
|
|
}
|
|
with set_http_connect(*status_codes, body_iter=body_iter,
|
|
headers=headers, expect_headers=expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 416)
|
|
self.assertEqual(resp.content_length, len(range_not_satisfiable_body))
|
|
self.assertEqual(resp.body, range_not_satisfiable_body)
|
|
self.assertEqual(resp.etag, body_etag)
|
|
self.assertEqual(resp.headers['Accept-Ranges'], 'bytes')
|
|
|
|
def test_non_durable_ec_response_bucket(self):
|
|
ts = self.ts()
|
|
bucket = obj.ECGetResponseBucket(self.policy, ts)
|
|
self.assertEqual(bucket.shortfall, self.policy.ec_ndata)
|
|
for i in range(1, self.policy.ec_ndata - self.policy.ec_nparity + 1):
|
|
stub_getter = mock.MagicMock(last_status=200, last_headers={
|
|
'X-Backend-Timestamp': ts.internal,
|
|
'X-Object-Sysmeta-Ec-Etag': 'the-etag',
|
|
'X-Object-Sysmeta-Ec-Frag-Index': str(i),
|
|
})
|
|
bucket.add_response(stub_getter, None)
|
|
self.assertEqual(bucket.shortfall, self.policy.ec_ndata - i)
|
|
self.assertEqual(bucket.shortfall, self.policy.ec_nparity)
|
|
self.assertFalse(bucket.durable)
|
|
expectations = (
|
|
4, # 7
|
|
4, # 8
|
|
4, # 9
|
|
4, # 10
|
|
3, # 11
|
|
2, # 12
|
|
1, # 13
|
|
1, # 14
|
|
)
|
|
for i, expected in zip(range(
|
|
self.policy.ec_ndata - self.policy.ec_nparity + 1,
|
|
self.policy.object_ring.replica_count + 1), expectations):
|
|
stub_getter = mock.MagicMock(last_status=200, last_headers={
|
|
'X-Backend-Timestamp': ts.internal,
|
|
'X-Object-Sysmeta-Ec-Etag': 'the-etag',
|
|
'X-Object-Sysmeta-Ec-Frag-Index': str(i),
|
|
})
|
|
bucket.add_response(stub_getter, None)
|
|
msg = 'With %r resp, expected shortfall %s != %s' % (
|
|
bucket.gets.keys(), expected, bucket.shortfall)
|
|
self.assertEqual(bucket.shortfall, expected, msg)
|
|
|
|
|
|
class TestECFunctions(unittest.TestCase):
|
|
def test_chunk_transformer(self):
|
|
def do_test(dup_factor, segments):
|
|
segment_size = 1024
|
|
orig_chunks = []
|
|
for i in range(segments):
|
|
orig_chunks.append(
|
|
chr(i + 97).encode('latin-1') * segment_size)
|
|
policy = ECStoragePolicy(0, 'ec8-2', ec_type=DEFAULT_TEST_EC_TYPE,
|
|
ec_ndata=8, ec_nparity=2,
|
|
object_ring=FakeRing(
|
|
replicas=10 * dup_factor),
|
|
ec_segment_size=segment_size,
|
|
ec_duplication_factor=dup_factor)
|
|
encoded_chunks = [[] for _ in range(policy.ec_n_unique_fragments)]
|
|
for orig_chunk in orig_chunks:
|
|
# each segment produces a set of frags
|
|
frag_set = policy.pyeclib_driver.encode(orig_chunk)
|
|
for frag_index, frag_data in enumerate(frag_set):
|
|
encoded_chunks[frag_index].append(frag_data)
|
|
# chunk_transformer buffers and concatenates multiple frags
|
|
expected = [b''.join(frags) for frags in encoded_chunks]
|
|
|
|
transform = obj.chunk_transformer(policy)
|
|
transform.send(None)
|
|
backend_chunks = transform.send(b''.join(orig_chunks))
|
|
self.assertIsNotNone(backend_chunks) # sanity
|
|
self.assertEqual(
|
|
len(backend_chunks), policy.ec_n_unique_fragments)
|
|
self.assertEqual(expected, backend_chunks)
|
|
|
|
# flush out last chunk buffer
|
|
backend_chunks = transform.send(b'')
|
|
self.assertEqual(
|
|
len(backend_chunks), policy.ec_n_unique_fragments)
|
|
self.assertEqual([b''] * policy.ec_n_unique_fragments,
|
|
backend_chunks)
|
|
|
|
do_test(dup_factor=1, segments=1)
|
|
do_test(dup_factor=2, segments=1)
|
|
do_test(dup_factor=3, segments=1)
|
|
do_test(dup_factor=1, segments=2)
|
|
do_test(dup_factor=2, segments=2)
|
|
do_test(dup_factor=3, segments=2)
|
|
|
|
def test_chunk_transformer_non_aligned_last_chunk(self):
|
|
last_chunk = b'a' * 128
|
|
|
|
def do_test(dup):
|
|
policy = ECStoragePolicy(0, 'ec8-2', ec_type=DEFAULT_TEST_EC_TYPE,
|
|
ec_ndata=8, ec_nparity=2,
|
|
object_ring=FakeRing(replicas=10 * dup),
|
|
ec_segment_size=1024,
|
|
ec_duplication_factor=dup)
|
|
expected = policy.pyeclib_driver.encode(last_chunk)
|
|
transform = obj.chunk_transformer(policy)
|
|
transform.send(None)
|
|
|
|
transform.send(last_chunk)
|
|
# flush out last chunk buffer
|
|
backend_chunks = transform.send(b'')
|
|
|
|
self.assertEqual(
|
|
len(backend_chunks), policy.ec_n_unique_fragments)
|
|
self.assertEqual(expected, backend_chunks)
|
|
|
|
do_test(1)
|
|
do_test(2)
|
|
|
|
def test_client_range_to_segment_range(self):
|
|
actual = obj.client_range_to_segment_range(100, 700, 512)
|
|
self.assertEqual(actual, (0, 1023))
|
|
self.assertEqual([type(x) for x in actual], [int, int])
|
|
|
|
actual = obj.client_range_to_segment_range(100, 700, 256)
|
|
self.assertEqual(actual, (0, 767))
|
|
self.assertEqual([type(x) for x in actual], [int, int])
|
|
|
|
actual = obj.client_range_to_segment_range(300, None, 256)
|
|
self.assertEqual(actual, (256, None))
|
|
self.assertEqual([type(x) for x in actual], [int, type(None)])
|
|
|
|
def test_segment_range_to_fragment_range(self):
|
|
actual = obj.segment_range_to_fragment_range(0, 1023, 512, 300)
|
|
self.assertEqual(actual, (0, 599))
|
|
self.assertEqual([type(x) for x in actual], [int, int])
|
|
|
|
actual = obj.segment_range_to_fragment_range(0, 767, 256, 100)
|
|
self.assertEqual(actual, (0, 299))
|
|
self.assertEqual([type(x) for x in actual], [int, int])
|
|
|
|
actual = obj.segment_range_to_fragment_range(256, None, 256, 100)
|
|
self.assertEqual(actual, (100, None))
|
|
self.assertEqual([type(x) for x in actual], [int, type(None)])
|
|
|
|
|
|
@patch_policies([ECStoragePolicy(0, name='ec', is_default=True,
|
|
ec_type=DEFAULT_TEST_EC_TYPE, ec_ndata=10,
|
|
ec_nparity=4, ec_segment_size=4096,
|
|
ec_duplication_factor=2),
|
|
StoragePolicy(1, name='unu')],
|
|
fake_ring_args=[{'replicas': 28}, {}])
|
|
class TestECDuplicationObjController(
|
|
ECObjectControllerMixin, unittest.TestCase):
|
|
container_info = {
|
|
'status': 200,
|
|
'read_acl': None,
|
|
'write_acl': None,
|
|
'sync_key': None,
|
|
'versions': None,
|
|
'storage_policy': '0',
|
|
}
|
|
|
|
controller_cls = obj.ECObjectController
|
|
|
|
def _test_GET_with_duplication_factor(self, node_frags, obj):
|
|
# This is basic tests in the healthy backends status
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj['etag'])
|
|
self.assertEqual(
|
|
md5(resp.body, usedforsecurity=False).hexdigest(),
|
|
obj['etag'])
|
|
|
|
collected_responses = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
index = conn.resp.headers['X-Object-Sysmeta-Ec-Frag-Index']
|
|
collected_responses[etag].add(index)
|
|
|
|
# the backend requests should be >= num_data_fragments
|
|
self.assertGreaterEqual(len(log), self.policy.ec_ndata)
|
|
# but <= # of replicas
|
|
self.assertLessEqual(len(log), self.replicas())
|
|
self.assertEqual(len(collected_responses), 1)
|
|
|
|
etag, frags = list(collected_responses.items())[0]
|
|
# the backend requests will stop at enough ec_ndata responses
|
|
self.assertEqual(
|
|
len(frags), self.policy.ec_ndata,
|
|
'collected %s frags (expected %s) for etag %s' % (
|
|
len(frags), self.policy.ec_ndata, etag))
|
|
|
|
# TODO: actually "frags" in node_frags is meaning "node_index" right now
|
|
# in following tests. Reconsidering the name and semantics change needed.
|
|
# Or, just mapping to be correct as frag_index is enough?.
|
|
def test_GET_with_duplication_factor(self):
|
|
obj = self._make_ec_object_stub()
|
|
node_frags = [
|
|
{'obj': obj, 'frag': 0},
|
|
{'obj': obj, 'frag': 1},
|
|
{'obj': obj, 'frag': 2},
|
|
{'obj': obj, 'frag': 3},
|
|
{'obj': obj, 'frag': 4},
|
|
{'obj': obj, 'frag': 5},
|
|
{'obj': obj, 'frag': 6},
|
|
{'obj': obj, 'frag': 7},
|
|
{'obj': obj, 'frag': 8},
|
|
{'obj': obj, 'frag': 9},
|
|
{'obj': obj, 'frag': 10},
|
|
{'obj': obj, 'frag': 11},
|
|
{'obj': obj, 'frag': 12},
|
|
{'obj': obj, 'frag': 13},
|
|
] * 2 # duplicated!
|
|
self._test_GET_with_duplication_factor(node_frags, obj)
|
|
|
|
def test_GET_with_duplication_factor_almost_duplicate_dispersion(self):
|
|
obj = self._make_ec_object_stub()
|
|
node_frags = [
|
|
# first half of # of replicas are 0, 1, 2, 3, 4, 5, 6
|
|
{'obj': obj, 'frag': 0},
|
|
{'obj': obj, 'frag': 0},
|
|
{'obj': obj, 'frag': 1},
|
|
{'obj': obj, 'frag': 1},
|
|
{'obj': obj, 'frag': 2},
|
|
{'obj': obj, 'frag': 2},
|
|
{'obj': obj, 'frag': 3},
|
|
{'obj': obj, 'frag': 3},
|
|
{'obj': obj, 'frag': 4},
|
|
{'obj': obj, 'frag': 4},
|
|
{'obj': obj, 'frag': 5},
|
|
{'obj': obj, 'frag': 5},
|
|
{'obj': obj, 'frag': 6},
|
|
{'obj': obj, 'frag': 6},
|
|
# second half of # of replicas are 7, 8, 9, 10, 11, 12, 13
|
|
{'obj': obj, 'frag': 7},
|
|
{'obj': obj, 'frag': 7},
|
|
{'obj': obj, 'frag': 8},
|
|
{'obj': obj, 'frag': 8},
|
|
{'obj': obj, 'frag': 9},
|
|
{'obj': obj, 'frag': 9},
|
|
{'obj': obj, 'frag': 10},
|
|
{'obj': obj, 'frag': 10},
|
|
{'obj': obj, 'frag': 11},
|
|
{'obj': obj, 'frag': 11},
|
|
{'obj': obj, 'frag': 12},
|
|
{'obj': obj, 'frag': 12},
|
|
{'obj': obj, 'frag': 13},
|
|
{'obj': obj, 'frag': 13},
|
|
]
|
|
# ...but it still works!
|
|
self._test_GET_with_duplication_factor(node_frags, obj)
|
|
|
|
def test_GET_with_missing_and_mixed_frags_will_dig_deep_but_stop(self):
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
|
|
# both of obj1 and obj2 has only 9 frags which is not able to decode
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0},
|
|
{'obj': obj2, 'frag': 0},
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj2, 'frag': 1},
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj2, 'frag': 2},
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj2, 'frag': 3},
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj2, 'frag': 4},
|
|
{'obj': obj1, 'frag': 5},
|
|
{'obj': obj2, 'frag': 5},
|
|
{'obj': obj1, 'frag': 6},
|
|
{'obj': obj2, 'frag': 6},
|
|
{'obj': obj1, 'frag': 7},
|
|
{'obj': obj2, 'frag': 7},
|
|
{'obj': obj1, 'frag': 8},
|
|
{'obj': obj2, 'frag': 8},
|
|
]
|
|
# ... and the rest are 404s which is limited by request_count
|
|
# (2 * replicas in default) rather than max_extra_requests limitation
|
|
# because the retries will be in ResumingGetter if the responses
|
|
# are 404s
|
|
node_frags += [[]] * (self.replicas() * 2 - len(node_frags))
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
collected_responses = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
index = conn.resp.headers['X-Object-Sysmeta-Ec-Frag-Index']
|
|
collected_responses[etag].add(index)
|
|
|
|
# default node_iter will exhaust to the last of handoffs
|
|
self.assertEqual(len(log), self.replicas() * 2)
|
|
# we have obj1, obj2, and 404 NotFound in collected_responses
|
|
self.assertEqual(len(list(collected_responses.keys())), 3)
|
|
self.assertIn(obj1['etag'], collected_responses)
|
|
self.assertIn(obj2['etag'], collected_responses)
|
|
self.assertIn(None, collected_responses)
|
|
|
|
# ... regardless we should never need to fetch more than ec_ndata
|
|
# frags for any given etag
|
|
for etag, frags in collected_responses.items():
|
|
self.assertLessEqual(len(frags), self.policy.ec_ndata,
|
|
'collected %s frags for etag %s' % (
|
|
len(frags), etag))
|
|
|
|
def test_GET_with_many_missed_overwrite_will_need_handoff(self):
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
# primaries
|
|
node_frags = [
|
|
{'obj': obj2, 'frag': 0},
|
|
{'obj': obj2, 'frag': 1},
|
|
{'obj': obj1, 'frag': 2}, # missed
|
|
{'obj': obj2, 'frag': 3},
|
|
{'obj': obj2, 'frag': 4},
|
|
{'obj': obj2, 'frag': 5},
|
|
{'obj': obj1, 'frag': 6}, # missed
|
|
{'obj': obj2, 'frag': 7},
|
|
{'obj': obj2, 'frag': 8},
|
|
{'obj': obj1, 'frag': 9}, # missed
|
|
{'obj': obj1, 'frag': 10}, # missed
|
|
{'obj': obj1, 'frag': 11}, # missed
|
|
{'obj': obj2, 'frag': 12},
|
|
{'obj': obj2, 'frag': 13},
|
|
]
|
|
|
|
node_frags = node_frags * 2 # 2 duplication
|
|
|
|
# so the primaries have indexes 0, 1, 3, 4, 5, 7, 8, 12, 13
|
|
# (9 indexes) for obj2 and then a handoff has index 6
|
|
node_frags += [
|
|
{'obj': obj2, 'frag': 6}, # handoff
|
|
]
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj2['etag'])
|
|
self.assertEqual(
|
|
md5(resp.body, usedforsecurity=False).hexdigest(),
|
|
obj2['etag'])
|
|
|
|
collected_responses = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
index = conn.resp.headers['X-Object-Sysmeta-Ec-Frag-Index']
|
|
collected_responses[etag].add(index)
|
|
|
|
# there's not enough of the obj2 etag on the primaries, we would
|
|
# have collected responses for both etags, and would have made
|
|
# one more request to the handoff node
|
|
self.assertEqual(len(log), self.replicas() + 1)
|
|
self.assertEqual(len(collected_responses), 2)
|
|
|
|
# ... regardless we should never need to fetch more than ec_ndata
|
|
# frags for any given etag
|
|
for etag, frags in collected_responses.items():
|
|
self.assertLessEqual(len(frags), self.policy.ec_ndata,
|
|
'collected %s frags for etag %s' % (
|
|
len(frags), etag))
|
|
|
|
def test_GET_with_missing_and_mixed_frags_will_dig_deep_but_succeed(self):
|
|
obj1 = self._make_ec_object_stub(pattern='obj1',
|
|
timestamp=self.ts())
|
|
obj2 = self._make_ec_object_stub(pattern='obj2',
|
|
timestamp=self.ts())
|
|
|
|
# 28 nodes are here
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0},
|
|
{'obj': obj2, 'frag': 0},
|
|
[],
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj2, 'frag': 1},
|
|
[],
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj2, 'frag': 2},
|
|
[],
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj2, 'frag': 3},
|
|
[],
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj2, 'frag': 4},
|
|
[],
|
|
{'obj': obj1, 'frag': 5},
|
|
{'obj': obj2, 'frag': 5},
|
|
[],
|
|
{'obj': obj1, 'frag': 6},
|
|
{'obj': obj2, 'frag': 6},
|
|
[],
|
|
{'obj': obj1, 'frag': 7},
|
|
{'obj': obj2, 'frag': 7},
|
|
[],
|
|
{'obj': obj1, 'frag': 8},
|
|
{'obj': obj2, 'frag': 8},
|
|
[],
|
|
[],
|
|
]
|
|
|
|
node_frags += [[]] * 13 # Plus 13 nodes in handoff
|
|
|
|
# finally 10th fragment for obj2 found
|
|
node_frags += [[{'obj': obj2, 'frag': 9}]]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
self.assertEqual(resp.headers['etag'], obj2['etag'])
|
|
self.assertEqual(
|
|
md5(resp.body, usedforsecurity=False).hexdigest(),
|
|
obj2['etag'])
|
|
|
|
collected_responses = defaultdict(set)
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
index = conn.resp.headers['X-Object-Sysmeta-Ec-Frag-Index']
|
|
collected_responses[etag].add(index)
|
|
|
|
# we go exactly as long as we have to, finding two different
|
|
# etags and some 404's (i.e. collected_responses[None])
|
|
self.assertEqual(len(log), len(node_frags))
|
|
self.assertEqual(len(collected_responses), 3)
|
|
|
|
# ... regardless we should never need to fetch more than ec_ndata
|
|
# frags for any given etag
|
|
for etag, frags in collected_responses.items():
|
|
self.assertLessEqual(len(frags), self.policy.ec_ndata,
|
|
'collected %s frags for etag %s' % (
|
|
len(frags), etag))
|
|
|
|
def test_GET_with_mixed_frags_and_no_quorum_will_503(self):
|
|
# all nodes have a frag but there is no one set that reaches quorum,
|
|
# which means there is no backend 404 response, but proxy should still
|
|
# return 404 rather than 503
|
|
stub_objects = [
|
|
self._make_ec_object_stub(pattern='obj1'),
|
|
self._make_ec_object_stub(pattern='obj2'),
|
|
self._make_ec_object_stub(pattern='obj3'),
|
|
self._make_ec_object_stub(pattern='obj4'),
|
|
self._make_ec_object_stub(pattern='obj5'),
|
|
self._make_ec_object_stub(pattern='obj6'),
|
|
self._make_ec_object_stub(pattern='obj7'),
|
|
]
|
|
etags = collections.Counter(stub['etag'] for stub in stub_objects)
|
|
self.assertEqual(len(etags), 7, etags) # sanity
|
|
|
|
# primaries and handoffs for required nodes
|
|
# this is 10-4 * 2 case so that 56 requests (2 * replicas) required
|
|
# to give up. we prepares 7 different objects above so responses
|
|
# will have 8 fragments for each object
|
|
required_nodes = self.replicas() * 2
|
|
# fill them out to the primary and handoff nodes
|
|
node_frags = []
|
|
for frag in range(8):
|
|
for stub_obj in stub_objects:
|
|
if len(node_frags) >= required_nodes:
|
|
# we already have enough responses
|
|
break
|
|
node_frags.append({'obj': stub_obj, 'frag': frag})
|
|
|
|
# sanity
|
|
self.assertEqual(required_nodes, len(node_frags))
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
collected_etags = set()
|
|
collected_status = set()
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
collected_etags.add(etag)
|
|
collected_status.add(conn.resp.status)
|
|
|
|
self.assertEqual(required_nodes, len(log))
|
|
self.assertEqual(len(collected_etags), 7)
|
|
self.assertEqual({200}, collected_status)
|
|
|
|
def test_GET_with_no_durable_files(self):
|
|
# verify that at least one durable is necessary for a successful GET
|
|
obj1 = self._make_ec_object_stub()
|
|
node_frags = [
|
|
{'obj': obj1, 'frag': 0, 'durable': False},
|
|
{'obj': obj1, 'frag': 1, 'durable': False},
|
|
{'obj': obj1, 'frag': 2, 'durable': False},
|
|
{'obj': obj1, 'frag': 3, 'durable': False},
|
|
{'obj': obj1, 'frag': 4, 'durable': False},
|
|
{'obj': obj1, 'frag': 5, 'durable': False},
|
|
{'obj': obj1, 'frag': 6, 'durable': False},
|
|
{'obj': obj1, 'frag': 7, 'durable': False},
|
|
{'obj': obj1, 'frag': 8, 'durable': False},
|
|
{'obj': obj1, 'frag': 9, 'durable': False},
|
|
{'obj': obj1, 'frag': 10, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 11, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 12, 'durable': False}, # parity
|
|
{'obj': obj1, 'frag': 13, 'durable': False}, # parity
|
|
]
|
|
|
|
node_frags = node_frags * 2 # 2 duplications
|
|
|
|
node_frags += [[]] * self.replicas() # handoffs
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 404)
|
|
# all 28 nodes tried with an optimistic get, none are durable and none
|
|
# report having a durable timestamp
|
|
self.assertEqual(self.replicas() * 2, len(log))
|
|
|
|
def test_GET_with_missing_and_mixed_frags_may_503(self):
|
|
obj1 = self._make_ec_object_stub(pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(pattern='obj2')
|
|
obj3 = self._make_ec_object_stub(pattern='obj3')
|
|
obj4 = self._make_ec_object_stub(pattern='obj4')
|
|
# we get a 503 when all the handoffs return 200
|
|
node_frags = [[]] * self.replicas() # primaries have no frags
|
|
# plus, 4 different objects and 7 indexes will b 28 node responses
|
|
# here for handoffs
|
|
node_frags = node_frags + [ # handoffs all have frags
|
|
{'obj': obj1, 'frag': 0},
|
|
{'obj': obj2, 'frag': 0},
|
|
{'obj': obj3, 'frag': 0},
|
|
{'obj': obj4, 'frag': 0},
|
|
{'obj': obj1, 'frag': 1},
|
|
{'obj': obj2, 'frag': 1},
|
|
{'obj': obj3, 'frag': 1},
|
|
{'obj': obj4, 'frag': 1},
|
|
{'obj': obj1, 'frag': 2},
|
|
{'obj': obj2, 'frag': 2},
|
|
{'obj': obj3, 'frag': 2},
|
|
{'obj': obj4, 'frag': 2},
|
|
{'obj': obj1, 'frag': 3},
|
|
{'obj': obj2, 'frag': 3},
|
|
{'obj': obj3, 'frag': 3},
|
|
{'obj': obj4, 'frag': 3},
|
|
{'obj': obj1, 'frag': 4},
|
|
{'obj': obj2, 'frag': 4},
|
|
{'obj': obj3, 'frag': 4},
|
|
{'obj': obj4, 'frag': 4},
|
|
{'obj': obj1, 'frag': 5},
|
|
{'obj': obj2, 'frag': 5},
|
|
{'obj': obj3, 'frag': 5},
|
|
{'obj': obj4, 'frag': 5},
|
|
{'obj': obj1, 'frag': 6},
|
|
{'obj': obj2, 'frag': 6},
|
|
{'obj': obj3, 'frag': 6},
|
|
{'obj': obj4, 'frag': 6},
|
|
]
|
|
|
|
fake_response = self._fake_ec_node_response(node_frags)
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 503)
|
|
# never get a quorum so all nodes are searched
|
|
self.assertEqual(len(log), 2 * self.replicas())
|
|
collected_indexes = defaultdict(list)
|
|
for conn in log:
|
|
fi = conn.resp.headers.get('X-Object-Sysmeta-Ec-Frag-Index')
|
|
if fi is not None:
|
|
collected_indexes[fi].append(conn)
|
|
self.assertEqual(len(collected_indexes), 7)
|
|
|
|
def test_GET_with_mixed_etags_at_same_timestamp(self):
|
|
# the difference from parent class is only handoff stub length
|
|
|
|
ts = self.ts() # force equal timestamps for two objects
|
|
obj1 = self._make_ec_object_stub(timestamp=ts, pattern='obj1')
|
|
obj2 = self._make_ec_object_stub(timestamp=ts, pattern='obj2')
|
|
self.assertNotEqual(obj1['etag'], obj2['etag']) # sanity
|
|
|
|
node_frags = [
|
|
# 7 frags of obj2 are available and durable
|
|
{'obj': obj2, 'frag': 0, 'durable': True},
|
|
{'obj': obj2, 'frag': 1, 'durable': True},
|
|
{'obj': obj2, 'frag': 2, 'durable': True},
|
|
{'obj': obj2, 'frag': 3, 'durable': True},
|
|
{'obj': obj2, 'frag': 4, 'durable': True},
|
|
{'obj': obj2, 'frag': 5, 'durable': True},
|
|
{'obj': obj2, 'frag': 6, 'durable': True},
|
|
# 7 frags of obj1 are available and durable
|
|
{'obj': obj1, 'frag': 7, 'durable': True},
|
|
{'obj': obj1, 'frag': 8, 'durable': True},
|
|
{'obj': obj1, 'frag': 9, 'durable': True},
|
|
{'obj': obj1, 'frag': 10, 'durable': True},
|
|
{'obj': obj1, 'frag': 11, 'durable': True},
|
|
{'obj': obj1, 'frag': 12, 'durable': True},
|
|
{'obj': obj1, 'frag': 13, 'durable': True},
|
|
# handoffs
|
|
]
|
|
|
|
node_frags += [[]] * (self.replicas() * 2 - len(node_frags))
|
|
|
|
fake_response = self._fake_ec_node_response(list(node_frags))
|
|
|
|
req = swob.Request.blank('/v1/a/c/o')
|
|
with capture_http_requests(fake_response) as log:
|
|
resp = req.get_response(self.app)
|
|
# read body to provoke any EC decode errors
|
|
self.assertTrue(resp.body)
|
|
|
|
self.assertEqual(resp.status_int, 503)
|
|
self.assertEqual(len(log), self.replicas() * 2)
|
|
collected_etags = set()
|
|
for conn in log:
|
|
etag = conn.resp.headers['X-Object-Sysmeta-Ec-Etag']
|
|
collected_etags.add(etag) # will be None from handoffs
|
|
self.assertEqual({obj1['etag'], obj2['etag'], None}, collected_etags)
|
|
log_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(log_lines,
|
|
['Problem with fragment response: ETag mismatch'] * 7
|
|
+ ['Object returning 503 for []'])
|
|
|
|
def _test_determine_chunk_destinations_prioritize(
|
|
self, missing_two, missing_one):
|
|
# This scenario is only likely for ec_duplication_factor >= 2. If we
|
|
# have multiple failures such that the putters collection is missing
|
|
# two primary nodes for frag index 'missing_two' and missing one
|
|
# primary node for frag index 'missing_one', then we should prioritize
|
|
# finding a handoff for frag index 'missing_two'.
|
|
|
|
class FakePutter(object):
|
|
def __init__(self, index):
|
|
self.node_index = index
|
|
|
|
controller = self.controller_cls(self.app, 'a', 'c', 'o')
|
|
|
|
# sanity, caller must set missing_two < than ec_num_unique_fragments
|
|
self.assertLess(missing_two, self.policy.ec_n_unique_fragments)
|
|
|
|
# create a dummy list of putters, check no handoffs
|
|
putters = []
|
|
for index in range(self.policy.object_ring.replica_count):
|
|
putters.append(FakePutter(index))
|
|
|
|
# sanity - all putters have primary nodes
|
|
got = controller._determine_chunk_destinations(putters, self.policy)
|
|
expected = {}
|
|
for i, p in enumerate(putters):
|
|
expected[p] = self.policy.get_backend_index(i)
|
|
self.assertEqual(got, expected)
|
|
|
|
# now, for fragment index that is missing two copies, lets make one
|
|
# putter be a handoff
|
|
handoff_putter = putters[missing_two]
|
|
handoff_putter.node_index = None
|
|
|
|
# and then pop another putter for a copy of same fragment index
|
|
putters.pop(missing_two + self.policy.ec_n_unique_fragments)
|
|
|
|
# also pop one copy of a different fragment to make one missing hole
|
|
putters.pop(missing_one)
|
|
|
|
# then determine chunk destinations: we have 26 putters here;
|
|
# missing_two frag index is missing two copies; missing_one frag index
|
|
# is missing one copy, therefore the handoff node should be assigned to
|
|
# missing_two frag index
|
|
got = controller._determine_chunk_destinations(putters, self.policy)
|
|
# N.B. len(putters) is now len(expected - 2) due to pop twice
|
|
self.assertEqual(len(putters), len(got))
|
|
# sanity, no node index - for handoff putter
|
|
self.assertIsNone(handoff_putter.node_index)
|
|
self.assertEqual(got[handoff_putter], missing_two)
|
|
# sanity, other nodes except handoff_putter have node_index
|
|
self.assertTrue(all(
|
|
[putter.node_index is not None for putter in got if
|
|
putter != handoff_putter]))
|
|
|
|
def test_determine_chunk_destinations_prioritize_more_missing(self):
|
|
# drop node_index 0, 14 and 1 should work
|
|
self._test_determine_chunk_destinations_prioritize(0, 1)
|
|
# drop node_index 1, 15 and 0 should work, too
|
|
self._test_determine_chunk_destinations_prioritize(1, 0)
|
|
|
|
|
|
class ECCommonPutterMixin(object):
|
|
# EC PUT tests common to both Mime and PUT+POST protocols
|
|
expect_headers = {}
|
|
|
|
def test_PUT_ec_error_during_transfer_data(self):
|
|
class FakeReader(object):
|
|
def read(self, size):
|
|
raise IOError('error message')
|
|
|
|
req = swob.Request.blank('/v1/a/c/o.jpg', method='PUT',
|
|
body=b'test body')
|
|
|
|
req.environ['wsgi.input'] = FakeReader()
|
|
req.headers['content-length'] = '6'
|
|
codes = [201] * self.replicas()
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 499)
|
|
|
|
def test_PUT_ec_chunkreadtimeout_during_transfer_data(self):
|
|
class FakeReader(object):
|
|
def read(self, size):
|
|
raise exceptions.ChunkReadTimeout()
|
|
|
|
req = swob.Request.blank('/v1/a/c/o.jpg', method='PUT',
|
|
body=b'test body')
|
|
|
|
req.environ['wsgi.input'] = FakeReader()
|
|
req.headers['content-length'] = '6'
|
|
codes = [201] * self.replicas()
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 408)
|
|
|
|
def test_PUT_ec_timeout_during_transfer_data(self):
|
|
class FakeReader(object):
|
|
def read(self, size):
|
|
raise exceptions.Timeout()
|
|
|
|
req = swob.Request.blank('/v1/a/c/o.jpg', method='PUT',
|
|
body=b'test body')
|
|
|
|
req.environ['wsgi.input'] = FakeReader()
|
|
req.headers['content-length'] = '6'
|
|
codes = [201] * self.replicas()
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 499)
|
|
|
|
def test_PUT_ec_exception_during_transfer_data(self):
|
|
class FakeReader(object):
|
|
def read(self, size):
|
|
raise Exception('exception message')
|
|
|
|
req = swob.Request.blank('/v1/a/c/o.jpg', method='PUT',
|
|
body=b'test body')
|
|
|
|
req.environ['wsgi.input'] = FakeReader()
|
|
req.headers['content-length'] = '6'
|
|
codes = [201] * self.replicas()
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 500)
|
|
|
|
|
|
# This is how CommonObjectControllerMixin is supposed to be used:
|
|
# @patch_policies(with_ec_default=True)
|
|
# class TestECObjControllerDoublePutter(BaseObjectControllerMixin,
|
|
# ECCommonPutterMixin,
|
|
# unittest.TestCase):
|
|
# # tests specific to the PUT+POST protocol
|
|
#
|
|
# def setUp(self):
|
|
# super(TestECObjControllerDoublePutter, self).setUp()
|
|
# # force use of the DoublePutter class
|
|
# self.app.use_put_v1 = True
|
|
|
|
|
|
@patch_policies(with_ec_default=True)
|
|
class TestECObjControllerMimePutter(BaseObjectControllerMixin,
|
|
ECCommonPutterMixin,
|
|
unittest.TestCase):
|
|
# tests specific to the older PUT protocol using a MimePutter
|
|
expect_headers = {
|
|
'X-Obj-Metadata-Footer': 'yes',
|
|
'X-Obj-Multiphase-Commit': 'yes'
|
|
}
|
|
|
|
def setUp(self):
|
|
super(TestECObjControllerMimePutter, self).setUp()
|
|
# force use of the MimePutter class
|
|
self.app.use_put_v1 = False
|
|
|
|
def test_PUT_simple(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
codes = [201] * self.replicas()
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_PUT_with_body_and_bad_etag(self):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_body = (b'asdf' * segment_size)[:-10]
|
|
codes = [201] * self.replicas()
|
|
conns = []
|
|
|
|
def capture_expect(conn):
|
|
# stash the backend connection so we can verify that it is closed
|
|
# (no data will be sent)
|
|
conns.append(conn)
|
|
|
|
# send a bad etag in the request headers
|
|
headers = {'Etag': 'bad etag'}
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers=headers, body=test_body)
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers,
|
|
give_expect=capture_expect):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(422, resp.status_int)
|
|
self.assertEqual(self.replicas(), len(conns))
|
|
for conn in conns:
|
|
self.assertTrue(conn.closed)
|
|
|
|
# make the footers callback send the correct etag
|
|
footers_callback = make_footers_callback(test_body)
|
|
env = {'swift.callback.update_footers': footers_callback}
|
|
headers = {'Etag': 'bad etag'}
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', headers=headers, environ=env,
|
|
body=test_body)
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(201, resp.status_int)
|
|
|
|
# make the footers callback send a bad Etag footer
|
|
footers_callback = make_footers_callback(b'not the test body')
|
|
env = {'swift.callback.update_footers': footers_callback}
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', environ=env, body=test_body)
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(422, resp.status_int)
|
|
|
|
def test_txn_id_logging_ECPUT(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
self.app.logger.txn_id = req.environ['swift.trans_id'] = 'test-txn-id'
|
|
codes = [(100, Timeout(), 503, 503)] * self.replicas()
|
|
stdout = StringIO()
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers), \
|
|
mock.patch('sys.stdout', stdout):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
for line in stdout.getvalue().splitlines():
|
|
self.assertIn('test-txn-id', line)
|
|
self.assertIn('Trying to get ', stdout.getvalue())
|
|
|
|
def test_PUT_with_explicit_commit_status(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
codes = [(100, 100, 201)] * self.replicas()
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_PUT_mostly_success(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
codes = [201] * self.quorum()
|
|
codes += [503] * (self.replicas() - len(codes))
|
|
random.shuffle(codes)
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_PUT_error_commit(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
codes = [(100, 503, Exception('not used'))] * self.replicas()
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
def test_PUT_mostly_success_commit(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
codes = [201] * self.quorum()
|
|
codes += [(100, 503, Exception('not used'))] * (
|
|
self.replicas() - len(codes))
|
|
random.shuffle(codes)
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_PUT_mostly_error_commit(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
codes = [(100, 503, Exception('not used'))] * self.quorum()
|
|
if isinstance(self.policy, ECStoragePolicy):
|
|
codes *= self.policy.ec_duplication_factor
|
|
codes += [201] * (self.replicas() - len(codes))
|
|
random.shuffle(codes)
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
def test_PUT_commit_timeout(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
codes = [201] * (self.replicas() - 1)
|
|
codes.append((100, Timeout(), Exception('not used')))
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_PUT_commit_exception(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
codes = [201] * (self.replicas() - 1)
|
|
codes.append((100, Exception('kaboom!'), Exception('not used')))
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_PUT_with_body(self):
|
|
self._test_PUT_with_body()
|
|
|
|
def test_PUT_with_chunked_body(self):
|
|
self._test_PUT_with_body(chunked=True, content_length=False)
|
|
|
|
def test_PUT_with_both_body(self):
|
|
self._test_PUT_with_body(chunked=True, content_length=True)
|
|
|
|
def _test_PUT_with_body(self, chunked=False, content_length=True):
|
|
segment_size = self.policy.ec_segment_size
|
|
test_body = (b'asdf' * segment_size)[:-10]
|
|
# make the footers callback not include Etag footer so that we can
|
|
# verify that the correct EC-calculated Etag is included in footers
|
|
# sent to backend
|
|
footers_callback = make_footers_callback()
|
|
env = {'swift.callback.update_footers': footers_callback}
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', environ=env)
|
|
etag = md5(test_body, usedforsecurity=False).hexdigest()
|
|
size = len(test_body)
|
|
req.body = test_body
|
|
if chunked:
|
|
req.headers['Transfer-Encoding'] = 'chunked'
|
|
if not content_length:
|
|
del req.headers['Content-Length']
|
|
codes = [201] * self.replicas()
|
|
resp_headers = {
|
|
'Some-Other-Header': 'Four',
|
|
'Etag': 'ignored',
|
|
}
|
|
|
|
put_requests = defaultdict(lambda: {'boundary': None, 'chunks': []})
|
|
|
|
def capture_body(conn, chunk):
|
|
put_requests[conn.connection_id]['chunks'].append(chunk)
|
|
|
|
def capture_headers(ip, port, device, part, method, path, headers,
|
|
**kwargs):
|
|
conn_id = kwargs['connection_id']
|
|
put_requests[conn_id]['boundary'] = headers[
|
|
'X-Backend-Obj-Multipart-Mime-Boundary']
|
|
put_requests[conn_id]['backend-content-length'] = headers.get(
|
|
'X-Backend-Obj-Content-Length')
|
|
put_requests[conn_id]['x-timestamp'] = headers[
|
|
'X-Timestamp']
|
|
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers,
|
|
give_send=capture_body,
|
|
give_connect=capture_headers,
|
|
headers=resp_headers):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
timestamps = {captured_req['x-timestamp']
|
|
for captured_req in put_requests.values()}
|
|
self.assertEqual(1, len(timestamps), timestamps)
|
|
self.assertEqual(dict(resp.headers), {
|
|
'Content-Type': 'text/html; charset=UTF-8',
|
|
'Content-Length': '0',
|
|
'Last-Modified': time.strftime(
|
|
"%a, %d %b %Y %H:%M:%S GMT",
|
|
time.gmtime(math.ceil(float(timestamps.pop())))),
|
|
'Etag': etag,
|
|
})
|
|
frag_archives = []
|
|
for connection_id, info in put_requests.items():
|
|
body = unchunk_body(b''.join(info['chunks']))
|
|
self.assertIsNotNone(info['boundary'],
|
|
"didn't get boundary for conn %r" % (
|
|
connection_id,))
|
|
|
|
# email.parser.FeedParser doesn't know how to take a multipart
|
|
# message and boundary together and parse it; it only knows how
|
|
# to take a string, parse the headers, and figure out the
|
|
# boundary on its own.
|
|
parser = EmailFeedParser()
|
|
parser.feed(
|
|
("Content-Type: multipart/nobodycares; boundary=%s\r\n\r\n" %
|
|
info['boundary']).encode('ascii'))
|
|
parser.feed(body)
|
|
message = parser.close()
|
|
|
|
self.assertTrue(message.is_multipart()) # sanity check
|
|
mime_parts = message.get_payload()
|
|
self.assertEqual(len(mime_parts), 3)
|
|
obj_part, footer_part, commit_part = mime_parts
|
|
|
|
# attach the body to frag_archives list
|
|
self.assertEqual(obj_part['X-Document'], 'object body')
|
|
obj_payload = obj_part.get_payload(decode=True)
|
|
frag_archives.append(obj_payload)
|
|
|
|
if chunked:
|
|
self.assertIsNone(info['backend-content-length'])
|
|
else:
|
|
self.assertTrue(
|
|
size > int(info['backend-content-length']) > 0,
|
|
"invalid backend-content-length for conn %r" % (
|
|
connection_id,))
|
|
# assert length was correct for this connection
|
|
self.assertEqual(int(info['backend-content-length']),
|
|
len(frag_archives[-1]))
|
|
# assert length was the same for all connections
|
|
self.assertEqual(int(info['backend-content-length']),
|
|
len(frag_archives[0]))
|
|
|
|
# validate some footer metadata
|
|
self.assertEqual(footer_part['X-Document'], 'object metadata')
|
|
footer_metadata = json.loads(footer_part.get_payload())
|
|
self.assertTrue(footer_metadata)
|
|
expected = {}
|
|
# update expected with footers from the callback...
|
|
footers_callback(expected)
|
|
expected.update({
|
|
'X-Object-Sysmeta-Ec-Content-Length': str(size),
|
|
'X-Backend-Container-Update-Override-Size': str(size),
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Backend-Container-Update-Override-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Segment-Size': str(segment_size),
|
|
'Etag': md5(obj_payload, usedforsecurity=False).hexdigest()})
|
|
for header, value in expected.items():
|
|
self.assertEqual(footer_metadata[header], value)
|
|
|
|
# sanity on commit message
|
|
self.assertEqual(commit_part['X-Document'], 'put commit')
|
|
|
|
self.assertEqual(len(frag_archives), self.replicas())
|
|
fragment_size = self.policy.fragment_size
|
|
node_payloads = []
|
|
for fa in frag_archives:
|
|
payload = [fa[x:x + fragment_size]
|
|
for x in range(0, len(fa), fragment_size)]
|
|
node_payloads.append(payload)
|
|
fragment_payloads = zip(*node_payloads)
|
|
|
|
expected_body = b''
|
|
for fragment_payload in fragment_payloads:
|
|
self.assertEqual(len(fragment_payload), self.replicas())
|
|
if True:
|
|
fragment_payload = list(fragment_payload)
|
|
expected_body += self.policy.pyeclib_driver.decode(
|
|
fragment_payload)
|
|
|
|
self.assertEqual(len(test_body), len(expected_body))
|
|
self.assertEqual(test_body, expected_body)
|
|
|
|
def test_PUT_with_footers(self):
|
|
# verify footers supplied by a footers callback being added to
|
|
# trailing metadata
|
|
segment_size = self.policy.ec_segment_size
|
|
test_body = (b'asdf' * segment_size)[:-10]
|
|
etag = md5(test_body, usedforsecurity=False).hexdigest()
|
|
size = len(test_body)
|
|
codes = [201] * self.replicas()
|
|
resp_headers = {
|
|
'Some-Other-Header': 'Four',
|
|
'Etag': 'ignored',
|
|
}
|
|
|
|
def do_test(footers_to_add, expect_added):
|
|
put_requests = defaultdict(
|
|
lambda: {'boundary': None, 'chunks': []})
|
|
|
|
def capture_body(conn, chunk):
|
|
put_requests[conn.connection_id]['chunks'].append(chunk)
|
|
|
|
def capture_headers(ip, port, device, part, method, path, headers,
|
|
**kwargs):
|
|
conn_id = kwargs['connection_id']
|
|
put_requests[conn_id]['boundary'] = headers[
|
|
'X-Backend-Obj-Multipart-Mime-Boundary']
|
|
put_requests[conn_id]['x-timestamp'] = headers[
|
|
'X-Timestamp']
|
|
|
|
def footers_callback(footers):
|
|
footers.update(footers_to_add)
|
|
env = {'swift.callback.update_footers': footers_callback}
|
|
req = swift.common.swob.Request.blank(
|
|
'/v1/a/c/o', method='PUT', environ=env, body=test_body)
|
|
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers,
|
|
give_send=capture_body,
|
|
give_connect=capture_headers,
|
|
headers=resp_headers):
|
|
resp = req.get_response(self.app)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
timestamps = {captured_req['x-timestamp']
|
|
for captured_req in put_requests.values()}
|
|
self.assertEqual(1, len(timestamps), timestamps)
|
|
self.assertEqual(dict(resp.headers), {
|
|
'Content-Type': 'text/html; charset=UTF-8',
|
|
'Content-Length': '0',
|
|
'Last-Modified': time.strftime(
|
|
"%a, %d %b %Y %H:%M:%S GMT",
|
|
time.gmtime(math.ceil(float(timestamps.pop())))),
|
|
'Etag': etag,
|
|
})
|
|
for connection_id, info in put_requests.items():
|
|
body = unchunk_body(b''.join(info['chunks']))
|
|
# email.parser.FeedParser doesn't know how to take a multipart
|
|
# message and boundary together and parse it; it only knows how
|
|
# to take a string, parse the headers, and figure out the
|
|
# boundary on its own.
|
|
parser = EmailFeedParser()
|
|
parser.feed(
|
|
("Content-Type: multipart/nobodycares; boundary=%s\r\n\r\n"
|
|
% info['boundary']).encode('ascii'))
|
|
parser.feed(body)
|
|
message = parser.close()
|
|
|
|
self.assertTrue(message.is_multipart()) # sanity check
|
|
mime_parts = message.get_payload()
|
|
self.assertEqual(len(mime_parts), 3)
|
|
obj_part, footer_part, commit_part = mime_parts
|
|
|
|
# validate EC footer metadata - should always be present
|
|
self.assertEqual(footer_part['X-Document'], 'object metadata')
|
|
footer_metadata = json.loads(footer_part.get_payload())
|
|
self.assertIsNotNone(
|
|
footer_metadata.pop('X-Object-Sysmeta-Ec-Frag-Index'))
|
|
expected = {
|
|
'X-Object-Sysmeta-Ec-Scheme':
|
|
self.policy.ec_scheme_description,
|
|
'X-Object-Sysmeta-Ec-Content-Length': str(size),
|
|
'X-Object-Sysmeta-Ec-Etag': etag,
|
|
'X-Object-Sysmeta-Ec-Segment-Size': str(segment_size),
|
|
'Etag': md5(
|
|
obj_part.get_payload(decode=True),
|
|
usedforsecurity=False).hexdigest()}
|
|
expected.update(expect_added)
|
|
for header, value in expected.items():
|
|
self.assertIn(header, footer_metadata)
|
|
self.assertEqual(value, footer_metadata[header])
|
|
footer_metadata.pop(header)
|
|
self.assertFalse(footer_metadata)
|
|
|
|
# sanity check - middleware sets no footer, expect EC overrides
|
|
footers_to_add = {}
|
|
expect_added = {
|
|
'X-Backend-Container-Update-Override-Size': str(size),
|
|
'X-Backend-Container-Update-Override-Etag': etag}
|
|
do_test(footers_to_add, expect_added)
|
|
|
|
# middleware cannot overwrite any EC sysmeta
|
|
footers_to_add = {
|
|
'X-Object-Sysmeta-Ec-Content-Length': str(size + 1),
|
|
'X-Object-Sysmeta-Ec-Etag': 'other etag',
|
|
'X-Object-Sysmeta-Ec-Segment-Size': str(segment_size + 1),
|
|
'X-Object-Sysmeta-Ec-Unused-But-Reserved': 'ignored'}
|
|
do_test(footers_to_add, expect_added)
|
|
|
|
# middleware can add x-object-sysmeta- headers including
|
|
# x-object-sysmeta-container-update-override headers
|
|
footers_to_add = {
|
|
'X-Object-Sysmeta-Foo': 'bar',
|
|
'X-Object-Sysmeta-Container-Update-Override-Size':
|
|
str(size + 1),
|
|
'X-Object-Sysmeta-Container-Update-Override-Etag': 'other etag',
|
|
'X-Object-Sysmeta-Container-Update-Override-Ping': 'pong'
|
|
}
|
|
expect_added.update(footers_to_add)
|
|
do_test(footers_to_add, expect_added)
|
|
|
|
# middleware can also overwrite x-backend-container-update-override
|
|
# headers
|
|
override_footers = {
|
|
'X-Backend-Container-Update-Override-Wham': 'bam',
|
|
'X-Backend-Container-Update-Override-Size': str(size + 2),
|
|
'X-Backend-Container-Update-Override-Etag': 'another etag'}
|
|
footers_to_add.update(override_footers)
|
|
expect_added.update(override_footers)
|
|
do_test(footers_to_add, expect_added)
|
|
|
|
def test_PUT_old_obj_server(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
responses = [
|
|
# one server will response 100-continue but not include the
|
|
# needful expect headers and the connection will be dropped
|
|
((100, Exception('not used')), {}),
|
|
] + [
|
|
# and pleanty of successful responses too
|
|
(201, {
|
|
'X-Obj-Metadata-Footer': 'yes',
|
|
'X-Obj-Multiphase-Commit': 'yes',
|
|
}),
|
|
] * self.replicas()
|
|
random.shuffle(responses)
|
|
if responses[-1][0] != 201:
|
|
# whoops, stupid random
|
|
responses = responses[1:] + [responses[0]]
|
|
codes, expect_headers = zip(*responses)
|
|
with set_http_connect(*codes, expect_headers=expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_PUT_with_slow_commits(self):
|
|
# It's important that this timeout be much less than the delay in
|
|
# the slow commit responses so that the slow commits are not waited
|
|
# for.
|
|
self.app.post_quorum_timeout = 0.01
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
# plenty of slow commits
|
|
response_sleep = 5.0
|
|
codes = [FakeStatus(201, response_sleep=response_sleep)
|
|
for i in range(self.replicas())]
|
|
# swap out some with regular fast responses
|
|
number_of_fast_responses_needed_to_be_quick_enough = \
|
|
self.policy.quorum
|
|
fast_indexes = random.sample(
|
|
range(self.replicas()),
|
|
number_of_fast_responses_needed_to_be_quick_enough)
|
|
for i in fast_indexes:
|
|
codes[i] = 201
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
start = time.time()
|
|
resp = req.get_response(self.app)
|
|
response_time = time.time() - start
|
|
self.assertEqual(resp.status_int, 201)
|
|
self.assertLess(response_time, response_sleep)
|
|
|
|
def test_PUT_with_just_enough_durable_responses(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
|
|
codes = [201] * (self.policy.ec_ndata + 1)
|
|
codes += [503] * (self.policy.ec_nparity - 1)
|
|
self.assertEqual(len(codes), self.policy.ec_n_unique_fragments)
|
|
random.shuffle(codes)
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
def test_PUT_with_less_durable_responses(self):
|
|
req = swift.common.swob.Request.blank('/v1/a/c/o', method='PUT',
|
|
body=b'')
|
|
|
|
codes = [201] * (self.policy.ec_ndata)
|
|
codes += [503] * (self.policy.ec_nparity)
|
|
self.assertEqual(len(codes), self.policy.ec_n_unique_fragments)
|
|
random.shuffle(codes)
|
|
with set_http_connect(*codes, expect_headers=self.expect_headers):
|
|
resp = req.get_response(self.app)
|
|
self.assertEqual(resp.status_int, 503)
|
|
|
|
|
|
class TestNumContainerUpdates(unittest.TestCase):
|
|
def test_it(self):
|
|
test_cases = [
|
|
# (container replicas, object replicas, object quorum, expected)
|
|
(3, 17, 13, 6), # EC 12+5
|
|
(3, 9, 4, 7), # EC 3+6
|
|
(3, 14, 11, 5), # EC 10+4
|
|
(5, 14, 11, 6), # EC 10+4, 5 container replicas
|
|
(7, 14, 11, 7), # EC 10+4, 7 container replicas
|
|
(3, 19, 16, 5), # EC 15+4
|
|
(5, 19, 16, 6), # EC 15+4, 5 container replicas
|
|
(3, 28, 22, 8), # EC (10+4)x2
|
|
(5, 28, 22, 9), # EC (10+4)x2, 5 container replicas
|
|
(3, 1, 1, 3), # 1 object replica
|
|
(3, 2, 1, 3), # 2 object replicas
|
|
(3, 3, 2, 3), # 3 object replicas
|
|
(3, 4, 2, 4), # 4 object replicas
|
|
(3, 5, 3, 4), # 5 object replicas
|
|
(3, 6, 3, 5), # 6 object replicas
|
|
(3, 7, 4, 5), # 7 object replicas
|
|
]
|
|
|
|
for c_replica, o_replica, o_quorum, exp in test_cases:
|
|
c_quorum = utils.quorum_size(c_replica)
|
|
got = obj.num_container_updates(c_replica, c_quorum,
|
|
o_replica, o_quorum)
|
|
self.assertEqual(
|
|
exp, got,
|
|
"Failed for c_replica=%d, o_replica=%d, o_quorum=%d" % (
|
|
c_replica, o_replica, o_quorum))
|
|
|
|
|
|
@patch_policies(with_ec_default=True)
|
|
class TestECFragGetter(BaseObjectControllerMixin, unittest.TestCase):
|
|
def setUp(self):
|
|
super(TestECFragGetter, self).setUp()
|
|
req = Request.blank(path='/v1/a/c/o')
|
|
self.getter = obj.ECFragGetter(
|
|
self.app, req, None, None, self.policy, 'a/c/o',
|
|
{}, None, self.logger.thread_locals,
|
|
self.logger)
|
|
|
|
def test_init_node_timeout(self):
|
|
app = argparse.Namespace(node_timeout=2, recoverable_node_timeout=3)
|
|
getter = obj.ECFragGetter(
|
|
app, None, None, None, self.policy, 'a/c/o',
|
|
{}, None, None, self.logger)
|
|
self.assertEqual(3, getter.node_timeout)
|
|
|
|
def test_iter_bytes_from_response_part(self):
|
|
part = FileLikeIter([b'some', b'thing'])
|
|
it = self.getter._iter_bytes_from_response_part(part, nbytes=None)
|
|
self.assertEqual(b'something', b''.join(it))
|
|
|
|
def test_iter_bytes_from_response_part_insufficient_bytes(self):
|
|
part = FileLikeIter([b'some', b'thing'])
|
|
it = self.getter._iter_bytes_from_response_part(part, nbytes=100)
|
|
with mock.patch.object(self.getter, '_find_source',
|
|
return_value=False):
|
|
with self.assertRaises(ShortReadError) as cm:
|
|
b''.join(it)
|
|
self.assertEqual('Too few bytes; read 9, expecting 100',
|
|
str(cm.exception))
|
|
|
|
def test_iter_bytes_from_response_part_read_timeout(self):
|
|
part = FileLikeIter([b'some', b'thing'])
|
|
it = self.getter._iter_bytes_from_response_part(part, nbytes=9)
|
|
exc = ChunkReadTimeout()
|
|
# set this after __init__ to keep it off the eventlet scheduler
|
|
exc.seconds = 9
|
|
with mock.patch.object(self.getter, '_find_source',
|
|
return_value=False):
|
|
with mock.patch.object(part, 'read', side_effect=[b'some', exc]):
|
|
with self.assertRaises(ChunkReadTimeout) as cm:
|
|
b''.join(it)
|
|
self.assertEqual('9 seconds', str(cm.exception))
|
|
|
|
def test_iter_bytes_from_response_part_small_fragment_size(self):
|
|
self.getter.fragment_size = 4
|
|
part = FileLikeIter([b'some', b'thing', b''])
|
|
it = self.getter._iter_bytes_from_response_part(part, nbytes=None)
|
|
self.assertEqual([b'some', b'thin', b'g'], [ch for ch in it])
|
|
self.getter.fragment_size = 1
|
|
part = FileLikeIter([b'some', b'thing', b''])
|
|
it = self.getter._iter_bytes_from_response_part(part, nbytes=None)
|
|
self.assertEqual([c.encode() for c in 'something'], [ch for ch in it])
|
|
|
|
def test_fragment_size(self):
|
|
source = FakeSource((
|
|
b'abcd', b'1234', b'abc', b'd1', b'234abcd1234abcd1', b'2'))
|
|
|
|
def mock_source_gen():
|
|
yield GetterSource(self.app, source, {})
|
|
|
|
self.getter.fragment_size = 8
|
|
with mock.patch.object(self.getter, '_source_gen', mock_source_gen):
|
|
it = self.getter.response_parts_iter()
|
|
fragments = list(next(it)['part_iter'])
|
|
|
|
self.assertEqual(fragments, [
|
|
b'abcd1234', b'abcd1234', b'abcd1234', b'abcd12'])
|
|
|
|
def test_fragment_size_resuming(self):
|
|
node = {'ip': '1.2.3.4', 'port': 6200, 'device': 'sda'}
|
|
|
|
source1 = FakeSource([b'abcd', b'1234', None,
|
|
b'efgh', b'5678', b'lots', b'more', b'data'])
|
|
# incomplete reads of fragment_size will be re-fetched
|
|
source2 = FakeSource([b'efgh', b'5678', b'lots', None])
|
|
source3 = FakeSource([b'lots', b'more', b'data'])
|
|
range_headers = []
|
|
sources = [GetterSource(self.app, src, node)
|
|
for src in (source1, source2, source3)]
|
|
|
|
def mock_source_gen():
|
|
for source in sources:
|
|
range_headers.append(self.getter.backend_headers.get('Range'))
|
|
yield source
|
|
|
|
self.getter.fragment_size = 8
|
|
with mock.patch.object(self.getter, '_source_gen',
|
|
mock_source_gen):
|
|
it = self.getter.response_parts_iter()
|
|
fragments = list(next(it)['part_iter'])
|
|
|
|
self.assertEqual(fragments, [
|
|
b'abcd1234', b'efgh5678', b'lotsmore', b'data'])
|
|
self.assertEqual(range_headers, [None, 'bytes=8-27', 'bytes=16-27'])
|
|
|
|
def test_fragment_size_resuming_chunked(self):
|
|
node = {'ip': '1.2.3.4', 'port': 6200, 'device': 'sda'}
|
|
headers = {'transfer-encoding': 'chunked',
|
|
'content-type': 'text/plain'}
|
|
source1 = FakeSource([b'abcd', b'1234', b'abc', None], headers=headers)
|
|
source2 = FakeSource([b'efgh5678'], headers=headers)
|
|
range_headers = []
|
|
sources = [GetterSource(self.app, src, node)
|
|
for src in (source1, source2)]
|
|
|
|
def mock_source_gen():
|
|
for source in sources:
|
|
range_headers.append(self.getter.backend_headers.get('Range'))
|
|
yield source
|
|
|
|
self.getter.fragment_size = 8
|
|
with mock.patch.object(self.getter, '_source_gen',
|
|
mock_source_gen):
|
|
it = self.getter.response_parts_iter()
|
|
fragments = list(next(it)['part_iter'])
|
|
self.assertEqual(fragments, [b'abcd1234', b'efgh5678'])
|
|
self.assertEqual(range_headers, [None, 'bytes=8-'])
|
|
|
|
|
|
@patch_policies()
|
|
class TestGetUpdateShard(BaseObjectControllerMixin, unittest.TestCase):
|
|
bound_prefix = 'x'
|
|
item = 'x1_test'
|
|
|
|
def setUp(self):
|
|
super(TestGetUpdateShard, self).setUp()
|
|
self.ctrl = obj.BaseObjectController(self.app, 'a', 'c', 'o')
|
|
self.memcache = FakeMemcache()
|
|
ts_iter = make_timestamp_iter()
|
|
# NB: these shard ranges have gaps
|
|
self.shard_ranges = [ShardRange(
|
|
'.sharded_a/sr%d' % i, next(ts_iter),
|
|
self.bound_prefix + u'%d_lower' % i,
|
|
self.bound_prefix + u'%d_upper' % i,
|
|
object_count=i, bytes_used=1024 * i,
|
|
meta_timestamp=next(ts_iter))
|
|
for i in range(3)]
|
|
|
|
def _create_response_data(self, shards, includes=None):
|
|
resp_headers = {'X-Backend-Record-Type': 'shard',
|
|
'X-Backend-Record-Shard-Format': 'namespace'}
|
|
namespaces = [Namespace(sr.name, sr.lower, sr.upper)
|
|
for sr in shards]
|
|
if includes is not None:
|
|
namespaces = [ns for ns in namespaces if includes in ns]
|
|
body = json.dumps([dict(ns) for ns in namespaces]).encode('ascii')
|
|
return body, resp_headers
|
|
|
|
def test_get_update_shard_cache_writing(self):
|
|
# verify case when complete set of shards is returned
|
|
req = Request.blank('/v1/a/c/o', method='PUT',
|
|
environ={'swift.cache': self.memcache})
|
|
body, resp_headers = self._create_response_data(self.shard_ranges)
|
|
with mocked_http_conn(
|
|
200, 200, body_iter=iter([b'', body]),
|
|
headers=resp_headers) as fake_conn:
|
|
actual = self.ctrl._get_update_shard(req, 'a', 'c', self.item)
|
|
|
|
# account info
|
|
captured = fake_conn.requests
|
|
self.assertEqual('HEAD', captured[0]['method'])
|
|
self.assertEqual('a', captured[0]['path'][7:])
|
|
# container GET
|
|
self.assertEqual('GET', captured[1]['method'])
|
|
self.assertEqual('a/c', captured[1]['path'][7:])
|
|
params = sorted(captured[1]['qs'].split('&'))
|
|
self.assertEqual(
|
|
['format=json', 'states=updating'], params)
|
|
captured_hdrs = captured[1]['headers']
|
|
self.assertEqual('shard', captured_hdrs.get('X-Backend-Record-Type'))
|
|
self.assertEqual('namespace',
|
|
captured_hdrs.get('X-Backend-Record-Shard-Format'))
|
|
exp_bounds = NamespaceBoundList.parse(self.shard_ranges).bounds
|
|
self.assertEqual(json.loads(json.dumps(exp_bounds)),
|
|
self.memcache.get('shard-updating-v2/a/c'))
|
|
exp_ns = Namespace(self.shard_ranges[1].name,
|
|
self.shard_ranges[1].lower,
|
|
self.shard_ranges[2].lower)
|
|
self.assertEqual(exp_ns, actual)
|
|
self.assertFalse(self.app.logger.get_lines_for_level('error'))
|
|
|
|
def test_get_update_shard_cache_reading(self):
|
|
# verify case when complete set of shards is in cache
|
|
cached_bounds = NamespaceBoundList.parse(self.shard_ranges).bounds
|
|
self.memcache.set('shard-updating-v2/a/c', cached_bounds)
|
|
req = Request.blank('/v1/a/c/o', method='PUT',
|
|
environ={'swift.cache': self.memcache})
|
|
|
|
actual = self.ctrl._get_update_shard(req, 'a', 'c', self.item)
|
|
|
|
self.assertEqual(json.loads(json.dumps(cached_bounds)),
|
|
self.memcache.get('shard-updating-v2/a/c'))
|
|
exp_ns = Namespace(self.shard_ranges[1].name,
|
|
self.shard_ranges[1].lower,
|
|
self.shard_ranges[2].lower)
|
|
self.assertEqual(exp_ns, actual)
|
|
self.assertFalse(self.app.logger.get_lines_for_level('error'))
|
|
|
|
def test_get_update_shard_cache_recheck_time_zero(self):
|
|
# verify case when shard caching is disabled
|
|
req = Request.blank('/v1/a/c/o', method='PUT',
|
|
environ={'swift.cache': self.memcache})
|
|
body, resp_headers = self._create_response_data(
|
|
self.shard_ranges, self.item)
|
|
with mock.patch.object(self.app, 'recheck_updating_shard_ranges', 0):
|
|
with mocked_http_conn(
|
|
200, 200, body_iter=iter([b'', body]),
|
|
headers=resp_headers) as fake_conn:
|
|
actual = self.ctrl._get_update_shard(req, 'a', 'c', self.item)
|
|
|
|
# account info
|
|
captured = fake_conn.requests
|
|
self.assertEqual('HEAD', captured[0]['method'])
|
|
self.assertEqual('a', captured[0]['path'][7:])
|
|
# container GET
|
|
self.assertEqual('GET', captured[1]['method'])
|
|
self.assertEqual('a/c', captured[1]['path'][7:])
|
|
params = sorted(captured[1]['qs'].split('&'))
|
|
self.assertEqual(
|
|
['format=json', 'includes=' + quote(self.item), 'states=updating'],
|
|
params)
|
|
captured_hdrs = captured[1]['headers']
|
|
self.assertEqual('shard', captured_hdrs.get('X-Backend-Record-Type'))
|
|
self.assertEqual('namespace',
|
|
captured_hdrs.get('X-Backend-Record-Shard-Format'))
|
|
self.assertIsNone(self.memcache.get('shard-updating-v2/a/c'))
|
|
exp_ns = Namespace(self.shard_ranges[1].name,
|
|
self.shard_ranges[1].lower,
|
|
self.shard_ranges[1].upper)
|
|
self.assertEqual(exp_ns, actual)
|
|
self.assertFalse(self.app.logger.get_lines_for_level('error'))
|
|
|
|
def test_get_update_shard_cache_not_available(self):
|
|
# verify case when memcache is not available
|
|
req = Request.blank('/v1/a/c/o', method='PUT')
|
|
body, resp_headers = self._create_response_data(self.shard_ranges)
|
|
with mocked_http_conn(
|
|
200, 200, body_iter=iter([b'', body]),
|
|
headers=resp_headers) as fake_conn:
|
|
actual = self.ctrl._get_update_shard(req, 'a', 'c', self.item)
|
|
|
|
# account info
|
|
captured = fake_conn.requests
|
|
self.assertEqual('HEAD', captured[0]['method'])
|
|
self.assertEqual('a', captured[0]['path'][7:])
|
|
# container GET
|
|
self.assertEqual('GET', captured[1]['method'])
|
|
self.assertEqual('a/c', captured[1]['path'][7:])
|
|
params = sorted(captured[1]['qs'].split('&'))
|
|
self.assertEqual(
|
|
['format=json', 'states=updating'], params)
|
|
captured_hdrs = captured[1]['headers']
|
|
self.assertEqual('shard', captured_hdrs.get('X-Backend-Record-Type'))
|
|
self.assertEqual('namespace',
|
|
captured_hdrs.get('X-Backend-Record-Shard-Format'))
|
|
self.assertIsNone(self.memcache.get('shard-updating-v2/a/c'))
|
|
exp_ns = Namespace(self.shard_ranges[1].name,
|
|
self.shard_ranges[1].lower,
|
|
self.shard_ranges[2].lower)
|
|
self.assertEqual(exp_ns, actual)
|
|
self.assertFalse(self.app.logger.get_lines_for_level('error'))
|
|
|
|
def test_get_update_shard_empty_body(self):
|
|
# verify case when no shards are returned
|
|
req = Request.blank('/v1/a/c/o', method='PUT',
|
|
environ={'swift.cache': self.memcache})
|
|
|
|
body, resp_headers = self._create_response_data(self.shard_ranges)
|
|
with mocked_http_conn(
|
|
200, 200, body_iter=[b'', b''],
|
|
headers=resp_headers) as fake_conn:
|
|
actual = self.ctrl._get_update_shard(req, 'a', 'c', self.item)
|
|
|
|
# account info
|
|
captured = fake_conn.requests
|
|
self.assertEqual('HEAD', captured[0]['method'])
|
|
self.assertEqual('a', captured[0]['path'][7:])
|
|
# container GET
|
|
self.assertEqual('GET', captured[1]['method'])
|
|
self.assertEqual('a/c', captured[1]['path'][7:])
|
|
params = sorted(captured[1]['qs'].split('&'))
|
|
self.assertEqual(
|
|
['format=json', 'states=updating'], params)
|
|
captured_hdrs = captured[1]['headers']
|
|
self.assertEqual('shard', captured_hdrs.get('X-Backend-Record-Type'))
|
|
self.assertEqual('namespace',
|
|
captured_hdrs.get('X-Backend-Record-Shard-Format'))
|
|
self.assertIsNone(self.memcache.get('shard-updating-v2/a/c'))
|
|
self.assertIsNone(actual)
|
|
lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertEqual(1, len(lines))
|
|
self.assertIn('Problem with listing response from /v1/a/c/o', lines[0])
|
|
|
|
|
|
class TestGetUpdateShardUTF8(TestGetUpdateShard):
|
|
bound_prefix = u'\u1234'
|
|
item = wsgi_to_str('\xe1\x88\xb41_test')
|
|
|
|
|
|
class TestGetUpdateShardLegacy(TestGetUpdateShard):
|
|
def _create_response_data(self, shards, includes=None):
|
|
# older container servers never return the shorter 'namespace' format
|
|
# nor the 'X-Backend-Record-Shard-Format' header
|
|
resp_headers = {'X-Backend-Record-Type': 'shard'}
|
|
if includes is not None:
|
|
shards = [sr for sr in shards if includes in sr]
|
|
body = json.dumps([dict(sr) for sr in shards]).encode('ascii')
|
|
return body, resp_headers
|
|
|
|
|
|
class TestGetUpdateShardLegacyUTF8(TestGetUpdateShard):
|
|
bound_prefix = u'\u1234'
|
|
item = wsgi_to_str('\xe1\x88\xb41_test')
|
|
|
|
|
|
@patch_policies()
|
|
class TestGetUpdatingNamespacesErrors(BaseObjectControllerMixin,
|
|
unittest.TestCase):
|
|
def setUp(self):
|
|
super(TestGetUpdatingNamespacesErrors, self).setUp()
|
|
self.ctrl = obj.BaseObjectController(self.app, 'a', 'c', 'o')
|
|
|
|
def _check_get_namespaces_bad_data(self, body):
|
|
req = Request.blank('/v1/a/c/o', method='PUT')
|
|
# empty response
|
|
resp_headers = {'X-Backend-Record-Type': 'shard'}
|
|
with mocked_http_conn(200, 200, body_iter=iter([b'', body]),
|
|
headers=resp_headers):
|
|
actual, resp = self.ctrl._get_updating_namespaces(
|
|
req, 'a', 'c', '1_test')
|
|
self.assertEqual(200, resp.status_int)
|
|
self.assertIsNone(actual)
|
|
lines = self.app.logger.get_lines_for_level('error')
|
|
return lines
|
|
|
|
def test_get_namespaces_empty_body(self):
|
|
error_lines = self._check_get_namespaces_bad_data(b'')
|
|
self.assertIn('Problem with listing response', error_lines[0])
|
|
if six.PY2:
|
|
self.assertIn('No JSON', error_lines[0])
|
|
else:
|
|
self.assertIn('JSONDecodeError', error_lines[0])
|
|
self.assertFalse(error_lines[1:])
|
|
|
|
def test_get_namespaces_not_a_list(self):
|
|
body = json.dumps({}).encode('ascii')
|
|
error_lines = self._check_get_namespaces_bad_data(body)
|
|
self.assertIn('Problem with listing response', error_lines[0])
|
|
self.assertIn('not a list', error_lines[0])
|
|
self.assertFalse(error_lines[1:])
|
|
|
|
def test_get_namespaces_key_missing(self):
|
|
body = json.dumps([{}]).encode('ascii')
|
|
error_lines = self._check_get_namespaces_bad_data(body)
|
|
self.assertIn('Failed to get namespaces', error_lines[0])
|
|
self.assertIn('KeyError', error_lines[0])
|
|
self.assertFalse(error_lines[1:])
|
|
|
|
def test_get_namespaces_invalid_shard_range(self):
|
|
# lower > upper !
|
|
bad_ns_data = {'name': 'name', 'lower': 'z', 'upper': 'a'}
|
|
body = json.dumps([bad_ns_data]).encode('ascii')
|
|
error_lines = self._check_get_namespaces_bad_data(body)
|
|
self.assertIn('Failed to get namespaces', error_lines[0])
|
|
self.assertIn('ValueError', error_lines[0])
|
|
self.assertFalse(error_lines[1:])
|
|
|
|
def test_get_namespaces_missing_record_type(self):
|
|
req = Request.blank('/v1/a/c/o', method='PUT')
|
|
sr = utils.Namespace('a/c', 'l', 'u')
|
|
body = json.dumps([dict(sr)]).encode('ascii')
|
|
with mocked_http_conn(
|
|
200, 200, body_iter=iter([b'', body])):
|
|
actual, resp = self.ctrl._get_updating_namespaces(
|
|
req, 'a', 'c', '1_test')
|
|
self.assertEqual(200, resp.status_int)
|
|
self.assertIsNone(actual)
|
|
error_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertIn('Failed to get shard ranges', error_lines[0])
|
|
self.assertIn('unexpected record type', error_lines[0])
|
|
self.assertIn('/a/c', error_lines[0])
|
|
self.assertFalse(error_lines[1:])
|
|
|
|
def test_get_namespaces_wrong_record_type(self):
|
|
req = Request.blank('/v1/a/c/o', method='PUT')
|
|
sr = utils.Namespace('a/c', 'l', 'u')
|
|
body = json.dumps([dict(sr)]).encode('ascii')
|
|
headers = {'X-Backend-Record-Type': 'object'}
|
|
with mocked_http_conn(
|
|
200, 200, body_iter=iter([b'', body]),
|
|
headers=headers):
|
|
actual, resp = self.ctrl._get_updating_namespaces(
|
|
req, 'a', 'c', '1_test')
|
|
self.assertEqual(200, resp.status_int)
|
|
self.assertIsNone(actual)
|
|
error_lines = self.app.logger.get_lines_for_level('error')
|
|
self.assertIn('Failed to get shard ranges', error_lines[0])
|
|
self.assertIn('unexpected record type', error_lines[0])
|
|
self.assertIn('/a/c', error_lines[0])
|
|
self.assertFalse(error_lines[1:])
|
|
|
|
def test_get_namespaces_request_failed(self):
|
|
req = Request.blank('/v1/a/c/o', method='PUT')
|
|
with mocked_http_conn(200, 404, 404, 404):
|
|
actual, resp = self.ctrl._get_updating_namespaces(
|
|
req, 'a', 'c', '1_test')
|
|
self.assertEqual(404, resp.status_int)
|
|
self.assertIsNone(actual)
|
|
self.assertFalse(self.app.logger.get_lines_for_level('error'))
|
|
warning_lines = self.app.logger.get_lines_for_level('warning')
|
|
self.assertIn('Failed to get container listing', warning_lines[0])
|
|
self.assertIn('/a/c', warning_lines[0])
|
|
self.assertFalse(warning_lines[1:])
|
|
|
|
|
|
if __name__ == '__main__':
|
|
unittest.main()
|