d74e6dc103
If checksum and size provided in image meta data, fuel-agent will do checksum comparing to ensure that data written on disk and downloaded image are identical. Partially implements blueprint: ibp-image-checksums Depends on: Ia9c8325a2f71af2a1af2ad54a1ae414edf49a7c6 Change-Id: Id120848537cfc4ed908b9e524624855fb7308603
226 lines
6.9 KiB
Python
226 lines
6.9 KiB
Python
# Copyright 2014 Mirantis, Inc.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
import abc
|
|
import os
|
|
import tarfile
|
|
import tempfile
|
|
import zlib
|
|
|
|
from oslo.config import cfg
|
|
|
|
from fuel_agent import errors
|
|
from fuel_agent.openstack.common import log as logging
|
|
from fuel_agent.utils import utils
|
|
|
|
LOG = logging.getLogger(__name__)
|
|
|
|
au_opts = [
|
|
cfg.IntOpt(
|
|
'data_chunk_size',
|
|
default=1048576,
|
|
help='Size of data chunk to operate with images'
|
|
),
|
|
]
|
|
|
|
CONF = cfg.CONF
|
|
CONF.register_opts(au_opts)
|
|
|
|
|
|
class Target(object):
|
|
__metaclass__ = abc.ABCMeta
|
|
|
|
def __iter__(self):
|
|
return self
|
|
|
|
def next(self):
|
|
raise StopIteration()
|
|
|
|
def target(self, filename='/dev/null'):
|
|
LOG.debug('Opening file: %s for write' % filename)
|
|
with open(filename, 'wb') as f:
|
|
count = 0
|
|
for chunk in self:
|
|
LOG.debug('Next chunk: %s' % count)
|
|
f.write(chunk)
|
|
count += 1
|
|
LOG.debug('Flushing file: %s' % filename)
|
|
f.flush()
|
|
# ensure data to be written to disk
|
|
os.fsync(f.fileno())
|
|
LOG.debug('File is written: %s' % filename)
|
|
|
|
|
|
class LocalFile(Target):
|
|
def __init__(self, filename):
|
|
if filename.startswith('file://'):
|
|
self.filename = filename[7:]
|
|
else:
|
|
self.filename = filename
|
|
self.fileobj = None
|
|
|
|
def next(self):
|
|
if not self.fileobj:
|
|
self.fileobj = open(self.filename, 'rb')
|
|
buffer = self.fileobj.read(CONF.data_chunk_size)
|
|
if buffer:
|
|
return buffer
|
|
else:
|
|
self.fileobj.close()
|
|
raise StopIteration()
|
|
|
|
|
|
class HttpUrl(Target):
|
|
def __init__(self, url):
|
|
self.url = str(url)
|
|
self.response_obj = utils.init_http_request(self.url)
|
|
self.processed_bytes = 0
|
|
try:
|
|
self.length = int(self.response_obj.headers['content-length'])
|
|
except (ValueError, KeyError):
|
|
raise errors.HttpUrlInvalidContentLength(
|
|
'Can not get content length for %s' % self.url)
|
|
else:
|
|
LOG.debug('Expected content length %s for %s' % (self.length,
|
|
self.url))
|
|
|
|
def next(self):
|
|
while self.processed_bytes < self.length:
|
|
try:
|
|
data = self.response_obj.raw.read(CONF.data_chunk_size)
|
|
if not data:
|
|
raise errors.HttpUrlConnectionError(
|
|
'Could not receive data: URL=%s, range=%s' %
|
|
(self.url, self.processed_bytes))
|
|
except Exception as exc:
|
|
LOG.exception(exc)
|
|
self.response_obj = utils.init_http_request(
|
|
self.url, self.processed_bytes)
|
|
continue
|
|
else:
|
|
self.processed_bytes += len(data)
|
|
return data
|
|
raise StopIteration()
|
|
|
|
|
|
class GunzipStream(Target):
|
|
def __init__(self, stream):
|
|
self.stream = iter(stream)
|
|
#NOTE(agordeev): toggle automatic header detection on
|
|
self.decompressor = zlib.decompressobj(zlib.MAX_WBITS | 32)
|
|
|
|
def next(self):
|
|
try:
|
|
return self.decompressor.decompress(self.stream.next())
|
|
except StopIteration:
|
|
raise
|
|
|
|
|
|
class ForwardFileStream(Target):
|
|
def __init__(self, stream):
|
|
self.stream = iter(stream)
|
|
self.position = 0
|
|
self.chunk = ''
|
|
self.closed = False
|
|
|
|
def next(self):
|
|
buffer = self.read()
|
|
if buffer:
|
|
return buffer
|
|
else:
|
|
raise StopIteration()
|
|
|
|
def close(self):
|
|
self.closed = True
|
|
|
|
def tell(self):
|
|
if self.closed:
|
|
raise ValueError('I/O operation on closed file')
|
|
return self.position
|
|
|
|
def seek(self, position):
|
|
if self.closed:
|
|
raise ValueError('I/O operation on closed file')
|
|
|
|
if position < self.position:
|
|
raise ValueError('Backward seek operation is impossible')
|
|
elif position < self.position + len(self.chunk):
|
|
self.chunk = self.chunk[(position - self.position):]
|
|
self.position = position
|
|
else:
|
|
try:
|
|
current = self.position + len(self.chunk)
|
|
while True:
|
|
chunk = self.stream.next()
|
|
if current + len(chunk) >= position:
|
|
self.chunk = chunk[(position - current):]
|
|
self.position = position
|
|
break
|
|
current += len(chunk)
|
|
except StopIteration:
|
|
self.chunk = None
|
|
self.position = position
|
|
|
|
def read(self, length=CONF.data_chunk_size):
|
|
# NOTE(kozhukalov): default lenght = 1048576 is not usual behaviour,
|
|
# but that is ok for our use case.
|
|
if self.closed:
|
|
raise ValueError('I/O operation on closed file')
|
|
if self.chunk is None:
|
|
return None
|
|
try:
|
|
while len(self.chunk) < length:
|
|
self.chunk += self.stream.next()
|
|
result = self.chunk[:length]
|
|
self.chunk = self.chunk[length:]
|
|
except StopIteration:
|
|
result = self.chunk
|
|
self.chunk = None
|
|
self.position += len(result)
|
|
return result
|
|
|
|
|
|
class TarStream(Target):
|
|
def __init__(self, stream):
|
|
self.stream = iter(stream)
|
|
self.tarobj = None
|
|
|
|
def target(self, filename=None):
|
|
if not self.tarobj:
|
|
self.tarobj = tarfile.open(
|
|
fileobj=ForwardFileStream(self.stream), mode='r:')
|
|
self.tarobj.extractall(path=(filename or tempfile.gettempdir()))
|
|
|
|
|
|
class Chain(object):
|
|
def __init__(self):
|
|
self.processors = []
|
|
|
|
def append(self, processor):
|
|
self.processors.append(processor)
|
|
|
|
def process(self):
|
|
def jump(proc, next_proc):
|
|
# if next_proc is just a string we assume it is a filename
|
|
# and we save stream into a file
|
|
if isinstance(next_proc, (str, unicode)):
|
|
LOG.debug('Processor target: %s' % next_proc)
|
|
proc.target(next_proc)
|
|
return LocalFile(next_proc)
|
|
# if next_proc is not a string we return new instance
|
|
# initialized with the previous one
|
|
else:
|
|
return next_proc(proc)
|
|
return reduce(jump, self.processors)
|