python-botocore/botocore/response.py

202 lines
7.1 KiB
Python
Raw Normal View History

# Copyright (c) 2012-2013 Mitch Garnaat http://garnaat.org/
2015-10-08 20:15:31 +02:00
# Copyright 2012-2014 Amazon.com, Inc. or its affiliates. All Rights Reserved.
#
2015-10-08 20:15:31 +02:00
# Licensed under the Apache License, Version 2.0 (the "License"). You
# may not use this file except in compliance with the License. A copy of
# the License is located at
#
2015-10-08 20:15:31 +02:00
# http://aws.amazon.com/apache2.0/
#
2015-10-08 20:15:31 +02:00
# or in the "license" file accompanying this file. This file is
# distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF
# ANY KIND, either express or implied. See the License for the specific
# language governing permissions and limitations under the License.
import logging
2022-05-26 00:10:07 +02:00
from io import IOBase
2022-05-26 00:10:07 +02:00
from urllib3.exceptions import ProtocolError as URLLib3ProtocolError
from urllib3.exceptions import ReadTimeoutError as URLLib3ReadTimeoutError
from botocore import parsers
2021-11-03 18:14:15 +01:00
from botocore.compat import set_socket_timeout
2022-01-12 23:38:07 +01:00
from botocore.exceptions import (
2022-05-26 00:10:07 +02:00
IncompleteReadError,
ReadTimeoutError,
ResponseStreamingError,
2022-01-12 23:38:07 +01:00
)
2015-10-08 20:15:31 +02:00
2021-11-03 18:14:15 +01:00
# Keep these imported. There's pre-existing code that uses them.
2022-05-26 00:10:07 +02:00
from botocore import ScalarTypes # noqa
from botocore.compat import XMLParseError # noqa
from botocore.hooks import first_non_none_response # noqa
2021-11-03 18:14:15 +01:00
2015-10-08 20:15:31 +02:00
logger = logging.getLogger(__name__)
2022-05-26 00:10:07 +02:00
class StreamingBody(IOBase):
2015-10-08 20:15:31 +02:00
"""Wrapper class for an http response body.
This provides a few additional conveniences that do not exist
in the urllib3 model:
* Set the timeout on the socket (i.e read() timeouts)
* Auto validation of content length, if the amount of bytes
we read does not match the content length, an exception
is raised.
"""
2022-05-26 00:10:07 +02:00
2018-07-11 08:25:50 +02:00
_DEFAULT_CHUNK_SIZE = 1024
2015-10-08 20:15:31 +02:00
def __init__(self, raw_stream, content_length):
self._raw_stream = raw_stream
self._content_length = content_length
self._amount_read = 0
2022-05-26 00:10:07 +02:00
def __del__(self):
# Extending destructor in order to preserve the underlying raw_stream.
# The ability to add custom cleanup logic introduced in Python3.4+.
# https://www.python.org/dev/peps/pep-0442/
pass
2015-10-08 20:15:31 +02:00
def set_socket_timeout(self, timeout):
"""Set the timeout seconds on the socket."""
# The problem we're trying to solve is to prevent .read() calls from
# hanging. This can happen in rare cases. What we'd like to ideally
# do is set a timeout on the .read() call so that callers can retry
# the request.
# Unfortunately, this isn't currently possible in requests.
# See: https://github.com/kennethreitz/requests/issues/1803
# So what we're going to do is reach into the guts of the stream and
# grab the socket object, which we can set the timeout on. We're
# putting in a check here so in case this interface goes away, we'll
# know.
try:
set_socket_timeout(self._raw_stream, timeout)
except AttributeError:
2022-05-26 00:10:07 +02:00
logger.error(
"Cannot access the socket object of "
"a streaming response. It's possible "
"the interface has changed.",
exc_info=True,
)
2015-10-08 20:15:31 +02:00
raise
2022-05-26 00:10:07 +02:00
def readable(self):
try:
return self._raw_stream.readable()
except AttributeError:
return False
2015-10-08 20:15:31 +02:00
def read(self, amt=None):
"""Read at most amt bytes from the stream.
If the amt argument is omitted, read all data.
"""
2018-10-04 08:50:52 +02:00
try:
chunk = self._raw_stream.read(amt)
except URLLib3ReadTimeoutError as e:
# TODO: the url will be None as urllib3 isn't setting it yet
raise ReadTimeoutError(endpoint_url=e.url, error=e)
2022-01-12 23:38:07 +01:00
except URLLib3ProtocolError as e:
raise ResponseStreamingError(error=e)
2015-10-08 20:15:31 +02:00
self._amount_read += len(chunk)
2018-01-15 17:34:17 +01:00
if amt is None or (not chunk and amt > 0):
2015-10-08 20:15:31 +02:00
# If the server sends empty contents or
# we ask to read all of the contents, then we know
# we need to verify the content length.
self._verify_content_length()
return chunk
2022-05-26 00:10:07 +02:00
def readlines(self):
return self._raw_stream.readlines()
2018-07-11 08:25:50 +02:00
def __iter__(self):
2022-05-26 00:10:07 +02:00
"""Return an iterator to yield 1k chunks from the raw stream."""
2018-07-11 08:25:50 +02:00
return self.iter_chunks(self._DEFAULT_CHUNK_SIZE)
2018-12-28 08:05:06 +01:00
def __next__(self):
2022-05-26 00:10:07 +02:00
"""Return the next 1k chunk from the raw stream."""
2018-12-28 08:05:06 +01:00
current_chunk = self.read(self._DEFAULT_CHUNK_SIZE)
if current_chunk:
return current_chunk
raise StopIteration()
2022-05-26 00:10:07 +02:00
def __enter__(self):
return self._raw_stream
def __exit__(self, type, value, traceback):
self._raw_stream.close()
2018-12-28 08:05:06 +01:00
next = __next__
2022-05-26 00:10:07 +02:00
def iter_lines(self, chunk_size=_DEFAULT_CHUNK_SIZE, keepends=False):
2018-07-11 08:25:50 +02:00
"""Return an iterator to yield lines from the raw stream.
This is achieved by reading chunk of bytes (of size chunk_size) at a
time from the raw stream, and then yielding lines from there.
"""
2018-12-28 08:05:06 +01:00
pending = b''
2018-07-11 08:25:50 +02:00
for chunk in self.iter_chunks(chunk_size):
2018-12-28 08:05:06 +01:00
lines = (pending + chunk).splitlines(True)
for line in lines[:-1]:
2021-01-26 16:12:20 +01:00
yield line.splitlines(keepends)[0]
2018-12-28 08:05:06 +01:00
pending = lines[-1]
2019-02-27 08:30:11 +01:00
if pending:
2021-01-26 16:12:20 +01:00
yield pending.splitlines(keepends)[0]
2018-07-11 08:25:50 +02:00
def iter_chunks(self, chunk_size=_DEFAULT_CHUNK_SIZE):
"""Return an iterator to yield chunks of chunk_size bytes from the raw
stream.
"""
while True:
current_chunk = self.read(chunk_size)
if current_chunk == b"":
break
yield current_chunk
2015-10-08 20:15:31 +02:00
def _verify_content_length(self):
# See: https://github.com/kennethreitz/requests/issues/1855
# Basically, our http library doesn't do this for us, so we have
# to do this ourself.
2022-05-26 00:10:07 +02:00
if self._content_length is not None and self._amount_read != int(
self._content_length
):
2015-10-08 20:15:31 +02:00
raise IncompleteReadError(
actual_bytes=self._amount_read,
2022-05-26 00:10:07 +02:00
expected_bytes=int(self._content_length),
)
def tell(self):
return self._raw_stream.tell()
2015-10-08 20:15:31 +02:00
def close(self):
"""Close the underlying http response stream."""
self._raw_stream.close()
2015-10-08 20:15:31 +02:00
2015-10-08 20:16:07 +02:00
def get_response(operation_model, http_response):
protocol = operation_model.metadata['protocol']
response_dict = {
'headers': http_response.headers,
'status_code': http_response.status_code,
}
# TODO: Unfortunately, we have to have error logic here.
# If it looks like an error, in the streaming response case we
# need to actually grab the contents.
if response_dict['status_code'] >= 300:
response_dict['body'] = http_response.content
elif operation_model.has_streaming_output:
response_dict['body'] = StreamingBody(
2022-05-26 00:10:07 +02:00
http_response.raw, response_dict['headers'].get('content-length')
)
2015-10-08 20:16:07 +02:00
else:
response_dict['body'] = http_response.content
parser = parsers.create_parser(protocol)
2022-05-26 00:10:07 +02:00
return http_response, parser.parse(
response_dict, operation_model.output_shape
)