aboutsummaryrefslogtreecommitdiff
path: root/.venv/lib/python3.12/site-packages/botocore/response.py
diff options
context:
space:
mode:
authorS. Solomon Darnell2025-03-28 21:52:21 -0500
committerS. Solomon Darnell2025-03-28 21:52:21 -0500
commit4a52a71956a8d46fcb7294ac71734504bb09bcc2 (patch)
treeee3dc5af3b6313e921cd920906356f5d4febc4ed /.venv/lib/python3.12/site-packages/botocore/response.py
parentcc961e04ba734dd72309fb548a2f97d67d578813 (diff)
downloadgn-ai-master.tar.gz
two version of R2R are hereHEADmaster
Diffstat (limited to '.venv/lib/python3.12/site-packages/botocore/response.py')
-rw-r--r--.venv/lib/python3.12/site-packages/botocore/response.py201
1 files changed, 201 insertions, 0 deletions
diff --git a/.venv/lib/python3.12/site-packages/botocore/response.py b/.venv/lib/python3.12/site-packages/botocore/response.py
new file mode 100644
index 00000000..ba3fac9b
--- /dev/null
+++ b/.venv/lib/python3.12/site-packages/botocore/response.py
@@ -0,0 +1,201 @@
+# Copyright (c) 2012-2013 Mitch Garnaat http://garnaat.org/
+# Copyright 2012-2014 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"). You
+# may not use this file except in compliance with the License. A copy of
+# the License is located at
+#
+# http://aws.amazon.com/apache2.0/
+#
+# or in the "license" file accompanying this file. This file is
+# distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF
+# ANY KIND, either express or implied. See the License for the specific
+# language governing permissions and limitations under the License.
+
+import logging
+from io import IOBase
+
+from urllib3.exceptions import ProtocolError as URLLib3ProtocolError
+from urllib3.exceptions import ReadTimeoutError as URLLib3ReadTimeoutError
+
+from botocore import parsers
+from botocore.compat import set_socket_timeout
+from botocore.exceptions import (
+ IncompleteReadError,
+ ReadTimeoutError,
+ ResponseStreamingError,
+)
+
+# Keep these imported. There's pre-existing code that uses them.
+from botocore import ScalarTypes # noqa
+from botocore.compat import XMLParseError # noqa
+from botocore.hooks import first_non_none_response # noqa
+
+
+logger = logging.getLogger(__name__)
+
+
+class StreamingBody(IOBase):
+ """Wrapper class for an http response body.
+
+ This provides a few additional conveniences that do not exist
+ in the urllib3 model:
+
+ * Set the timeout on the socket (i.e read() timeouts)
+ * Auto validation of content length, if the amount of bytes
+ we read does not match the content length, an exception
+ is raised.
+
+ """
+
+ _DEFAULT_CHUNK_SIZE = 1024
+
+ def __init__(self, raw_stream, content_length):
+ self._raw_stream = raw_stream
+ self._content_length = content_length
+ self._amount_read = 0
+
+ def __del__(self):
+ # Extending destructor in order to preserve the underlying raw_stream.
+ # The ability to add custom cleanup logic introduced in Python3.4+.
+ # https://www.python.org/dev/peps/pep-0442/
+ pass
+
+ def set_socket_timeout(self, timeout):
+ """Set the timeout seconds on the socket."""
+ # The problem we're trying to solve is to prevent .read() calls from
+ # hanging. This can happen in rare cases. What we'd like to ideally
+ # do is set a timeout on the .read() call so that callers can retry
+ # the request.
+ # Unfortunately, this isn't currently possible in requests.
+ # See: https://github.com/kennethreitz/requests/issues/1803
+ # So what we're going to do is reach into the guts of the stream and
+ # grab the socket object, which we can set the timeout on. We're
+ # putting in a check here so in case this interface goes away, we'll
+ # know.
+ try:
+ set_socket_timeout(self._raw_stream, timeout)
+ except AttributeError:
+ logger.error(
+ "Cannot access the socket object of "
+ "a streaming response. It's possible "
+ "the interface has changed.",
+ exc_info=True,
+ )
+ raise
+
+ def readable(self):
+ try:
+ return self._raw_stream.readable()
+ except AttributeError:
+ return False
+
+ def read(self, amt=None):
+ """Read at most amt bytes from the stream.
+
+ If the amt argument is omitted, read all data.
+ """
+ try:
+ chunk = self._raw_stream.read(amt)
+ except URLLib3ReadTimeoutError as e:
+ # TODO: the url will be None as urllib3 isn't setting it yet
+ raise ReadTimeoutError(endpoint_url=e.url, error=e)
+ except URLLib3ProtocolError as e:
+ raise ResponseStreamingError(error=e)
+ self._amount_read += len(chunk)
+ if amt is None or (not chunk and amt > 0):
+ # If the server sends empty contents or
+ # we ask to read all of the contents, then we know
+ # we need to verify the content length.
+ self._verify_content_length()
+ return chunk
+
+ def readlines(self):
+ return self._raw_stream.readlines()
+
+ def __iter__(self):
+ """Return an iterator to yield 1k chunks from the raw stream."""
+ return self.iter_chunks(self._DEFAULT_CHUNK_SIZE)
+
+ def __next__(self):
+ """Return the next 1k chunk from the raw stream."""
+ current_chunk = self.read(self._DEFAULT_CHUNK_SIZE)
+ if current_chunk:
+ return current_chunk
+ raise StopIteration()
+
+ def __enter__(self):
+ return self._raw_stream
+
+ def __exit__(self, type, value, traceback):
+ self._raw_stream.close()
+
+ next = __next__
+
+ def iter_lines(self, chunk_size=_DEFAULT_CHUNK_SIZE, keepends=False):
+ """Return an iterator to yield lines from the raw stream.
+
+ This is achieved by reading chunk of bytes (of size chunk_size) at a
+ time from the raw stream, and then yielding lines from there.
+ """
+ pending = b''
+ for chunk in self.iter_chunks(chunk_size):
+ lines = (pending + chunk).splitlines(True)
+ for line in lines[:-1]:
+ yield line.splitlines(keepends)[0]
+ pending = lines[-1]
+ if pending:
+ yield pending.splitlines(keepends)[0]
+
+ def iter_chunks(self, chunk_size=_DEFAULT_CHUNK_SIZE):
+ """Return an iterator to yield chunks of chunk_size bytes from the raw
+ stream.
+ """
+ while True:
+ current_chunk = self.read(chunk_size)
+ if current_chunk == b"":
+ break
+ yield current_chunk
+
+ def _verify_content_length(self):
+ # See: https://github.com/kennethreitz/requests/issues/1855
+ # Basically, our http library doesn't do this for us, so we have
+ # to do this ourself.
+ if self._content_length is not None and self._amount_read != int(
+ self._content_length
+ ):
+ raise IncompleteReadError(
+ actual_bytes=self._amount_read,
+ expected_bytes=int(self._content_length),
+ )
+
+ def tell(self):
+ return self._raw_stream.tell()
+
+ def close(self):
+ """Close the underlying http response stream."""
+ self._raw_stream.close()
+
+
+def get_response(operation_model, http_response):
+ protocol = operation_model.metadata['protocol']
+ response_dict = {
+ 'headers': http_response.headers,
+ 'status_code': http_response.status_code,
+ }
+ # TODO: Unfortunately, we have to have error logic here.
+ # If it looks like an error, in the streaming response case we
+ # need to actually grab the contents.
+ if response_dict['status_code'] >= 300:
+ response_dict['body'] = http_response.content
+ elif operation_model.has_streaming_output:
+ response_dict['body'] = StreamingBody(
+ http_response.raw, response_dict['headers'].get('content-length')
+ )
+ else:
+ response_dict['body'] = http_response.content
+
+ parser = parsers.create_parser(protocol)
+ return http_response, parser.parse(
+ response_dict, operation_model.output_shape
+ )