Created
September 9, 2021 13:17
-
-
Save paulbastian/add8c12cc47509f3b41cb726287e05b5 to your computer and use it in GitHub Desktop.
patched aiohttp-client.py for NO_PROXY support
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| """HTTP Client for asyncio.""" | |
| import asyncio | |
| import base64 | |
| import hashlib | |
| import json | |
| import os | |
| import sys | |
| import traceback | |
| import warnings | |
| from types import SimpleNamespace, TracebackType | |
| from typing import ( | |
| Any, | |
| Awaitable, | |
| Callable, | |
| Coroutine, | |
| FrozenSet, | |
| Generator, | |
| Generic, | |
| Iterable, | |
| List, | |
| Mapping, | |
| Optional, | |
| Set, | |
| Tuple, | |
| Type, | |
| TypeVar, | |
| Union, | |
| ) | |
| import attr | |
| from multidict import CIMultiDict, MultiDict, MultiDictProxy, istr | |
| from yarl import URL | |
| from urllib.request import proxy_bypass | |
| from . import hdrs, http, payload | |
| from .abc import AbstractCookieJar | |
| from .client_exceptions import ( | |
| ClientConnectionError as ClientConnectionError, | |
| ClientConnectorCertificateError as ClientConnectorCertificateError, | |
| ClientConnectorError as ClientConnectorError, | |
| ClientConnectorSSLError as ClientConnectorSSLError, | |
| ClientError as ClientError, | |
| ClientHttpProxyError as ClientHttpProxyError, | |
| ClientOSError as ClientOSError, | |
| ClientPayloadError as ClientPayloadError, | |
| ClientProxyConnectionError as ClientProxyConnectionError, | |
| ClientResponseError as ClientResponseError, | |
| ClientSSLError as ClientSSLError, | |
| ContentTypeError as ContentTypeError, | |
| InvalidURL as InvalidURL, | |
| ServerConnectionError as ServerConnectionError, | |
| ServerDisconnectedError as ServerDisconnectedError, | |
| ServerFingerprintMismatch as ServerFingerprintMismatch, | |
| ServerTimeoutError as ServerTimeoutError, | |
| TooManyRedirects as TooManyRedirects, | |
| WSServerHandshakeError as WSServerHandshakeError, | |
| ) | |
| from .client_reqrep import ( | |
| ClientRequest as ClientRequest, | |
| ClientResponse as ClientResponse, | |
| Fingerprint as Fingerprint, | |
| RequestInfo as RequestInfo, | |
| _merge_ssl_params, | |
| ) | |
| from .client_ws import ClientWebSocketResponse as ClientWebSocketResponse | |
| from .connector import ( | |
| BaseConnector as BaseConnector, | |
| NamedPipeConnector as NamedPipeConnector, | |
| TCPConnector as TCPConnector, | |
| UnixConnector as UnixConnector, | |
| ) | |
| from .cookiejar import CookieJar | |
| from .helpers import ( | |
| DEBUG, | |
| PY_36, | |
| BasicAuth, | |
| CeilTimeout, | |
| TimeoutHandle, | |
| get_running_loop, | |
| proxies_from_env, | |
| sentinel, | |
| strip_auth_from_url, | |
| ) | |
| from .http import WS_KEY, HttpVersion, WebSocketReader, WebSocketWriter | |
| from .http_websocket import WSHandshakeError, WSMessage, ws_ext_gen, ws_ext_parse | |
| from .streams import FlowControlDataQueue | |
| from .tracing import Trace, TraceConfig | |
| from .typedefs import JSONEncoder, LooseCookies, LooseHeaders, StrOrURL | |
| __all__ = ( | |
| # client_exceptions | |
| "ClientConnectionError", | |
| "ClientConnectorCertificateError", | |
| "ClientConnectorError", | |
| "ClientConnectorSSLError", | |
| "ClientError", | |
| "ClientHttpProxyError", | |
| "ClientOSError", | |
| "ClientPayloadError", | |
| "ClientProxyConnectionError", | |
| "ClientResponseError", | |
| "ClientSSLError", | |
| "ContentTypeError", | |
| "InvalidURL", | |
| "ServerConnectionError", | |
| "ServerDisconnectedError", | |
| "ServerFingerprintMismatch", | |
| "ServerTimeoutError", | |
| "TooManyRedirects", | |
| "WSServerHandshakeError", | |
| # client_reqrep | |
| "ClientRequest", | |
| "ClientResponse", | |
| "Fingerprint", | |
| "RequestInfo", | |
| # connector | |
| "BaseConnector", | |
| "TCPConnector", | |
| "UnixConnector", | |
| "NamedPipeConnector", | |
| # client_ws | |
| "ClientWebSocketResponse", | |
| # client | |
| "ClientSession", | |
| "ClientTimeout", | |
| "request", | |
| ) | |
| try: | |
| from ssl import SSLContext | |
| except ImportError: # pragma: no cover | |
| SSLContext = object # type: ignore | |
| @attr.s(auto_attribs=True, frozen=True, slots=True) | |
| class ClientTimeout: | |
| total: Optional[float] = None | |
| connect: Optional[float] = None | |
| sock_read: Optional[float] = None | |
| sock_connect: Optional[float] = None | |
| # pool_queue_timeout: Optional[float] = None | |
| # dns_resolution_timeout: Optional[float] = None | |
| # socket_connect_timeout: Optional[float] = None | |
| # connection_acquiring_timeout: Optional[float] = None | |
| # new_connection_timeout: Optional[float] = None | |
| # http_header_timeout: Optional[float] = None | |
| # response_body_timeout: Optional[float] = None | |
| # to create a timeout specific for a single request, either | |
| # - create a completely new one to overwrite the default | |
| # - or use http://www.attrs.org/en/stable/api.html#attr.evolve | |
| # to overwrite the defaults | |
| # 5 Minute default read timeout | |
| DEFAULT_TIMEOUT = ClientTimeout(total=5 * 60) | |
| _RetType = TypeVar("_RetType") | |
| class ClientSession: | |
| """First-class interface for making HTTP requests.""" | |
| ATTRS = frozenset( | |
| [ | |
| "_source_traceback", | |
| "_connector", | |
| "requote_redirect_url", | |
| "_loop", | |
| "_cookie_jar", | |
| "_connector_owner", | |
| "_default_auth", | |
| "_version", | |
| "_json_serialize", | |
| "_requote_redirect_url", | |
| "_timeout", | |
| "_raise_for_status", | |
| "_auto_decompress", | |
| "_trust_env", | |
| "_default_headers", | |
| "_skip_auto_headers", | |
| "_request_class", | |
| "_response_class", | |
| "_ws_response_class", | |
| "_trace_configs", | |
| "_read_bufsize", | |
| ] | |
| ) | |
| _source_traceback = None | |
| def __init__( | |
| self, | |
| *, | |
| connector: Optional[BaseConnector] = None, | |
| loop: Optional[asyncio.AbstractEventLoop] = None, | |
| cookies: Optional[LooseCookies] = None, | |
| headers: Optional[LooseHeaders] = None, | |
| skip_auto_headers: Optional[Iterable[str]] = None, | |
| auth: Optional[BasicAuth] = None, | |
| json_serialize: JSONEncoder = json.dumps, | |
| request_class: Type[ClientRequest] = ClientRequest, | |
| response_class: Type[ClientResponse] = ClientResponse, | |
| ws_response_class: Type[ClientWebSocketResponse] = ClientWebSocketResponse, | |
| version: HttpVersion = http.HttpVersion11, | |
| cookie_jar: Optional[AbstractCookieJar] = None, | |
| connector_owner: bool = True, | |
| raise_for_status: bool = False, | |
| read_timeout: Union[float, object] = sentinel, | |
| conn_timeout: Optional[float] = None, | |
| timeout: Union[object, ClientTimeout] = sentinel, | |
| auto_decompress: bool = True, | |
| trust_env: bool = False, | |
| requote_redirect_url: bool = True, | |
| trace_configs: Optional[List[TraceConfig]] = None, | |
| read_bufsize: int = 2 ** 16, | |
| ) -> None: | |
| if loop is None: | |
| if connector is not None: | |
| loop = connector._loop | |
| loop = get_running_loop(loop) | |
| if connector is None: | |
| connector = TCPConnector(loop=loop) | |
| if connector._loop is not loop: | |
| raise RuntimeError("Session and connector has to use same event loop") | |
| self._loop = loop | |
| if loop.get_debug(): | |
| self._source_traceback = traceback.extract_stack(sys._getframe(1)) | |
| if cookie_jar is None: | |
| cookie_jar = CookieJar(loop=loop) | |
| self._cookie_jar = cookie_jar | |
| if cookies is not None: | |
| self._cookie_jar.update_cookies(cookies) | |
| self._connector = connector # type: Optional[BaseConnector] | |
| self._connector_owner = connector_owner | |
| self._default_auth = auth | |
| self._version = version | |
| self._json_serialize = json_serialize | |
| if timeout is sentinel: | |
| self._timeout = DEFAULT_TIMEOUT | |
| if read_timeout is not sentinel: | |
| warnings.warn( | |
| "read_timeout is deprecated, " "use timeout argument instead", | |
| DeprecationWarning, | |
| stacklevel=2, | |
| ) | |
| self._timeout = attr.evolve(self._timeout, total=read_timeout) | |
| if conn_timeout is not None: | |
| self._timeout = attr.evolve(self._timeout, connect=conn_timeout) | |
| warnings.warn( | |
| "conn_timeout is deprecated, " "use timeout argument instead", | |
| DeprecationWarning, | |
| stacklevel=2, | |
| ) | |
| else: | |
| self._timeout = timeout # type: ignore | |
| if read_timeout is not sentinel: | |
| raise ValueError( | |
| "read_timeout and timeout parameters " | |
| "conflict, please setup " | |
| "timeout.read" | |
| ) | |
| if conn_timeout is not None: | |
| raise ValueError( | |
| "conn_timeout and timeout parameters " | |
| "conflict, please setup " | |
| "timeout.connect" | |
| ) | |
| self._raise_for_status = raise_for_status | |
| self._auto_decompress = auto_decompress | |
| self._trust_env = trust_env | |
| self._requote_redirect_url = requote_redirect_url | |
| self._read_bufsize = read_bufsize | |
| # Convert to list of tuples | |
| if headers: | |
| real_headers = CIMultiDict(headers) # type: CIMultiDict[str] | |
| else: | |
| real_headers = CIMultiDict() | |
| self._default_headers = real_headers # type: CIMultiDict[str] | |
| if skip_auto_headers is not None: | |
| self._skip_auto_headers = frozenset([istr(i) for i in skip_auto_headers]) | |
| else: | |
| self._skip_auto_headers = frozenset() | |
| self._request_class = request_class | |
| self._response_class = response_class | |
| self._ws_response_class = ws_response_class | |
| self._trace_configs = trace_configs or [] | |
| for trace_config in self._trace_configs: | |
| trace_config.freeze() | |
| def __init_subclass__(cls: Type["ClientSession"]) -> None: | |
| warnings.warn( | |
| "Inheritance class {} from ClientSession " | |
| "is discouraged".format(cls.__name__), | |
| DeprecationWarning, | |
| stacklevel=2, | |
| ) | |
| if DEBUG: | |
| def __setattr__(self, name: str, val: Any) -> None: | |
| if name not in self.ATTRS: | |
| warnings.warn( | |
| "Setting custom ClientSession.{} attribute " | |
| "is discouraged".format(name), | |
| DeprecationWarning, | |
| stacklevel=2, | |
| ) | |
| super().__setattr__(name, val) | |
| def __del__(self, _warnings: Any = warnings) -> None: | |
| if not self.closed: | |
| if PY_36: | |
| kwargs = {"source": self} | |
| else: | |
| kwargs = {} | |
| _warnings.warn( | |
| f"Unclosed client session {self!r}", ResourceWarning, **kwargs | |
| ) | |
| context = {"client_session": self, "message": "Unclosed client session"} | |
| if self._source_traceback is not None: | |
| context["source_traceback"] = self._source_traceback | |
| self._loop.call_exception_handler(context) | |
| def request( | |
| self, method: str, url: StrOrURL, **kwargs: Any | |
| ) -> "_RequestContextManager": | |
| """Perform HTTP request.""" | |
| return _RequestContextManager(self._request(method, url, **kwargs)) | |
| async def _request( | |
| self, | |
| method: str, | |
| str_or_url: StrOrURL, | |
| *, | |
| params: Optional[Mapping[str, str]] = None, | |
| data: Any = None, | |
| json: Any = None, | |
| cookies: Optional[LooseCookies] = None, | |
| headers: Optional[LooseHeaders] = None, | |
| skip_auto_headers: Optional[Iterable[str]] = None, | |
| auth: Optional[BasicAuth] = None, | |
| allow_redirects: bool = True, | |
| max_redirects: int = 10, | |
| compress: Optional[str] = None, | |
| chunked: Optional[bool] = None, | |
| expect100: bool = False, | |
| raise_for_status: Optional[bool] = None, | |
| read_until_eof: bool = True, | |
| proxy: Optional[StrOrURL] = None, | |
| proxy_auth: Optional[BasicAuth] = None, | |
| timeout: Union[ClientTimeout, object] = sentinel, | |
| verify_ssl: Optional[bool] = None, | |
| fingerprint: Optional[bytes] = None, | |
| ssl_context: Optional[SSLContext] = None, | |
| ssl: Optional[Union[SSLContext, bool, Fingerprint]] = None, | |
| proxy_headers: Optional[LooseHeaders] = None, | |
| trace_request_ctx: Optional[SimpleNamespace] = None, | |
| read_bufsize: Optional[int] = None, | |
| ) -> ClientResponse: | |
| # NOTE: timeout clamps existing connect and read timeouts. We cannot | |
| # set the default to None because we need to detect if the user wants | |
| # to use the existing timeouts by setting timeout to None. | |
| if self.closed: | |
| raise RuntimeError("Session is closed") | |
| ssl = _merge_ssl_params(ssl, verify_ssl, ssl_context, fingerprint) | |
| if data is not None and json is not None: | |
| raise ValueError( | |
| "data and json parameters can not be used at the same time" | |
| ) | |
| elif json is not None: | |
| data = payload.JsonPayload(json, dumps=self._json_serialize) | |
| if not isinstance(chunked, bool) and chunked is not None: | |
| warnings.warn("Chunk size is deprecated #1615", DeprecationWarning) | |
| redirects = 0 | |
| history = [] | |
| version = self._version | |
| # Merge with default headers and transform to CIMultiDict | |
| headers = self._prepare_headers(headers) | |
| proxy_headers = self._prepare_headers(proxy_headers) | |
| try: | |
| url = URL(str_or_url) | |
| except ValueError as e: | |
| raise InvalidURL(str_or_url) from e | |
| skip_headers = set(self._skip_auto_headers) | |
| if skip_auto_headers is not None: | |
| for i in skip_auto_headers: | |
| skip_headers.add(istr(i)) | |
| if proxy is not None: | |
| try: | |
| proxy = URL(proxy) | |
| except ValueError as e: | |
| raise InvalidURL(proxy) from e | |
| if timeout is sentinel: | |
| real_timeout = self._timeout # type: ClientTimeout | |
| else: | |
| if not isinstance(timeout, ClientTimeout): | |
| real_timeout = ClientTimeout(total=timeout) # type: ignore | |
| else: | |
| real_timeout = timeout | |
| # timeout is cumulative for all request operations | |
| # (request, redirects, responses, data consuming) | |
| tm = TimeoutHandle(self._loop, real_timeout.total) | |
| handle = tm.start() | |
| if read_bufsize is None: | |
| read_bufsize = self._read_bufsize | |
| traces = [ | |
| Trace( | |
| self, | |
| trace_config, | |
| trace_config.trace_config_ctx(trace_request_ctx=trace_request_ctx), | |
| ) | |
| for trace_config in self._trace_configs | |
| ] | |
| for trace in traces: | |
| await trace.send_request_start(method, url, headers) | |
| timer = tm.timer() | |
| try: | |
| with timer: | |
| while True: | |
| url, auth_from_url = strip_auth_from_url(url) | |
| if auth and auth_from_url: | |
| raise ValueError( | |
| "Cannot combine AUTH argument with " | |
| "credentials encoded in URL" | |
| ) | |
| if auth is None: | |
| auth = auth_from_url | |
| if auth is None: | |
| auth = self._default_auth | |
| # It would be confusing if we support explicit | |
| # Authorization header with auth argument | |
| if ( | |
| headers is not None | |
| and auth is not None | |
| and hdrs.AUTHORIZATION in headers | |
| ): | |
| raise ValueError( | |
| "Cannot combine AUTHORIZATION header " | |
| "with AUTH argument or credentials " | |
| "encoded in URL" | |
| ) | |
| all_cookies = self._cookie_jar.filter_cookies(url) | |
| if cookies is not None: | |
| tmp_cookie_jar = CookieJar() | |
| tmp_cookie_jar.update_cookies(cookies) | |
| req_cookies = tmp_cookie_jar.filter_cookies(url) | |
| if req_cookies: | |
| all_cookies.load(req_cookies) | |
| if proxy is not None: | |
| proxy = URL(proxy) | |
| elif self._trust_env: | |
| if not proxy_bypass(str(url.host)): | |
| for scheme, proxy_info in proxies_from_env().items(): | |
| if scheme == url.scheme: | |
| proxy = proxy_info.proxy | |
| proxy_auth = proxy_info.proxy_auth | |
| break | |
| req = self._request_class( | |
| method, | |
| url, | |
| params=params, | |
| headers=headers, | |
| skip_auto_headers=skip_headers, | |
| data=data, | |
| cookies=all_cookies, | |
| auth=auth, | |
| version=version, | |
| compress=compress, | |
| chunked=chunked, | |
| expect100=expect100, | |
| loop=self._loop, | |
| response_class=self._response_class, | |
| proxy=proxy, | |
| proxy_auth=proxy_auth, | |
| timer=timer, | |
| session=self, | |
| ssl=ssl, | |
| proxy_headers=proxy_headers, | |
| traces=traces, | |
| ) | |
| # connection timeout | |
| try: | |
| with CeilTimeout(real_timeout.connect, loop=self._loop): | |
| assert self._connector is not None | |
| conn = await self._connector.connect( | |
| req, traces=traces, timeout=real_timeout | |
| ) | |
| except asyncio.TimeoutError as exc: | |
| raise ServerTimeoutError( | |
| "Connection timeout " "to host {}".format(url) | |
| ) from exc | |
| assert conn.transport is not None | |
| assert conn.protocol is not None | |
| conn.protocol.set_response_params( | |
| timer=timer, | |
| skip_payload=method.upper() == "HEAD", | |
| read_until_eof=read_until_eof, | |
| auto_decompress=self._auto_decompress, | |
| read_timeout=real_timeout.sock_read, | |
| read_bufsize=read_bufsize, | |
| ) | |
| try: | |
| try: | |
| resp = await req.send(conn) | |
| try: | |
| await resp.start(conn) | |
| except BaseException: | |
| resp.close() | |
| raise | |
| except BaseException: | |
| conn.close() | |
| raise | |
| except ClientError: | |
| raise | |
| except OSError as exc: | |
| raise ClientOSError(*exc.args) from exc | |
| self._cookie_jar.update_cookies(resp.cookies, resp.url) | |
| # redirects | |
| if resp.status in (301, 302, 303, 307, 308) and allow_redirects: | |
| for trace in traces: | |
| await trace.send_request_redirect( | |
| method, url, headers, resp | |
| ) | |
| redirects += 1 | |
| history.append(resp) | |
| if max_redirects and redirects >= max_redirects: | |
| resp.close() | |
| raise TooManyRedirects( | |
| history[0].request_info, tuple(history) | |
| ) | |
| # For 301 and 302, mimic IE, now changed in RFC | |
| # https://github.com/kennethreitz/requests/pull/269 | |
| if (resp.status == 303 and resp.method != hdrs.METH_HEAD) or ( | |
| resp.status in (301, 302) and resp.method == hdrs.METH_POST | |
| ): | |
| method = hdrs.METH_GET | |
| data = None | |
| if headers.get(hdrs.CONTENT_LENGTH): | |
| headers.pop(hdrs.CONTENT_LENGTH) | |
| r_url = resp.headers.get(hdrs.LOCATION) or resp.headers.get( | |
| hdrs.URI | |
| ) | |
| if r_url is None: | |
| # see github.com/aio-libs/aiohttp/issues/2022 | |
| break | |
| else: | |
| # reading from correct redirection | |
| # response is forbidden | |
| resp.release() | |
| try: | |
| parsed_url = URL( | |
| r_url, encoded=not self._requote_redirect_url | |
| ) | |
| except ValueError as e: | |
| raise InvalidURL(r_url) from e | |
| scheme = parsed_url.scheme | |
| if scheme not in ("http", "https", ""): | |
| resp.close() | |
| raise ValueError("Can redirect only to http or https") | |
| elif not scheme: | |
| parsed_url = url.join(parsed_url) | |
| if url.origin() != parsed_url.origin(): | |
| auth = None | |
| headers.pop(hdrs.AUTHORIZATION, None) | |
| url = parsed_url | |
| params = None | |
| resp.release() | |
| continue | |
| break | |
| # check response status | |
| if raise_for_status is None: | |
| raise_for_status = self._raise_for_status | |
| if raise_for_status: | |
| resp.raise_for_status() | |
| # register connection | |
| if handle is not None: | |
| if resp.connection is not None: | |
| resp.connection.add_callback(handle.cancel) | |
| else: | |
| handle.cancel() | |
| resp._history = tuple(history) | |
| for trace in traces: | |
| await trace.send_request_end(method, url, headers, resp) | |
| return resp | |
| except BaseException as e: | |
| # cleanup timer | |
| tm.close() | |
| if handle: | |
| handle.cancel() | |
| handle = None | |
| for trace in traces: | |
| await trace.send_request_exception(method, url, headers, e) | |
| raise | |
| def ws_connect( | |
| self, | |
| url: StrOrURL, | |
| *, | |
| method: str = hdrs.METH_GET, | |
| protocols: Iterable[str] = (), | |
| timeout: float = 10.0, | |
| receive_timeout: Optional[float] = None, | |
| autoclose: bool = True, | |
| autoping: bool = True, | |
| heartbeat: Optional[float] = None, | |
| auth: Optional[BasicAuth] = None, | |
| origin: Optional[str] = None, | |
| headers: Optional[LooseHeaders] = None, | |
| proxy: Optional[StrOrURL] = None, | |
| proxy_auth: Optional[BasicAuth] = None, | |
| ssl: Union[SSLContext, bool, None, Fingerprint] = None, | |
| verify_ssl: Optional[bool] = None, | |
| fingerprint: Optional[bytes] = None, | |
| ssl_context: Optional[SSLContext] = None, | |
| proxy_headers: Optional[LooseHeaders] = None, | |
| compress: int = 0, | |
| max_msg_size: int = 4 * 1024 * 1024, | |
| ) -> "_WSRequestContextManager": | |
| """Initiate websocket connection.""" | |
| return _WSRequestContextManager( | |
| self._ws_connect( | |
| url, | |
| method=method, | |
| protocols=protocols, | |
| timeout=timeout, | |
| receive_timeout=receive_timeout, | |
| autoclose=autoclose, | |
| autoping=autoping, | |
| heartbeat=heartbeat, | |
| auth=auth, | |
| origin=origin, | |
| headers=headers, | |
| proxy=proxy, | |
| proxy_auth=proxy_auth, | |
| ssl=ssl, | |
| verify_ssl=verify_ssl, | |
| fingerprint=fingerprint, | |
| ssl_context=ssl_context, | |
| proxy_headers=proxy_headers, | |
| compress=compress, | |
| max_msg_size=max_msg_size, | |
| ) | |
| ) | |
| async def _ws_connect( | |
| self, | |
| url: StrOrURL, | |
| *, | |
| method: str = hdrs.METH_GET, | |
| protocols: Iterable[str] = (), | |
| timeout: float = 10.0, | |
| receive_timeout: Optional[float] = None, | |
| autoclose: bool = True, | |
| autoping: bool = True, | |
| heartbeat: Optional[float] = None, | |
| auth: Optional[BasicAuth] = None, | |
| origin: Optional[str] = None, | |
| headers: Optional[LooseHeaders] = None, | |
| proxy: Optional[StrOrURL] = None, | |
| proxy_auth: Optional[BasicAuth] = None, | |
| ssl: Union[SSLContext, bool, None, Fingerprint] = None, | |
| verify_ssl: Optional[bool] = None, | |
| fingerprint: Optional[bytes] = None, | |
| ssl_context: Optional[SSLContext] = None, | |
| proxy_headers: Optional[LooseHeaders] = None, | |
| compress: int = 0, | |
| max_msg_size: int = 4 * 1024 * 1024, | |
| ) -> ClientWebSocketResponse: | |
| if headers is None: | |
| real_headers = CIMultiDict() # type: CIMultiDict[str] | |
| else: | |
| real_headers = CIMultiDict(headers) | |
| default_headers = { | |
| hdrs.UPGRADE: "websocket", | |
| hdrs.CONNECTION: "upgrade", | |
| hdrs.SEC_WEBSOCKET_VERSION: "13", | |
| } | |
| for key, value in default_headers.items(): | |
| real_headers.setdefault(key, value) | |
| sec_key = base64.b64encode(os.urandom(16)) | |
| real_headers[hdrs.SEC_WEBSOCKET_KEY] = sec_key.decode() | |
| if protocols: | |
| real_headers[hdrs.SEC_WEBSOCKET_PROTOCOL] = ",".join(protocols) | |
| if origin is not None: | |
| real_headers[hdrs.ORIGIN] = origin | |
| if compress: | |
| extstr = ws_ext_gen(compress=compress) | |
| real_headers[hdrs.SEC_WEBSOCKET_EXTENSIONS] = extstr | |
| ssl = _merge_ssl_params(ssl, verify_ssl, ssl_context, fingerprint) | |
| # send request | |
| resp = await self.request( | |
| method, | |
| url, | |
| headers=real_headers, | |
| read_until_eof=False, | |
| auth=auth, | |
| proxy=proxy, | |
| proxy_auth=proxy_auth, | |
| ssl=ssl, | |
| proxy_headers=proxy_headers, | |
| ) | |
| try: | |
| # check handshake | |
| if resp.status != 101: | |
| raise WSServerHandshakeError( | |
| resp.request_info, | |
| resp.history, | |
| message="Invalid response status", | |
| status=resp.status, | |
| headers=resp.headers, | |
| ) | |
| if resp.headers.get(hdrs.UPGRADE, "").lower() != "websocket": | |
| raise WSServerHandshakeError( | |
| resp.request_info, | |
| resp.history, | |
| message="Invalid upgrade header", | |
| status=resp.status, | |
| headers=resp.headers, | |
| ) | |
| if resp.headers.get(hdrs.CONNECTION, "").lower() != "upgrade": | |
| raise WSServerHandshakeError( | |
| resp.request_info, | |
| resp.history, | |
| message="Invalid connection header", | |
| status=resp.status, | |
| headers=resp.headers, | |
| ) | |
| # key calculation | |
| r_key = resp.headers.get(hdrs.SEC_WEBSOCKET_ACCEPT, "") | |
| match = base64.b64encode(hashlib.sha1(sec_key + WS_KEY).digest()).decode() | |
| if r_key != match: | |
| raise WSServerHandshakeError( | |
| resp.request_info, | |
| resp.history, | |
| message="Invalid challenge response", | |
| status=resp.status, | |
| headers=resp.headers, | |
| ) | |
| # websocket protocol | |
| protocol = None | |
| if protocols and hdrs.SEC_WEBSOCKET_PROTOCOL in resp.headers: | |
| resp_protocols = [ | |
| proto.strip() | |
| for proto in resp.headers[hdrs.SEC_WEBSOCKET_PROTOCOL].split(",") | |
| ] | |
| for proto in resp_protocols: | |
| if proto in protocols: | |
| protocol = proto | |
| break | |
| # websocket compress | |
| notakeover = False | |
| if compress: | |
| compress_hdrs = resp.headers.get(hdrs.SEC_WEBSOCKET_EXTENSIONS) | |
| if compress_hdrs: | |
| try: | |
| compress, notakeover = ws_ext_parse(compress_hdrs) | |
| except WSHandshakeError as exc: | |
| raise WSServerHandshakeError( | |
| resp.request_info, | |
| resp.history, | |
| message=exc.args[0], | |
| status=resp.status, | |
| headers=resp.headers, | |
| ) from exc | |
| else: | |
| compress = 0 | |
| notakeover = False | |
| conn = resp.connection | |
| assert conn is not None | |
| conn_proto = conn.protocol | |
| assert conn_proto is not None | |
| transport = conn.transport | |
| assert transport is not None | |
| reader = FlowControlDataQueue( | |
| conn_proto, 2 ** 16, loop=self._loop | |
| ) # type: FlowControlDataQueue[WSMessage] | |
| conn_proto.set_parser(WebSocketReader(reader, max_msg_size), reader) | |
| writer = WebSocketWriter( | |
| conn_proto, | |
| transport, | |
| use_mask=True, | |
| compress=compress, | |
| notakeover=notakeover, | |
| ) | |
| except BaseException: | |
| resp.close() | |
| raise | |
| else: | |
| return self._ws_response_class( | |
| reader, | |
| writer, | |
| protocol, | |
| resp, | |
| timeout, | |
| autoclose, | |
| autoping, | |
| self._loop, | |
| receive_timeout=receive_timeout, | |
| heartbeat=heartbeat, | |
| compress=compress, | |
| client_notakeover=notakeover, | |
| ) | |
| def _prepare_headers(self, headers: Optional[LooseHeaders]) -> "CIMultiDict[str]": | |
| """Add default headers and transform it to CIMultiDict""" | |
| # Convert headers to MultiDict | |
| result = CIMultiDict(self._default_headers) | |
| if headers: | |
| if not isinstance(headers, (MultiDictProxy, MultiDict)): | |
| headers = CIMultiDict(headers) | |
| added_names = set() # type: Set[str] | |
| for key, value in headers.items(): | |
| if key in added_names: | |
| result.add(key, value) | |
| else: | |
| result[key] = value | |
| added_names.add(key) | |
| return result | |
| def get( | |
| self, url: StrOrURL, *, allow_redirects: bool = True, **kwargs: Any | |
| ) -> "_RequestContextManager": | |
| """Perform HTTP GET request.""" | |
| return _RequestContextManager( | |
| self._request(hdrs.METH_GET, url, allow_redirects=allow_redirects, **kwargs) | |
| ) | |
| def options( | |
| self, url: StrOrURL, *, allow_redirects: bool = True, **kwargs: Any | |
| ) -> "_RequestContextManager": | |
| """Perform HTTP OPTIONS request.""" | |
| return _RequestContextManager( | |
| self._request( | |
| hdrs.METH_OPTIONS, url, allow_redirects=allow_redirects, **kwargs | |
| ) | |
| ) | |
| def head( | |
| self, url: StrOrURL, *, allow_redirects: bool = False, **kwargs: Any | |
| ) -> "_RequestContextManager": | |
| """Perform HTTP HEAD request.""" | |
| return _RequestContextManager( | |
| self._request( | |
| hdrs.METH_HEAD, url, allow_redirects=allow_redirects, **kwargs | |
| ) | |
| ) | |
| def post( | |
| self, url: StrOrURL, *, data: Any = None, **kwargs: Any | |
| ) -> "_RequestContextManager": | |
| """Perform HTTP POST request.""" | |
| return _RequestContextManager( | |
| self._request(hdrs.METH_POST, url, data=data, **kwargs) | |
| ) | |
| def put( | |
| self, url: StrOrURL, *, data: Any = None, **kwargs: Any | |
| ) -> "_RequestContextManager": | |
| """Perform HTTP PUT request.""" | |
| return _RequestContextManager( | |
| self._request(hdrs.METH_PUT, url, data=data, **kwargs) | |
| ) | |
| def patch( | |
| self, url: StrOrURL, *, data: Any = None, **kwargs: Any | |
| ) -> "_RequestContextManager": | |
| """Perform HTTP PATCH request.""" | |
| return _RequestContextManager( | |
| self._request(hdrs.METH_PATCH, url, data=data, **kwargs) | |
| ) | |
| def delete(self, url: StrOrURL, **kwargs: Any) -> "_RequestContextManager": | |
| """Perform HTTP DELETE request.""" | |
| return _RequestContextManager(self._request(hdrs.METH_DELETE, url, **kwargs)) | |
| async def close(self) -> None: | |
| """Close underlying connector. | |
| Release all acquired resources. | |
| """ | |
| if not self.closed: | |
| if self._connector is not None and self._connector_owner: | |
| await self._connector.close() | |
| self._connector = None | |
| @property | |
| def closed(self) -> bool: | |
| """Is client session closed. | |
| A readonly property. | |
| """ | |
| return self._connector is None or self._connector.closed | |
| @property | |
| def connector(self) -> Optional[BaseConnector]: | |
| """Connector instance used for the session.""" | |
| return self._connector | |
| @property | |
| def cookie_jar(self) -> AbstractCookieJar: | |
| """The session cookies.""" | |
| return self._cookie_jar | |
| @property | |
| def version(self) -> Tuple[int, int]: | |
| """The session HTTP protocol version.""" | |
| return self._version | |
| @property | |
| def requote_redirect_url(self) -> bool: | |
| """Do URL requoting on redirection handling.""" | |
| return self._requote_redirect_url | |
| @requote_redirect_url.setter | |
| def requote_redirect_url(self, val: bool) -> None: | |
| """Do URL requoting on redirection handling.""" | |
| warnings.warn( | |
| "session.requote_redirect_url modification " "is deprecated #2778", | |
| DeprecationWarning, | |
| stacklevel=2, | |
| ) | |
| self._requote_redirect_url = val | |
| @property | |
| def loop(self) -> asyncio.AbstractEventLoop: | |
| """Session's loop.""" | |
| warnings.warn( | |
| "client.loop property is deprecated", DeprecationWarning, stacklevel=2 | |
| ) | |
| return self._loop | |
| @property | |
| def timeout(self) -> Union[object, ClientTimeout]: | |
| """Timeout for the session.""" | |
| return self._timeout | |
| @property | |
| def headers(self) -> "CIMultiDict[str]": | |
| """The default headers of the client session.""" | |
| return self._default_headers | |
| @property | |
| def skip_auto_headers(self) -> FrozenSet[istr]: | |
| """Headers for which autogeneration should be skipped""" | |
| return self._skip_auto_headers | |
| @property | |
| def auth(self) -> Optional[BasicAuth]: | |
| """An object that represents HTTP Basic Authorization""" | |
| return self._default_auth | |
| @property | |
| def json_serialize(self) -> JSONEncoder: | |
| """Json serializer callable""" | |
| return self._json_serialize | |
| @property | |
| def connector_owner(self) -> bool: | |
| """Should connector be closed on session closing""" | |
| return self._connector_owner | |
| @property | |
| def raise_for_status( | |
| self, | |
| ) -> Union[bool, Callable[[ClientResponse], Awaitable[None]]]: | |
| """ | |
| Should `ClientResponse.raise_for_status()` | |
| be called for each response | |
| """ | |
| return self._raise_for_status | |
| @property | |
| def auto_decompress(self) -> bool: | |
| """Should the body response be automatically decompressed""" | |
| return self._auto_decompress | |
| @property | |
| def trust_env(self) -> bool: | |
| """ | |
| Should get proxies information | |
| from HTTP_PROXY / HTTPS_PROXY environment variables | |
| or ~/.netrc file if present | |
| """ | |
| return self._trust_env | |
| @property | |
| def trace_configs(self) -> List[TraceConfig]: | |
| """A list of TraceConfig instances used for client tracing""" | |
| return self._trace_configs | |
| def detach(self) -> None: | |
| """Detach connector from session without closing the former. | |
| Session is switched to closed state anyway. | |
| """ | |
| self._connector = None | |
| def __enter__(self) -> None: | |
| raise TypeError("Use async with instead") | |
| def __exit__( | |
| self, | |
| exc_type: Optional[Type[BaseException]], | |
| exc_val: Optional[BaseException], | |
| exc_tb: Optional[TracebackType], | |
| ) -> None: | |
| # __exit__ should exist in pair with __enter__ but never executed | |
| pass # pragma: no cover | |
| async def __aenter__(self) -> "ClientSession": | |
| return self | |
| async def __aexit__( | |
| self, | |
| exc_type: Optional[Type[BaseException]], | |
| exc_val: Optional[BaseException], | |
| exc_tb: Optional[TracebackType], | |
| ) -> None: | |
| await self.close() | |
| class _BaseRequestContextManager(Coroutine[Any, Any, _RetType], Generic[_RetType]): | |
| __slots__ = ("_coro", "_resp") | |
| def __init__(self, coro: Coroutine["asyncio.Future[Any]", None, _RetType]) -> None: | |
| self._coro = coro | |
| def send(self, arg: None) -> "asyncio.Future[Any]": | |
| return self._coro.send(arg) | |
| def throw(self, arg: BaseException) -> None: # type: ignore | |
| self._coro.throw(arg) | |
| def close(self) -> None: | |
| return self._coro.close() | |
| def __await__(self) -> Generator[Any, None, _RetType]: | |
| ret = self._coro.__await__() | |
| return ret | |
| def __iter__(self) -> Generator[Any, None, _RetType]: | |
| return self.__await__() | |
| async def __aenter__(self) -> _RetType: | |
| self._resp = await self._coro | |
| return self._resp | |
| class _RequestContextManager(_BaseRequestContextManager[ClientResponse]): | |
| async def __aexit__( | |
| self, | |
| exc_type: Optional[Type[BaseException]], | |
| exc: Optional[BaseException], | |
| tb: Optional[TracebackType], | |
| ) -> None: | |
| # We're basing behavior on the exception as it can be caused by | |
| # user code unrelated to the status of the connection. If you | |
| # would like to close a connection you must do that | |
| # explicitly. Otherwise connection error handling should kick in | |
| # and close/recycle the connection as required. | |
| self._resp.release() | |
| class _WSRequestContextManager(_BaseRequestContextManager[ClientWebSocketResponse]): | |
| async def __aexit__( | |
| self, | |
| exc_type: Optional[Type[BaseException]], | |
| exc: Optional[BaseException], | |
| tb: Optional[TracebackType], | |
| ) -> None: | |
| await self._resp.close() | |
| class _SessionRequestContextManager: | |
| __slots__ = ("_coro", "_resp", "_session") | |
| def __init__( | |
| self, | |
| coro: Coroutine["asyncio.Future[Any]", None, ClientResponse], | |
| session: ClientSession, | |
| ) -> None: | |
| self._coro = coro | |
| self._resp = None # type: Optional[ClientResponse] | |
| self._session = session | |
| async def __aenter__(self) -> ClientResponse: | |
| try: | |
| self._resp = await self._coro | |
| except BaseException: | |
| await self._session.close() | |
| raise | |
| else: | |
| return self._resp | |
| async def __aexit__( | |
| self, | |
| exc_type: Optional[Type[BaseException]], | |
| exc: Optional[BaseException], | |
| tb: Optional[TracebackType], | |
| ) -> None: | |
| assert self._resp is not None | |
| self._resp.close() | |
| await self._session.close() | |
| def request( | |
| method: str, | |
| url: StrOrURL, | |
| *, | |
| params: Optional[Mapping[str, str]] = None, | |
| data: Any = None, | |
| json: Any = None, | |
| headers: Optional[LooseHeaders] = None, | |
| skip_auto_headers: Optional[Iterable[str]] = None, | |
| auth: Optional[BasicAuth] = None, | |
| allow_redirects: bool = True, | |
| max_redirects: int = 10, | |
| compress: Optional[str] = None, | |
| chunked: Optional[bool] = None, | |
| expect100: bool = False, | |
| raise_for_status: Optional[bool] = None, | |
| read_until_eof: bool = True, | |
| proxy: Optional[StrOrURL] = None, | |
| proxy_auth: Optional[BasicAuth] = None, | |
| timeout: Union[ClientTimeout, object] = sentinel, | |
| cookies: Optional[LooseCookies] = None, | |
| version: HttpVersion = http.HttpVersion11, | |
| connector: Optional[BaseConnector] = None, | |
| read_bufsize: Optional[int] = None, | |
| loop: Optional[asyncio.AbstractEventLoop] = None, | |
| ) -> _SessionRequestContextManager: | |
| """Constructs and sends a request. Returns response object. | |
| method - HTTP method | |
| url - request url | |
| params - (optional) Dictionary or bytes to be sent in the query | |
| string of the new request | |
| data - (optional) Dictionary, bytes, or file-like object to | |
| send in the body of the request | |
| json - (optional) Any json compatible python object | |
| headers - (optional) Dictionary of HTTP Headers to send with | |
| the request | |
| cookies - (optional) Dict object to send with the request | |
| auth - (optional) BasicAuth named tuple represent HTTP Basic Auth | |
| auth - aiohttp.helpers.BasicAuth | |
| allow_redirects - (optional) If set to False, do not follow | |
| redirects | |
| version - Request HTTP version. | |
| compress - Set to True if request has to be compressed | |
| with deflate encoding. | |
| chunked - Set to chunk size for chunked transfer encoding. | |
| expect100 - Expect 100-continue response from server. | |
| connector - BaseConnector sub-class instance to support | |
| connection pooling. | |
| read_until_eof - Read response until eof if response | |
| does not have Content-Length header. | |
| loop - Optional event loop. | |
| timeout - Optional ClientTimeout settings structure, 5min | |
| total timeout by default. | |
| Usage:: | |
| >>> import aiohttp | |
| >>> resp = await aiohttp.request('GET', 'http://python.org/') | |
| >>> resp | |
| <ClientResponse(python.org/) [200]> | |
| >>> data = await resp.read() | |
| """ | |
| connector_owner = False | |
| if connector is None: | |
| connector_owner = True | |
| connector = TCPConnector(loop=loop, force_close=True) | |
| session = ClientSession( | |
| loop=loop, | |
| cookies=cookies, | |
| version=version, | |
| timeout=timeout, | |
| connector=connector, | |
| connector_owner=connector_owner, | |
| ) | |
| return _SessionRequestContextManager( | |
| session._request( | |
| method, | |
| url, | |
| params=params, | |
| data=data, | |
| json=json, | |
| headers=headers, | |
| skip_auto_headers=skip_auto_headers, | |
| auth=auth, | |
| allow_redirects=allow_redirects, | |
| max_redirects=max_redirects, | |
| compress=compress, | |
| chunked=chunked, | |
| expect100=expect100, | |
| raise_for_status=raise_for_status, | |
| read_until_eof=read_until_eof, | |
| proxy=proxy, | |
| proxy_auth=proxy_auth, | |
| read_bufsize=read_bufsize, | |
| ), | |
| session, | |
| ) |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment