Merge branch 'feature/UpdateTornado' into dev

This commit is contained in:
JackDandy 2023-03-08 14:13:07 +00:00
commit 3f3dc75edc
23 changed files with 516 additions and 384 deletions

View file

@ -2,6 +2,7 @@
* Update html5lib 1.1 (f87487a) to 1.2-dev (3e500bb) * Update html5lib 1.1 (f87487a) to 1.2-dev (3e500bb)
* Update package resource API 63.2.0 (3ae44cd) to 67.5.1 (f51eccd) * Update package resource API 63.2.0 (3ae44cd) to 67.5.1 (f51eccd)
* Update Tornado Web Server 6.2.0 (a4f08a3) to 6.3.0 (7186b86)
* Update urllib3 1.26.13 (25fbd5f) to 1.26.14 (a06c05c) * Update urllib3 1.26.13 (25fbd5f) to 1.26.14 (a06c05c)
* Change remove calls to legacy py2 fix encoding function * Change remove calls to legacy py2 fix encoding function
* Change requirements for pure py3 * Change requirements for pure py3

View file

@ -22,5 +22,46 @@
# is zero for an official release, positive for a development branch, # is zero for an official release, positive for a development branch,
# or negative for a release candidate or beta (after the base version # or negative for a release candidate or beta (after the base version
# number has been incremented) # number has been incremented)
version = "6.2" version = "6.3.dev1"
version_info = (6, 2, 0, 0) version_info = (6, 3, 0, -100)
import importlib
import typing
__all__ = [
"auth",
"autoreload",
"concurrent",
"curl_httpclient",
"escape",
"gen",
"http1connection",
"httpclient",
"httpserver",
"httputil",
"ioloop",
"iostream",
"locale",
"locks",
"log",
"netutil",
"options",
"platform",
"process",
"queues",
"routing",
"simple_httpclient",
"tcpclient",
"tcpserver",
"template",
"testing",
"util",
"web",
]
# Copied from https://peps.python.org/pep-0562/
def __getattr__(name: str) -> typing.Any:
if name in __all__:
return importlib.import_module("." + name, __name__)
raise AttributeError(f"module {__name__!r} has no attribute {name!r}")

View file

@ -15,66 +15,66 @@
"""Data used by the tornado.locale module.""" """Data used by the tornado.locale module."""
LOCALE_NAMES = { LOCALE_NAMES = {
"af_ZA": {"name_en": u"Afrikaans", "name": u"Afrikaans"}, "af_ZA": {"name_en": "Afrikaans", "name": "Afrikaans"},
"am_ET": {"name_en": u"Amharic", "name": u"አማርኛ"}, "am_ET": {"name_en": "Amharic", "name": "አማርኛ"},
"ar_AR": {"name_en": u"Arabic", "name": u"العربية"}, "ar_AR": {"name_en": "Arabic", "name": "العربية"},
"bg_BG": {"name_en": u"Bulgarian", "name": u"Български"}, "bg_BG": {"name_en": "Bulgarian", "name": "Български"},
"bn_IN": {"name_en": u"Bengali", "name": u"বাংলা"}, "bn_IN": {"name_en": "Bengali", "name": "বাংলা"},
"bs_BA": {"name_en": u"Bosnian", "name": u"Bosanski"}, "bs_BA": {"name_en": "Bosnian", "name": "Bosanski"},
"ca_ES": {"name_en": u"Catalan", "name": u"Català"}, "ca_ES": {"name_en": "Catalan", "name": "Català"},
"cs_CZ": {"name_en": u"Czech", "name": u"Čeština"}, "cs_CZ": {"name_en": "Czech", "name": "Čeština"},
"cy_GB": {"name_en": u"Welsh", "name": u"Cymraeg"}, "cy_GB": {"name_en": "Welsh", "name": "Cymraeg"},
"da_DK": {"name_en": u"Danish", "name": u"Dansk"}, "da_DK": {"name_en": "Danish", "name": "Dansk"},
"de_DE": {"name_en": u"German", "name": u"Deutsch"}, "de_DE": {"name_en": "German", "name": "Deutsch"},
"el_GR": {"name_en": u"Greek", "name": u"Ελληνικά"}, "el_GR": {"name_en": "Greek", "name": "Ελληνικά"},
"en_GB": {"name_en": u"English (UK)", "name": u"English (UK)"}, "en_GB": {"name_en": "English (UK)", "name": "English (UK)"},
"en_US": {"name_en": u"English (US)", "name": u"English (US)"}, "en_US": {"name_en": "English (US)", "name": "English (US)"},
"es_ES": {"name_en": u"Spanish (Spain)", "name": u"Español (España)"}, "es_ES": {"name_en": "Spanish (Spain)", "name": "Español (España)"},
"es_LA": {"name_en": u"Spanish", "name": u"Español"}, "es_LA": {"name_en": "Spanish", "name": "Español"},
"et_EE": {"name_en": u"Estonian", "name": u"Eesti"}, "et_EE": {"name_en": "Estonian", "name": "Eesti"},
"eu_ES": {"name_en": u"Basque", "name": u"Euskara"}, "eu_ES": {"name_en": "Basque", "name": "Euskara"},
"fa_IR": {"name_en": u"Persian", "name": u"فارسی"}, "fa_IR": {"name_en": "Persian", "name": "فارسی"},
"fi_FI": {"name_en": u"Finnish", "name": u"Suomi"}, "fi_FI": {"name_en": "Finnish", "name": "Suomi"},
"fr_CA": {"name_en": u"French (Canada)", "name": u"Français (Canada)"}, "fr_CA": {"name_en": "French (Canada)", "name": "Français (Canada)"},
"fr_FR": {"name_en": u"French", "name": u"Français"}, "fr_FR": {"name_en": "French", "name": "Français"},
"ga_IE": {"name_en": u"Irish", "name": u"Gaeilge"}, "ga_IE": {"name_en": "Irish", "name": "Gaeilge"},
"gl_ES": {"name_en": u"Galician", "name": u"Galego"}, "gl_ES": {"name_en": "Galician", "name": "Galego"},
"he_IL": {"name_en": u"Hebrew", "name": u"עברית"}, "he_IL": {"name_en": "Hebrew", "name": "עברית"},
"hi_IN": {"name_en": u"Hindi", "name": u"हिन्दी"}, "hi_IN": {"name_en": "Hindi", "name": "हिन्दी"},
"hr_HR": {"name_en": u"Croatian", "name": u"Hrvatski"}, "hr_HR": {"name_en": "Croatian", "name": "Hrvatski"},
"hu_HU": {"name_en": u"Hungarian", "name": u"Magyar"}, "hu_HU": {"name_en": "Hungarian", "name": "Magyar"},
"id_ID": {"name_en": u"Indonesian", "name": u"Bahasa Indonesia"}, "id_ID": {"name_en": "Indonesian", "name": "Bahasa Indonesia"},
"is_IS": {"name_en": u"Icelandic", "name": u"Íslenska"}, "is_IS": {"name_en": "Icelandic", "name": "Íslenska"},
"it_IT": {"name_en": u"Italian", "name": u"Italiano"}, "it_IT": {"name_en": "Italian", "name": "Italiano"},
"ja_JP": {"name_en": u"Japanese", "name": u"日本語"}, "ja_JP": {"name_en": "Japanese", "name": "日本語"},
"ko_KR": {"name_en": u"Korean", "name": u"한국어"}, "ko_KR": {"name_en": "Korean", "name": "한국어"},
"lt_LT": {"name_en": u"Lithuanian", "name": u"Lietuvių"}, "lt_LT": {"name_en": "Lithuanian", "name": "Lietuvių"},
"lv_LV": {"name_en": u"Latvian", "name": u"Latviešu"}, "lv_LV": {"name_en": "Latvian", "name": "Latviešu"},
"mk_MK": {"name_en": u"Macedonian", "name": u"Македонски"}, "mk_MK": {"name_en": "Macedonian", "name": "Македонски"},
"ml_IN": {"name_en": u"Malayalam", "name": u"മലയാളം"}, "ml_IN": {"name_en": "Malayalam", "name": "മലയാളം"},
"ms_MY": {"name_en": u"Malay", "name": u"Bahasa Melayu"}, "ms_MY": {"name_en": "Malay", "name": "Bahasa Melayu"},
"nb_NO": {"name_en": u"Norwegian (bokmal)", "name": u"Norsk (bokmål)"}, "nb_NO": {"name_en": "Norwegian (bokmal)", "name": "Norsk (bokmål)"},
"nl_NL": {"name_en": u"Dutch", "name": u"Nederlands"}, "nl_NL": {"name_en": "Dutch", "name": "Nederlands"},
"nn_NO": {"name_en": u"Norwegian (nynorsk)", "name": u"Norsk (nynorsk)"}, "nn_NO": {"name_en": "Norwegian (nynorsk)", "name": "Norsk (nynorsk)"},
"pa_IN": {"name_en": u"Punjabi", "name": u"ਪੰਜਾਬੀ"}, "pa_IN": {"name_en": "Punjabi", "name": "ਪੰਜਾਬੀ"},
"pl_PL": {"name_en": u"Polish", "name": u"Polski"}, "pl_PL": {"name_en": "Polish", "name": "Polski"},
"pt_BR": {"name_en": u"Portuguese (Brazil)", "name": u"Português (Brasil)"}, "pt_BR": {"name_en": "Portuguese (Brazil)", "name": "Português (Brasil)"},
"pt_PT": {"name_en": u"Portuguese (Portugal)", "name": u"Português (Portugal)"}, "pt_PT": {"name_en": "Portuguese (Portugal)", "name": "Português (Portugal)"},
"ro_RO": {"name_en": u"Romanian", "name": u"Română"}, "ro_RO": {"name_en": "Romanian", "name": "Română"},
"ru_RU": {"name_en": u"Russian", "name": u"Русский"}, "ru_RU": {"name_en": "Russian", "name": "Русский"},
"sk_SK": {"name_en": u"Slovak", "name": u"Slovenčina"}, "sk_SK": {"name_en": "Slovak", "name": "Slovenčina"},
"sl_SI": {"name_en": u"Slovenian", "name": u"Slovenščina"}, "sl_SI": {"name_en": "Slovenian", "name": "Slovenščina"},
"sq_AL": {"name_en": u"Albanian", "name": u"Shqip"}, "sq_AL": {"name_en": "Albanian", "name": "Shqip"},
"sr_RS": {"name_en": u"Serbian", "name": u"Српски"}, "sr_RS": {"name_en": "Serbian", "name": "Српски"},
"sv_SE": {"name_en": u"Swedish", "name": u"Svenska"}, "sv_SE": {"name_en": "Swedish", "name": "Svenska"},
"sw_KE": {"name_en": u"Swahili", "name": u"Kiswahili"}, "sw_KE": {"name_en": "Swahili", "name": "Kiswahili"},
"ta_IN": {"name_en": u"Tamil", "name": u"தமிழ்"}, "ta_IN": {"name_en": "Tamil", "name": "தமிழ்"},
"te_IN": {"name_en": u"Telugu", "name": u"తెలుగు"}, "te_IN": {"name_en": "Telugu", "name": "తెలుగు"},
"th_TH": {"name_en": u"Thai", "name": u"ภาษาไทย"}, "th_TH": {"name_en": "Thai", "name": "ภาษาไทย"},
"tl_PH": {"name_en": u"Filipino", "name": u"Filipino"}, "tl_PH": {"name_en": "Filipino", "name": "Filipino"},
"tr_TR": {"name_en": u"Turkish", "name": u"Türkçe"}, "tr_TR": {"name_en": "Turkish", "name": "Türkçe"},
"uk_UA": {"name_en": u"Ukraini ", "name": u"Українська"}, "uk_UA": {"name_en": "Ukraini ", "name": "Українська"},
"vi_VN": {"name_en": u"Vietnamese", "name": u"Tiếng Việt"}, "vi_VN": {"name_en": "Vietnamese", "name": "Tiếng Việt"},
"zh_CN": {"name_en": u"Chinese (Simplified)", "name": u"中文(简体)"}, "zh_CN": {"name_en": "Chinese (Simplified)", "name": "中文(简体)"},
"zh_TW": {"name_en": u"Chinese (Traditional)", "name": u"中文(繁體)"}, "zh_TW": {"name_en": "Chinese (Traditional)", "name": "中文(繁體)"},
} }

View file

@ -42,7 +42,7 @@ Example usage for Google OAuth:
user = await self.get_authenticated_user( user = await self.get_authenticated_user(
redirect_uri='http://your.site.com/auth/google', redirect_uri='http://your.site.com/auth/google',
code=self.get_argument('code')) code=self.get_argument('code'))
# Save the user with e.g. set_secure_cookie # Save the user with e.g. set_signed_cookie
else: else:
self.authorize_redirect( self.authorize_redirect(
redirect_uri='http://your.site.com/auth/google', redirect_uri='http://your.site.com/auth/google',
@ -136,7 +136,7 @@ class OpenIdMixin(object):
args = dict( args = dict(
(k, v[-1]) for k, v in handler.request.arguments.items() (k, v[-1]) for k, v in handler.request.arguments.items()
) # type: Dict[str, Union[str, bytes]] ) # type: Dict[str, Union[str, bytes]]
args["openid.mode"] = u"check_authentication" args["openid.mode"] = "check_authentication"
url = self._OPENID_ENDPOINT # type: ignore url = self._OPENID_ENDPOINT # type: ignore
if http_client is None: if http_client is None:
http_client = self.get_auth_http_client() http_client = self.get_auth_http_client()
@ -211,14 +211,14 @@ class OpenIdMixin(object):
for key in handler.request.arguments: for key in handler.request.arguments:
if ( if (
key.startswith("openid.ns.") key.startswith("openid.ns.")
and handler.get_argument(key) == u"http://openid.net/srv/ax/1.0" and handler.get_argument(key) == "http://openid.net/srv/ax/1.0"
): ):
ax_ns = key[10:] ax_ns = key[10:]
break break
def get_ax_arg(uri: str) -> str: def get_ax_arg(uri: str) -> str:
if not ax_ns: if not ax_ns:
return u"" return ""
prefix = "openid." + ax_ns + ".type." prefix = "openid." + ax_ns + ".type."
ax_name = None ax_name = None
for name in handler.request.arguments.keys(): for name in handler.request.arguments.keys():
@ -227,8 +227,8 @@ class OpenIdMixin(object):
ax_name = "openid." + ax_ns + ".value." + part ax_name = "openid." + ax_ns + ".value." + part
break break
if not ax_name: if not ax_name:
return u"" return ""
return handler.get_argument(ax_name, u"") return handler.get_argument(ax_name, "")
email = get_ax_arg("http://axschema.org/contact/email") email = get_ax_arg("http://axschema.org/contact/email")
name = get_ax_arg("http://axschema.org/namePerson") name = get_ax_arg("http://axschema.org/namePerson")
@ -247,7 +247,7 @@ class OpenIdMixin(object):
if name: if name:
user["name"] = name user["name"] = name
elif name_parts: elif name_parts:
user["name"] = u" ".join(name_parts) user["name"] = " ".join(name_parts)
elif email: elif email:
user["name"] = email.split("@")[0] user["name"] = email.split("@")[0]
if email: if email:
@ -694,7 +694,7 @@ class TwitterMixin(OAuthMixin):
async def get(self): async def get(self):
if self.get_argument("oauth_token", None): if self.get_argument("oauth_token", None):
user = await self.get_authenticated_user() user = await self.get_authenticated_user()
# Save the user using e.g. set_secure_cookie() # Save the user using e.g. set_signed_cookie()
else: else:
await self.authorize_redirect() await self.authorize_redirect()
@ -855,8 +855,28 @@ class GoogleOAuth2Mixin(OAuth2Mixin):
_OAUTH_NO_CALLBACKS = False _OAUTH_NO_CALLBACKS = False
_OAUTH_SETTINGS_KEY = "google_oauth" _OAUTH_SETTINGS_KEY = "google_oauth"
def get_google_oauth_settings(self) -> Dict[str, str]:
"""Return the Google OAuth 2.0 credentials that you created with
[Google Cloud
Platform](https://console.cloud.google.com/apis/credentials). The dict
format is::
{
"key": "your_client_id", "secret": "your_client_secret"
}
If your credentials are stored differently (e.g. in a db) you can
override this method for custom provision.
"""
handler = cast(RequestHandler, self)
return handler.settings[self._OAUTH_SETTINGS_KEY]
async def get_authenticated_user( async def get_authenticated_user(
self, redirect_uri: str, code: str self,
redirect_uri: str,
code: str,
client_id: Optional[str] = None,
client_secret: Optional[str] = None,
) -> Dict[str, Any]: ) -> Dict[str, Any]:
"""Handles the login for the Google user, returning an access token. """Handles the login for the Google user, returning an access token.
@ -883,11 +903,11 @@ class GoogleOAuth2Mixin(OAuth2Mixin):
"https://www.googleapis.com/oauth2/v1/userinfo", "https://www.googleapis.com/oauth2/v1/userinfo",
access_token=access["access_token"]) access_token=access["access_token"])
# Save the user and access token with # Save the user and access token with
# e.g. set_secure_cookie. # e.g. set_signed_cookie.
else: else:
self.authorize_redirect( self.authorize_redirect(
redirect_uri='http://your.site.com/auth/google', redirect_uri='http://your.site.com/auth/google',
client_id=self.settings['google_oauth']['key'], client_id=self.get_google_oauth_settings()['key'],
scope=['profile', 'email'], scope=['profile', 'email'],
response_type='code', response_type='code',
extra_params={'approval_prompt': 'auto'}) extra_params={'approval_prompt': 'auto'})
@ -899,14 +919,20 @@ class GoogleOAuth2Mixin(OAuth2Mixin):
The ``callback`` argument was removed. Use the returned awaitable object instead. The ``callback`` argument was removed. Use the returned awaitable object instead.
""" # noqa: E501 """ # noqa: E501
handler = cast(RequestHandler, self)
if client_id is None or client_secret is None:
settings = self.get_google_oauth_settings()
if client_id is None:
client_id = settings["key"]
if client_secret is None:
client_secret = settings["secret"]
http = self.get_auth_http_client() http = self.get_auth_http_client()
body = urllib.parse.urlencode( body = urllib.parse.urlencode(
{ {
"redirect_uri": redirect_uri, "redirect_uri": redirect_uri,
"code": code, "code": code,
"client_id": handler.settings[self._OAUTH_SETTINGS_KEY]["key"], "client_id": client_id,
"client_secret": handler.settings[self._OAUTH_SETTINGS_KEY]["secret"], "client_secret": client_secret,
"grant_type": "authorization_code", "grant_type": "authorization_code",
} }
) )
@ -951,7 +977,7 @@ class FacebookGraphMixin(OAuth2Mixin):
client_id=self.settings["facebook_api_key"], client_id=self.settings["facebook_api_key"],
client_secret=self.settings["facebook_secret"], client_secret=self.settings["facebook_secret"],
code=self.get_argument("code")) code=self.get_argument("code"))
# Save the user with e.g. set_secure_cookie # Save the user with e.g. set_signed_cookie
else: else:
self.authorize_redirect( self.authorize_redirect(
redirect_uri='/auth/facebookgraph/', redirect_uri='/auth/facebookgraph/',

View file

@ -36,11 +36,11 @@ from tornado.httpclient import (
) )
from tornado.log import app_log from tornado.log import app_log
from typing import Dict, Any, Callable, Union, Tuple, Optional from typing import Dict, Any, Callable, Union, Optional
import typing import typing
if typing.TYPE_CHECKING: if typing.TYPE_CHECKING:
from typing import Deque # noqa: F401 from typing import Deque, Tuple # noqa: F401
curl_log = logging.getLogger("tornado.curl_httpclient") curl_log = logging.getLogger("tornado.curl_httpclient")

View file

@ -368,7 +368,7 @@ def linkify(
# have a status bar, such as Safari by default) # have a status bar, such as Safari by default)
params += ' title="%s"' % href params += ' title="%s"' % href
return u'<a href="%s"%s>%s</a>' % (href, params, url) return '<a href="%s"%s>%s</a>' % (href, params, url)
# First HTML-escape so that our strings are all safe. # First HTML-escape so that our strings are all safe.
# The regex is modified to avoid character entites other than &amp; so # The regex is modified to avoid character entites other than &amp; so

View file

@ -743,7 +743,7 @@ class Runner(object):
self.running = False self.running = False
self.finished = False self.finished = False
self.io_loop = IOLoop.current() self.io_loop = IOLoop.current()
if self.handle_yield(first_yielded): if self.ctx_run(self.handle_yield, first_yielded):
gen = result_future = first_yielded = None # type: ignore gen = result_future = first_yielded = None # type: ignore
self.ctx_run(self.run) self.ctx_run(self.run)
@ -763,21 +763,25 @@ class Runner(object):
return return
self.future = None self.future = None
try: try:
exc_info = None
try: try:
value = future.result() value = future.result()
except Exception: except Exception as e:
exc_info = sys.exc_info() # Save the exception for later. It's important that
# gen.throw() not be called inside this try/except block
# because that makes sys.exc_info behave unexpectedly.
exc: Optional[Exception] = e
else:
exc = None
finally:
future = None future = None
if exc_info is not None: if exc is not None:
try: try:
yielded = self.gen.throw(*exc_info) # type: ignore yielded = self.gen.throw(exc)
finally: finally:
# Break up a reference to itself # Break up a circular reference for faster GC on
# for faster GC on CPython. # CPython.
exc_info = None del exc
else: else:
yielded = self.gen.send(value) yielded = self.gen.send(value)

View file

@ -83,7 +83,7 @@ class IOLoop(Configurable):
import functools import functools
import socket import socket
import tornado.ioloop import tornado
from tornado.iostream import IOStream from tornado.iostream import IOStream
async def handle_connection(connection, address): async def handle_connection(connection, address):
@ -123,8 +123,7 @@ class IOLoop(Configurable):
and instead initialize the `asyncio` event loop and use `IOLoop.current()`. and instead initialize the `asyncio` event loop and use `IOLoop.current()`.
In some cases, such as in test frameworks when initializing an `IOLoop` In some cases, such as in test frameworks when initializing an `IOLoop`
to be run in a secondary thread, it may be appropriate to construct to be run in a secondary thread, it may be appropriate to construct
an `IOLoop` with ``IOLoop(make_current=False)``. Constructing an `IOLoop` an `IOLoop` with ``IOLoop(make_current=False)``.
without the ``make_current=False`` argument is deprecated since Tornado 6.2.
In general, an `IOLoop` cannot survive a fork or be shared across processes In general, an `IOLoop` cannot survive a fork or be shared across processes
in any way. When multiple processes are being used, each process should in any way. When multiple processes are being used, each process should
@ -145,12 +144,10 @@ class IOLoop(Configurable):
cannot be used on Python 3 except to redundantly specify the `asyncio` cannot be used on Python 3 except to redundantly specify the `asyncio`
event loop. event loop.
.. deprecated:: 6.2 .. versionchanged:: 6.3
It is deprecated to create an event loop that is "current" but not ``make_current=True`` is now the default when creating an IOLoop -
running. This means it is deprecated to pass previously the default was to make the event loop current if there wasn't
``make_current=True`` to the ``IOLoop`` constructor, or to create already a current one.
an ``IOLoop`` while no asyncio event loop is running unless
``make_current=False`` is used.
""" """
# These constants were originally based on constants from the epoll module. # These constants were originally based on constants from the epoll module.
@ -263,17 +260,20 @@ class IOLoop(Configurable):
""" """
try: try:
loop = asyncio.get_event_loop() loop = asyncio.get_event_loop()
except (RuntimeError, AssertionError): except RuntimeError:
if not instance: if not instance:
return None return None
raise # Create a new asyncio event loop for this thread.
loop = asyncio.new_event_loop()
asyncio.set_event_loop(loop)
try: try:
return IOLoop._ioloop_for_asyncio[loop] return IOLoop._ioloop_for_asyncio[loop]
except KeyError: except KeyError:
if instance: if instance:
from tornado.platform.asyncio import AsyncIOMainLoop from tornado.platform.asyncio import AsyncIOMainLoop
current = AsyncIOMainLoop(make_current=True) # type: Optional[IOLoop] current = AsyncIOMainLoop() # type: Optional[IOLoop]
else: else:
current = None current = None
return current return current
@ -295,12 +295,17 @@ class IOLoop(Configurable):
This method also sets the current `asyncio` event loop. This method also sets the current `asyncio` event loop.
.. deprecated:: 6.2 .. deprecated:: 6.2
The concept of an event loop that is "current" without Setting and clearing the current event loop through Tornado is
currently running is deprecated in asyncio since Python deprecated. Use ``asyncio.set_event_loop`` instead if you need this.
3.10. All related functionality in Tornado is also
deprecated. Instead, start the event loop with `asyncio.run`
before interacting with it.
""" """
warnings.warn(
"make_current is deprecated; start the event loop first",
DeprecationWarning,
stacklevel=2,
)
self._make_current()
def _make_current(self) -> None:
# The asyncio event loops override this method. # The asyncio event loops override this method.
raise NotImplementedError() raise NotImplementedError()
@ -344,16 +349,9 @@ class IOLoop(Configurable):
return AsyncIOLoop return AsyncIOLoop
def initialize(self, make_current: Optional[bool] = None) -> None: def initialize(self, make_current: bool = True) -> None:
if make_current is None: if make_current:
if IOLoop.current(instance=False) is None: self._make_current()
self.make_current()
elif make_current:
current = IOLoop.current(instance=False)
# AsyncIO loops can already be current by this point.
if current is not None and current is not self:
raise RuntimeError("current IOLoop already exists")
self.make_current()
def close(self, all_fds: bool = False) -> None: def close(self, all_fds: bool = False) -> None:
"""Closes the `IOLoop`, freeing any resources used. """Closes the `IOLoop`, freeing any resources used.

View file

@ -195,9 +195,7 @@ class _StreamBuffer(object):
pos += size pos += size
size = 0 size = 0
else: else:
# Amortized O(1) shrink for Python 2
pos += size pos += size
if len(b) <= 2 * pos:
del typing.cast(bytearray, b)[:pos] del typing.cast(bytearray, b)[:pos]
pos = 0 pos = 0
size = 0 size = 0
@ -254,7 +252,6 @@ class BaseIOStream(object):
self.max_write_buffer_size = max_write_buffer_size self.max_write_buffer_size = max_write_buffer_size
self.error = None # type: Optional[BaseException] self.error = None # type: Optional[BaseException]
self._read_buffer = bytearray() self._read_buffer = bytearray()
self._read_buffer_pos = 0
self._read_buffer_size = 0 self._read_buffer_size = 0
self._user_read_buffer = False self._user_read_buffer = False
self._after_user_read_buffer = None # type: Optional[bytearray] self._after_user_read_buffer = None # type: Optional[bytearray]
@ -451,21 +448,17 @@ class BaseIOStream(object):
available_bytes = self._read_buffer_size available_bytes = self._read_buffer_size
n = len(buf) n = len(buf)
if available_bytes >= n: if available_bytes >= n:
end = self._read_buffer_pos + n buf[:] = memoryview(self._read_buffer)[:n]
buf[:] = memoryview(self._read_buffer)[self._read_buffer_pos : end] del self._read_buffer[:n]
del self._read_buffer[:end]
self._after_user_read_buffer = self._read_buffer self._after_user_read_buffer = self._read_buffer
elif available_bytes > 0: elif available_bytes > 0:
buf[:available_bytes] = memoryview(self._read_buffer)[ buf[:available_bytes] = memoryview(self._read_buffer)[:]
self._read_buffer_pos :
]
# Set up the supplied buffer as our temporary read buffer. # Set up the supplied buffer as our temporary read buffer.
# The original (if it had any data remaining) has been # The original (if it had any data remaining) has been
# saved for later. # saved for later.
self._user_read_buffer = True self._user_read_buffer = True
self._read_buffer = buf self._read_buffer = buf
self._read_buffer_pos = 0
self._read_buffer_size = available_bytes self._read_buffer_size = available_bytes
self._read_bytes = n self._read_bytes = n
self._read_partial = partial self._read_partial = partial
@ -818,7 +811,6 @@ class BaseIOStream(object):
if self._user_read_buffer: if self._user_read_buffer:
self._read_buffer = self._after_user_read_buffer or bytearray() self._read_buffer = self._after_user_read_buffer or bytearray()
self._after_user_read_buffer = None self._after_user_read_buffer = None
self._read_buffer_pos = 0
self._read_buffer_size = len(self._read_buffer) self._read_buffer_size = len(self._read_buffer)
self._user_read_buffer = False self._user_read_buffer = False
result = size # type: Union[int, bytes] result = size # type: Union[int, bytes]
@ -931,20 +923,17 @@ class BaseIOStream(object):
# since large merges are relatively expensive and get undone in # since large merges are relatively expensive and get undone in
# _consume(). # _consume().
if self._read_buffer: if self._read_buffer:
loc = self._read_buffer.find( loc = self._read_buffer.find(self._read_delimiter)
self._read_delimiter, self._read_buffer_pos
)
if loc != -1: if loc != -1:
loc -= self._read_buffer_pos
delimiter_len = len(self._read_delimiter) delimiter_len = len(self._read_delimiter)
self._check_max_bytes(self._read_delimiter, loc + delimiter_len) self._check_max_bytes(self._read_delimiter, loc + delimiter_len)
return loc + delimiter_len return loc + delimiter_len
self._check_max_bytes(self._read_delimiter, self._read_buffer_size) self._check_max_bytes(self._read_delimiter, self._read_buffer_size)
elif self._read_regex is not None: elif self._read_regex is not None:
if self._read_buffer: if self._read_buffer:
m = self._read_regex.search(self._read_buffer, self._read_buffer_pos) m = self._read_regex.search(self._read_buffer)
if m is not None: if m is not None:
loc = m.end() - self._read_buffer_pos loc = m.end()
self._check_max_bytes(self._read_regex, loc) self._check_max_bytes(self._read_regex, loc)
return loc return loc
self._check_max_bytes(self._read_regex, self._read_buffer_size) self._check_max_bytes(self._read_regex, self._read_buffer_size)
@ -1001,19 +990,9 @@ class BaseIOStream(object):
return b"" return b""
assert loc <= self._read_buffer_size assert loc <= self._read_buffer_size
# Slice the bytearray buffer into bytes, without intermediate copying # Slice the bytearray buffer into bytes, without intermediate copying
b = ( b = (memoryview(self._read_buffer)[:loc]).tobytes()
memoryview(self._read_buffer)[
self._read_buffer_pos : self._read_buffer_pos + loc
]
).tobytes()
self._read_buffer_pos += loc
self._read_buffer_size -= loc self._read_buffer_size -= loc
# Amortized O(1) shrink del self._read_buffer[:loc]
# (this heuristic is implemented natively in Python 3.4+
# but is replicated here for Python 2)
if self._read_buffer_pos > self._read_buffer_size:
del self._read_buffer[: self._read_buffer_pos]
self._read_buffer_pos = 0
return b return b
def _check_closed(self) -> None: def _check_closed(self) -> None:
@ -1092,9 +1071,8 @@ class IOStream(BaseIOStream):
.. testcode:: .. testcode::
import tornado.ioloop
import tornado.iostream
import socket import socket
import tornado
async def main(): async def main():
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM, 0) s = socket.socket(socket.AF_INET, socket.SOCK_STREAM, 0)

View file

@ -268,7 +268,7 @@ class Locale(object):
def __init__(self, code: str) -> None: def __init__(self, code: str) -> None:
self.code = code self.code = code
self.name = LOCALE_NAMES.get(code, {}).get("name", u"Unknown") self.name = LOCALE_NAMES.get(code, {}).get("name", "Unknown")
self.rtl = False self.rtl = False
for prefix in ["fa", "ar", "he"]: for prefix in ["fa", "ar", "he"]:
if self.code.startswith(prefix): if self.code.startswith(prefix):
@ -406,7 +406,7 @@ class Locale(object):
str_time = "%d:%02d" % (local_date.hour, local_date.minute) str_time = "%d:%02d" % (local_date.hour, local_date.minute)
elif self.code == "zh_CN": elif self.code == "zh_CN":
str_time = "%s%d:%02d" % ( str_time = "%s%d:%02d" % (
(u"\u4e0a\u5348", u"\u4e0b\u5348")[local_date.hour >= 12], ("\u4e0a\u5348", "\u4e0b\u5348")[local_date.hour >= 12],
local_date.hour % 12 or 12, local_date.hour % 12 or 12,
local_date.minute, local_date.minute,
) )
@ -458,7 +458,7 @@ class Locale(object):
return "" return ""
if len(parts) == 1: if len(parts) == 1:
return parts[0] return parts[0]
comma = u" \u0648 " if self.code.startswith("fa") else u", " comma = " \u0648 " if self.code.startswith("fa") else ", "
return _("%(commas)s and %(last)s") % { return _("%(commas)s and %(last)s") % {
"commas": comma.join(parts[:-1]), "commas": comma.join(parts[:-1]),
"last": parts[len(parts) - 1], "last": parts[len(parts) - 1],

View file

@ -44,10 +44,10 @@ if hasattr(ssl, "OP_NO_COMPRESSION"):
# module-import time, the import lock is already held by the main thread, # module-import time, the import lock is already held by the main thread,
# leading to deadlock. Avoid it by caching the idna encoder on the main # leading to deadlock. Avoid it by caching the idna encoder on the main
# thread now. # thread now.
u"foo".encode("idna") "foo".encode("idna")
# For undiagnosed reasons, 'latin1' codec may also need to be preloaded. # For undiagnosed reasons, 'latin1' codec may also need to be preloaded.
u"foo".encode("latin1") "foo".encode("latin1")
# Default backlog used when calling sock.listen() # Default backlog used when calling sock.listen()
_DEFAULT_BACKLOG = 128 _DEFAULT_BACKLOG = 128
@ -115,7 +115,7 @@ def bind_sockets(
sys.platform == "darwin" sys.platform == "darwin"
and address == "localhost" and address == "localhost"
and af == socket.AF_INET6 and af == socket.AF_INET6
and sockaddr[3] != 0 and sockaddr[3] != 0 # type: ignore
): ):
# Mac OS X includes a link-local address fe80::1%lo0 in the # Mac OS X includes a link-local address fe80::1%lo0 in the
# getaddrinfo results for 'localhost'. However, the firewall # getaddrinfo results for 'localhost'. However, the firewall

View file

@ -56,7 +56,7 @@ Your ``main()`` method can parse the command line or parse a config file with
either `parse_command_line` or `parse_config_file`:: either `parse_command_line` or `parse_config_file`::
import myapp.db, myapp.server import myapp.db, myapp.server
import tornado.options import tornado
if __name__ == '__main__': if __name__ == '__main__':
tornado.options.parse_command_line() tornado.options.parse_command_line()
@ -427,7 +427,9 @@ class OptionParser(object):
% (option.name, option.type.__name__) % (option.name, option.type.__name__)
) )
if type(config[name]) == str and option.type != str: if type(config[name]) == str and (
option.type != str or option.multiple
):
option.parse(config[name]) option.parse(config[name])
else: else:
option.set(config[name]) option.set(config[name])

View file

@ -36,10 +36,10 @@ import warnings
from tornado.gen import convert_yielded from tornado.gen import convert_yielded
from tornado.ioloop import IOLoop, _Selectable from tornado.ioloop import IOLoop, _Selectable
from typing import Any, TypeVar, Awaitable, Callable, Union, Optional, List, Tuple, Dict from typing import Any, TypeVar, Awaitable, Callable, Union, Optional, List, Dict
if typing.TYPE_CHECKING: if typing.TYPE_CHECKING:
from typing import Set # noqa: F401 from typing import Set, Tuple # noqa: F401
from typing_extensions import Protocol from typing_extensions import Protocol
class _HasFileno(Protocol): class _HasFileno(Protocol):
@ -74,20 +74,6 @@ def _atexit_callback() -> None:
atexit.register(_atexit_callback) atexit.register(_atexit_callback)
if sys.version_info >= (3, 10):
def _get_event_loop() -> asyncio.AbstractEventLoop:
try:
return asyncio.get_running_loop()
except RuntimeError:
pass
return asyncio.get_event_loop_policy().get_event_loop()
else:
from asyncio import get_event_loop as _get_event_loop
class BaseAsyncIOLoop(IOLoop): class BaseAsyncIOLoop(IOLoop):
def initialize( # type: ignore def initialize( # type: ignore
@ -206,15 +192,7 @@ class BaseAsyncIOLoop(IOLoop):
handler_func(fileobj, events) handler_func(fileobj, events)
def start(self) -> None: def start(self) -> None:
try:
old_loop = _get_event_loop()
except (RuntimeError, AssertionError):
old_loop = None # type: ignore
try:
asyncio.set_event_loop(self.asyncio_loop)
self.asyncio_loop.run_forever() self.asyncio_loop.run_forever()
finally:
asyncio.set_event_loop(old_loop)
def stop(self) -> None: def stop(self) -> None:
self.asyncio_loop.stop() self.asyncio_loop.stop()
@ -298,7 +276,7 @@ class AsyncIOMainLoop(BaseAsyncIOLoop):
def initialize(self, **kwargs: Any) -> None: # type: ignore def initialize(self, **kwargs: Any) -> None: # type: ignore
super().initialize(asyncio.get_event_loop(), **kwargs) super().initialize(asyncio.get_event_loop(), **kwargs)
def make_current(self) -> None: def _make_current(self) -> None:
# AsyncIOMainLoop already refers to the current asyncio loop so # AsyncIOMainLoop already refers to the current asyncio loop so
# nothing to do here. # nothing to do here.
pass pass
@ -349,12 +327,7 @@ class AsyncIOLoop(BaseAsyncIOLoop):
self._clear_current() self._clear_current()
super().close(all_fds=all_fds) super().close(all_fds=all_fds)
def make_current(self) -> None: def _make_current(self) -> None:
warnings.warn(
"make_current is deprecated; start the event loop first",
DeprecationWarning,
stacklevel=2,
)
if not self.is_current: if not self.is_current:
try: try:
self.old_asyncio = asyncio.get_event_loop() self.old_asyncio = asyncio.get_event_loop()
@ -672,10 +645,18 @@ class AddThreadSelectorEventLoop(asyncio.AbstractEventLoop):
self._writers[fd] = functools.partial(callback, *args) self._writers[fd] = functools.partial(callback, *args)
self._wake_selector() self._wake_selector()
def remove_reader(self, fd: "_FileDescriptorLike") -> None: def remove_reader(self, fd: "_FileDescriptorLike") -> bool:
try:
del self._readers[fd] del self._readers[fd]
except KeyError:
return False
self._wake_selector() self._wake_selector()
return True
def remove_writer(self, fd: "_FileDescriptorLike") -> None: def remove_writer(self, fd: "_FileDescriptorLike") -> bool:
try:
del self._writers[fd] del self._writers[fd]
except KeyError:
return False
self._wake_selector() self._wake_selector()
return True

View file

@ -15,14 +15,15 @@ if typing.TYPE_CHECKING:
class CaresResolver(Resolver): class CaresResolver(Resolver):
"""Name resolver based on the c-ares library. """Name resolver based on the c-ares library.
This is a non-blocking and non-threaded resolver. It may not produce This is a non-blocking and non-threaded resolver. It may not produce the
the same results as the system resolver, but can be used for non-blocking same results as the system resolver, but can be used for non-blocking
resolution when threads cannot be used. resolution when threads cannot be used.
c-ares fails to resolve some names when ``family`` is ``AF_UNSPEC``, ``pycares`` will not return a mix of ``AF_INET`` and ``AF_INET6`` when
so it is only recommended for use in ``AF_INET`` (i.e. IPv4). This is ``family`` is ``AF_UNSPEC``, so it is only recommended for use in
the default for ``tornado.simple_httpclient``, but other libraries ``AF_INET`` (i.e. IPv4). This is the default for
may default to ``AF_UNSPEC``. ``tornado.simple_httpclient``, but other libraries may default to
``AF_UNSPEC``.
.. versionchanged:: 5.0 .. versionchanged:: 5.0
The ``io_loop`` argument (deprecated since version 4.1) has been removed. The ``io_loop`` argument (deprecated since version 4.1) has been removed.

View file

@ -381,7 +381,7 @@ class PriorityQueue(Queue):
def _put(self, item: _T) -> None: def _put(self, item: _T) -> None:
heapq.heappush(self._queue, item) heapq.heappush(self._queue, item)
def _get(self) -> _T: def _get(self) -> _T: # type: ignore[type-var]
return heapq.heappop(self._queue) return heapq.heappop(self._queue)
@ -418,5 +418,5 @@ class LifoQueue(Queue):
def _put(self, item: _T) -> None: def _put(self, item: _T) -> None:
self._queue.append(item) self._queue.append(item)
def _get(self) -> _T: def _get(self) -> _T: # type: ignore[type-var]
return self._queue.pop() return self._queue.pop()

View file

@ -547,7 +547,7 @@ class _HTTPConnection(httputil.HTTPMessageDelegate):
value: Optional[BaseException], value: Optional[BaseException],
tb: Optional[TracebackType], tb: Optional[TracebackType],
) -> bool: ) -> bool:
if self.final_callback: if self.final_callback is not None:
self._remove_timeout() self._remove_timeout()
if isinstance(value, StreamClosedError): if isinstance(value, StreamClosedError):
if value.real_error is None: if value.real_error is None:

View file

@ -21,6 +21,7 @@ import socket
import numbers import numbers
import datetime import datetime
import ssl import ssl
import typing
from tornado.concurrent import Future, future_add_done_callback from tornado.concurrent import Future, future_add_done_callback
from tornado.ioloop import IOLoop from tornado.ioloop import IOLoop
@ -29,7 +30,10 @@ from tornado import gen
from tornado.netutil import Resolver from tornado.netutil import Resolver
from tornado.gen import TimeoutError from tornado.gen import TimeoutError
from typing import Any, Union, Dict, Tuple, List, Callable, Iterator, Optional, Set from typing import Any, Union, Dict, Tuple, List, Callable, Iterator, Optional
if typing.TYPE_CHECKING:
from typing import Set # noqa(F401)
_INITIAL_CONNECT_TIMEOUT = 0.3 _INITIAL_CONNECT_TIMEOUT = 0.3

View file

@ -246,9 +246,7 @@ class TCPServer(object):
.. deprecated:: 6.2 .. deprecated:: 6.2
Use either ``listen()`` or ``add_sockets()`` instead of ``bind()`` Use either ``listen()`` or ``add_sockets()`` instead of ``bind()``
and ``start()``. The ``bind()/start()`` pattern depends on and ``start()``.
interfaces that have been deprecated in Python 3.10 and will be
removed in future versions of Python.
""" """
sockets = bind_sockets( sockets = bind_sockets(
port, port,
@ -295,9 +293,7 @@ class TCPServer(object):
.. deprecated:: 6.2 .. deprecated:: 6.2
Use either ``listen()`` or ``add_sockets()`` instead of ``bind()`` Use either ``listen()`` or ``add_sockets()`` instead of ``bind()``
and ``start()``. The ``bind()/start()`` pattern depends on and ``start()``.
interfaces that have been deprecated in Python 3.10 and will be
removed in future versions of Python.
""" """
assert not self._started assert not self._started
self._started = True self._started = True

View file

@ -135,7 +135,8 @@ class AsyncTestCase(unittest.TestCase):
By default, a new `.IOLoop` is constructed for each test and is available By default, a new `.IOLoop` is constructed for each test and is available
as ``self.io_loop``. If the code being tested requires a as ``self.io_loop``. If the code being tested requires a
global `.IOLoop`, subclasses should override `get_new_ioloop` to return it. reused global `.IOLoop`, subclasses should override `get_new_ioloop` to return it,
although this is deprecated as of Tornado 6.3.
The `.IOLoop`'s ``start`` and ``stop`` methods should not be The `.IOLoop`'s ``start`` and ``stop`` methods should not be
called directly. Instead, use `self.stop <stop>` and `self.wait called directly. Instead, use `self.stop <stop>` and `self.wait
@ -162,17 +163,6 @@ class AsyncTestCase(unittest.TestCase):
response = self.wait() response = self.wait()
# Test contents of response # Test contents of response
self.assertIn("FriendFeed", response.body) self.assertIn("FriendFeed", response.body)
.. deprecated:: 6.2
AsyncTestCase and AsyncHTTPTestCase are deprecated due to changes
in future versions of Python (after 3.10). The interfaces used
in this class are incompatible with the deprecation and intended
removal of certain methods related to the idea of a "current"
event loop while no event loop is actually running. Use
`unittest.IsolatedAsyncioTestCase` instead. Note that this class
does not emit DeprecationWarnings until better migration guidance
can be provided.
""" """
def __init__(self, methodName: str = "runTest") -> None: def __init__(self, methodName: str = "runTest") -> None:
@ -193,6 +183,10 @@ class AsyncTestCase(unittest.TestCase):
self._test_generator = None # type: Optional[Union[Generator, Coroutine]] self._test_generator = None # type: Optional[Union[Generator, Coroutine]]
def setUp(self) -> None: def setUp(self) -> None:
py_ver = sys.version_info
if ((3, 10, 0) <= py_ver < (3, 10, 9)) or ((3, 11, 0) <= py_ver <= (3, 11, 1)):
# Early releases in the Python 3.10 and 3.1 series had deprecation
# warnings that were later reverted; we must suppress them here.
setup_with_context_manager(self, warnings.catch_warnings()) setup_with_context_manager(self, warnings.catch_warnings())
warnings.filterwarnings( warnings.filterwarnings(
"ignore", "ignore",
@ -201,41 +195,10 @@ class AsyncTestCase(unittest.TestCase):
module=r"tornado\..*", module=r"tornado\..*",
) )
super().setUp() super().setUp()
# NOTE: this code attempts to navigate deprecation warnings introduced if type(self).get_new_ioloop is not AsyncTestCase.get_new_ioloop:
# in Python 3.10. The idea of an implicit current event loop is warnings.warn("get_new_ioloop is deprecated", DeprecationWarning)
# deprecated in that version, with the intention that tests like this self.io_loop = self.get_new_ioloop()
# explicitly create a new event loop and run on it. However, other asyncio.set_event_loop(self.io_loop.asyncio_loop) # type: ignore[attr-defined]
# packages such as pytest-asyncio (as of version 0.16.0) still rely on
# the implicit current event loop and we want to be compatible with them
# (even when run on 3.10, but not, of course, on the future version of
# python that removes the get/set_event_loop methods completely).
#
# Deprecation warnings were introduced inconsistently:
# asyncio.get_event_loop warns, but
# asyncio.get_event_loop_policy().get_event_loop does not. Similarly,
# none of the set_event_loop methods warn, although comments on
# https://bugs.python.org/issue39529 indicate that they are also
# intended for future removal.
#
# Therefore, we first attempt to access the event loop with the
# (non-warning) policy method, and if it fails, fall back to creating a
# new event loop. We do not have effective test coverage of the
# new event loop case; this will have to be watched when/if
# get_event_loop is actually removed.
self.should_close_asyncio_loop = False
try:
self.asyncio_loop = asyncio.get_event_loop_policy().get_event_loop()
except Exception:
self.asyncio_loop = asyncio.new_event_loop()
self.should_close_asyncio_loop = True
async def get_loop() -> IOLoop:
return self.get_new_ioloop()
self.io_loop = self.asyncio_loop.run_until_complete(get_loop())
with warnings.catch_warnings():
warnings.simplefilter("ignore", DeprecationWarning)
self.io_loop.make_current()
def tearDown(self) -> None: def tearDown(self) -> None:
# Native coroutines tend to produce warnings if they're not # Native coroutines tend to produce warnings if they're not
@ -270,17 +233,13 @@ class AsyncTestCase(unittest.TestCase):
# Clean up Subprocess, so it can be used again with a new ioloop. # Clean up Subprocess, so it can be used again with a new ioloop.
Subprocess.uninitialize() Subprocess.uninitialize()
with warnings.catch_warnings(): asyncio.set_event_loop(None)
warnings.simplefilter("ignore", DeprecationWarning)
self.io_loop.clear_current()
if not isinstance(self.io_loop, _NON_OWNED_IOLOOPS): if not isinstance(self.io_loop, _NON_OWNED_IOLOOPS):
# Try to clean up any file descriptors left open in the ioloop. # Try to clean up any file descriptors left open in the ioloop.
# This avoids leaks, especially when tests are run repeatedly # This avoids leaks, especially when tests are run repeatedly
# in the same process with autoreload (because curl does not # in the same process with autoreload (because curl does not
# set FD_CLOEXEC on its file descriptors) # set FD_CLOEXEC on its file descriptors)
self.io_loop.close(all_fds=True) self.io_loop.close(all_fds=True)
if self.should_close_asyncio_loop:
self.asyncio_loop.close()
super().tearDown() super().tearDown()
# In case an exception escaped or the StackContext caught an exception # In case an exception escaped or the StackContext caught an exception
# when there wasn't a wait() to re-raise it, do so here. # when there wasn't a wait() to re-raise it, do so here.
@ -298,6 +257,9 @@ class AsyncTestCase(unittest.TestCase):
singletons using the default `.IOLoop`) or if a per-test event singletons using the default `.IOLoop`) or if a per-test event
loop is being provided by another system (such as loop is being provided by another system (such as
``pytest-asyncio``). ``pytest-asyncio``).
.. deprecated:: 6.3
This method will be removed in Tornado 7.0.
""" """
return IOLoop(make_current=False) return IOLoop(make_current=False)
@ -435,10 +397,6 @@ class AsyncHTTPTestCase(AsyncTestCase):
like ``http_client.fetch()``, into a synchronous operation. If you need like ``http_client.fetch()``, into a synchronous operation. If you need
to do other asynchronous operations in tests, you'll probably need to use to do other asynchronous operations in tests, you'll probably need to use
``stop()`` and ``wait()`` yourself. ``stop()`` and ``wait()`` yourself.
.. deprecated:: 6.2
`AsyncTestCase` and `AsyncHTTPTestCase` are deprecated due to changes
in Python 3.10; see comments on `AsyncTestCase` for more details.
""" """
def setUp(self) -> None: def setUp(self) -> None:
@ -672,7 +630,7 @@ def gen_test( # noqa: F811
if self._test_generator is not None and getattr( if self._test_generator is not None and getattr(
self._test_generator, "cr_running", True self._test_generator, "cr_running", True
): ):
self._test_generator.throw(type(e), e) self._test_generator.throw(e)
# In case the test contains an overly broad except # In case the test contains an overly broad except
# clause, we may get back here. # clause, we may get back here.
# Coroutine was stopped or didn't raise a useful stack trace, # Coroutine was stopped or didn't raise a useful stack trace,
@ -724,28 +682,37 @@ class ExpectLog(logging.Filter):
) -> None: ) -> None:
"""Constructs an ExpectLog context manager. """Constructs an ExpectLog context manager.
:param logger: Logger object (or name of logger) to watch. Pass :param logger: Logger object (or name of logger) to watch. Pass an
an empty string to watch the root logger. empty string to watch the root logger.
:param regex: Regular expression to match. Any log entries on :param regex: Regular expression to match. Any log entries on the
the specified logger that match this regex will be suppressed. specified logger that match this regex will be suppressed.
:param required: If true, an exception will be raised if the end of :param required: If true, an exception will be raised if the end of the
the ``with`` statement is reached without matching any log entries. ``with`` statement is reached without matching any log entries.
:param level: A constant from the ``logging`` module indicating the :param level: A constant from the ``logging`` module indicating the
expected log level. If this parameter is provided, only log messages expected log level. If this parameter is provided, only log messages
at this level will be considered to match. Additionally, the at this level will be considered to match. Additionally, the
supplied ``logger`` will have its level adjusted if necessary supplied ``logger`` will have its level adjusted if necessary (for
(for the duration of the ``ExpectLog`` to enable the expected the duration of the ``ExpectLog`` to enable the expected message.
message.
.. versionchanged:: 6.1 .. versionchanged:: 6.1
Added the ``level`` parameter. Added the ``level`` parameter.
.. deprecated:: 6.3
In Tornado 7.0, only ``WARNING`` and higher logging levels will be
matched by default. To match ``INFO`` and lower levels, the ``level``
argument must be used. This is changing to minimize differences
between ``tornado.testing.main`` (which enables ``INFO`` logs by
default) and most other test runners (including those in IDEs)
which have ``INFO`` logs disabled by default.
""" """
if isinstance(logger, basestring_type): if isinstance(logger, basestring_type):
logger = logging.getLogger(logger) logger = logging.getLogger(logger)
self.logger = logger self.logger = logger
self.regex = re.compile(regex) self.regex = re.compile(regex)
self.required = required self.required = required
self.matched = False # matched and deprecated_level_matched are a counter for the respective event.
self.matched = 0
self.deprecated_level_matched = 0
self.logged_stack = False self.logged_stack = False
self.level = level self.level = level
self.orig_level = None # type: Optional[int] self.orig_level = None # type: Optional[int]
@ -755,13 +722,20 @@ class ExpectLog(logging.Filter):
self.logged_stack = True self.logged_stack = True
message = record.getMessage() message = record.getMessage()
if self.regex.match(message): if self.regex.match(message):
if self.level is None and record.levelno < logging.WARNING:
# We're inside the logging machinery here so generating a DeprecationWarning
# here won't be reported cleanly (if warnings-as-errors is enabled, the error
# just gets swallowed by the logging module), and even if it were it would
# have the wrong stack trace. Just remember this fact and report it in
# __exit__ instead.
self.deprecated_level_matched += 1
if self.level is not None and record.levelno != self.level: if self.level is not None and record.levelno != self.level:
app_log.warning( app_log.warning(
"Got expected log message %r at unexpected level (%s vs %s)" "Got expected log message %r at unexpected level (%s vs %s)"
% (message, logging.getLevelName(self.level), record.levelname) % (message, logging.getLevelName(self.level), record.levelname)
) )
return True return True
self.matched = True self.matched += 1
return False return False
return True return True
@ -783,6 +757,15 @@ class ExpectLog(logging.Filter):
self.logger.removeFilter(self) self.logger.removeFilter(self)
if not typ and self.required and not self.matched: if not typ and self.required and not self.matched:
raise Exception("did not get expected log message") raise Exception("did not get expected log message")
if (
not typ
and self.required
and (self.deprecated_level_matched >= self.matched)
):
warnings.warn(
"ExpectLog matched at INFO or below without level argument",
DeprecationWarning,
)
# From https://nedbatchelder.com/blog/201508/using_context_managers_in_test_setup.html # From https://nedbatchelder.com/blog/201508/using_context_managers_in_test_setup.html

View file

@ -23,7 +23,7 @@ Here is a simple "Hello, world" example app:
.. testcode:: .. testcode::
import asyncio import asyncio
import tornado.web import tornado
class MainHandler(tornado.web.RequestHandler): class MainHandler(tornado.web.RequestHandler):
def get(self): def get(self):
@ -166,7 +166,7 @@ May be overridden by passing a ``version`` keyword argument.
""" """
DEFAULT_SIGNED_VALUE_MIN_VERSION = 1 DEFAULT_SIGNED_VALUE_MIN_VERSION = 1
"""The oldest signed value accepted by `.RequestHandler.get_secure_cookie`. """The oldest signed value accepted by `.RequestHandler.get_signed_cookie`.
May be overridden by passing a ``min_version`` keyword argument. May be overridden by passing a ``min_version`` keyword argument.
@ -210,7 +210,7 @@ class RequestHandler(object):
self, self,
application: "Application", application: "Application",
request: httputil.HTTPServerRequest, request: httputil.HTTPServerRequest,
**kwargs: Any **kwargs: Any,
) -> None: ) -> None:
super().__init__() super().__init__()
@ -603,21 +603,28 @@ class RequestHandler(object):
expires: Optional[Union[float, Tuple, datetime.datetime]] = None, expires: Optional[Union[float, Tuple, datetime.datetime]] = None,
path: str = "/", path: str = "/",
expires_days: Optional[float] = None, expires_days: Optional[float] = None,
**kwargs: Any # Keyword-only args start here for historical reasons.
*,
max_age: Optional[int] = None,
httponly: bool = False,
secure: bool = False,
samesite: Optional[str] = None,
) -> None: ) -> None:
"""Sets an outgoing cookie name/value with the given options. """Sets an outgoing cookie name/value with the given options.
Newly-set cookies are not immediately visible via `get_cookie`; Newly-set cookies are not immediately visible via `get_cookie`;
they are not present until the next request. they are not present until the next request.
expires may be a numeric timestamp as returned by `time.time`, Most arguments are passed directly to `http.cookies.Morsel` directly.
a time tuple as returned by `time.gmtime`, or a See https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Set-Cookie
`datetime.datetime` object. for more information.
``expires`` may be a numeric timestamp as returned by `time.time`,
a time tuple as returned by `time.gmtime`, or a
`datetime.datetime` object. ``expires_days`` is provided as a convenience
to set an expiration time in days from today (if both are set, ``expires``
is used).
Additional keyword arguments are set on the cookies.Morsel
directly.
See https://docs.python.org/3/library/http.cookies.html#http.cookies.Morsel
for available attributes.
""" """
# The cookie library only accepts type str, in both python 2 and 3 # The cookie library only accepts type str, in both python 2 and 3
name = escape.native_str(name) name = escape.native_str(name)
@ -641,56 +648,82 @@ class RequestHandler(object):
morsel["expires"] = httputil.format_timestamp(expires) morsel["expires"] = httputil.format_timestamp(expires)
if path: if path:
morsel["path"] = path morsel["path"] = path
for k, v in kwargs.items(): if max_age:
if k == "max_age": # Note change from _ to -.
k = "max-age" morsel["max-age"] = str(max_age)
if httponly:
# Note that SimpleCookie ignores the value here. The presense of an
# httponly (or secure) key is treated as true.
morsel["httponly"] = True
if secure:
morsel["secure"] = True
if samesite:
morsel["samesite"] = samesite
# skip falsy values for httponly and secure flags because def clear_cookie(self, name: str, **kwargs: Any) -> None:
# SimpleCookie sets them regardless
if k in ["httponly", "secure"] and not v:
continue
morsel[k] = v
def clear_cookie(
self, name: str, path: str = "/", domain: Optional[str] = None
) -> None:
"""Deletes the cookie with the given name. """Deletes the cookie with the given name.
Due to limitations of the cookie protocol, you must pass the same This method accepts the same arguments as `set_cookie`, except for
path and domain to clear a cookie as were used when that cookie ``expires`` and ``max_age``. Clearing a cookie requires the same
was set (but there is no way to find out on the server side ``domain`` and ``path`` arguments as when it was set. In some cases the
which values were used for a given cookie). ``samesite`` and ``secure`` arguments are also required to match. Other
arguments are ignored.
Similar to `set_cookie`, the effect of this method will not be Similar to `set_cookie`, the effect of this method will not be
seen until the following request. seen until the following request.
.. versionchanged:: 6.3
Now accepts all keyword arguments that ``set_cookie`` does.
The ``samesite`` and ``secure`` flags have recently become
required for clearing ``samesite="none"`` cookies.
""" """
for excluded_arg in ["expires", "max_age"]:
if excluded_arg in kwargs:
raise TypeError(
f"clear_cookie() got an unexpected keyword argument '{excluded_arg}'"
)
expires = datetime.datetime.utcnow() - datetime.timedelta(days=365) expires = datetime.datetime.utcnow() - datetime.timedelta(days=365)
self.set_cookie(name, value="", path=path, expires=expires, domain=domain) self.set_cookie(name, value="", expires=expires, **kwargs)
def clear_all_cookies(self, path: str = "/", domain: Optional[str] = None) -> None: def clear_all_cookies(self, **kwargs: Any) -> None:
"""Deletes all the cookies the user sent with this request. """Attempt to delete all the cookies the user sent with this request.
See `clear_cookie` for more information on the path and domain See `clear_cookie` for more information on keyword arguments. Due to
parameters. limitations of the cookie protocol, it is impossible to determine on the
server side which values are necessary for the ``domain``, ``path``,
``samesite``, or ``secure`` arguments, this method can only be
successful if you consistently use the same values for these arguments
when setting cookies.
Similar to `set_cookie`, the effect of this method will not be Similar to `set_cookie`, the effect of this method will not be seen
seen until the following request. until the following request.
.. versionchanged:: 3.2 .. versionchanged:: 3.2
Added the ``path`` and ``domain`` parameters. Added the ``path`` and ``domain`` parameters.
.. versionchanged:: 6.3
Now accepts all keyword arguments that ``set_cookie`` does.
.. deprecated:: 6.3
The increasingly complex rules governing cookies have made it
impossible for a ``clear_all_cookies`` method to work reliably
since all we know about cookies are their names. Applications
should generally use ``clear_cookie`` one at a time instead.
""" """
for name in self.request.cookies: for name in self.request.cookies:
self.clear_cookie(name, path=path, domain=domain) self.clear_cookie(name, **kwargs)
def set_secure_cookie( def set_signed_cookie(
self, self,
name: str, name: str,
value: Union[str, bytes], value: Union[str, bytes],
expires_days: Optional[float] = 30, expires_days: Optional[float] = 30,
version: Optional[int] = None, version: Optional[int] = None,
**kwargs: Any **kwargs: Any,
) -> None: ) -> None:
"""Signs and timestamps a cookie so it cannot be forged. """Signs and timestamps a cookie so it cannot be forged.
@ -698,11 +731,11 @@ class RequestHandler(object):
to use this method. It should be a long, random sequence of bytes to use this method. It should be a long, random sequence of bytes
to be used as the HMAC secret for the signature. to be used as the HMAC secret for the signature.
To read a cookie set with this method, use `get_secure_cookie()`. To read a cookie set with this method, use `get_signed_cookie()`.
Note that the ``expires_days`` parameter sets the lifetime of the Note that the ``expires_days`` parameter sets the lifetime of the
cookie in the browser, but is independent of the ``max_age_days`` cookie in the browser, but is independent of the ``max_age_days``
parameter to `get_secure_cookie`. parameter to `get_signed_cookie`.
A value of None limits the lifetime to the current browser session. A value of None limits the lifetime to the current browser session.
Secure cookies may contain arbitrary byte values, not just unicode Secure cookies may contain arbitrary byte values, not just unicode
@ -715,22 +748,30 @@ class RequestHandler(object):
Added the ``version`` argument. Introduced cookie version 2 Added the ``version`` argument. Introduced cookie version 2
and made it the default. and made it the default.
.. versionchanged:: 6.3
Renamed from ``set_secure_cookie`` to ``set_signed_cookie`` to
avoid confusion with other uses of "secure" in cookie attributes
and prefixes. The old name remains as an alias.
""" """
self.set_cookie( self.set_cookie(
name, name,
self.create_signed_value(name, value, version=version), self.create_signed_value(name, value, version=version),
expires_days=expires_days, expires_days=expires_days,
**kwargs **kwargs,
) )
set_secure_cookie = set_signed_cookie
def create_signed_value( def create_signed_value(
self, name: str, value: Union[str, bytes], version: Optional[int] = None self, name: str, value: Union[str, bytes], version: Optional[int] = None
) -> bytes: ) -> bytes:
"""Signs and timestamps a string so it cannot be forged. """Signs and timestamps a string so it cannot be forged.
Normally used via set_secure_cookie, but provided as a separate Normally used via set_signed_cookie, but provided as a separate
method for non-cookie uses. To decode a value not stored method for non-cookie uses. To decode a value not stored
as a cookie use the optional value argument to get_secure_cookie. as a cookie use the optional value argument to get_signed_cookie.
.. versionchanged:: 3.2.1 .. versionchanged:: 3.2.1
@ -749,7 +790,7 @@ class RequestHandler(object):
secret, name, value, version=version, key_version=key_version secret, name, value, version=version, key_version=key_version
) )
def get_secure_cookie( def get_signed_cookie(
self, self,
name: str, name: str,
value: Optional[str] = None, value: Optional[str] = None,
@ -763,12 +804,19 @@ class RequestHandler(object):
Similar to `get_cookie`, this method only returns cookies that Similar to `get_cookie`, this method only returns cookies that
were present in the request. It does not see outgoing cookies set by were present in the request. It does not see outgoing cookies set by
`set_secure_cookie` in this handler. `set_signed_cookie` in this handler.
.. versionchanged:: 3.2.1 .. versionchanged:: 3.2.1
Added the ``min_version`` argument. Introduced cookie version 2; Added the ``min_version`` argument. Introduced cookie version 2;
both versions 1 and 2 are accepted by default. both versions 1 and 2 are accepted by default.
.. versionchanged:: 6.3
Renamed from ``get_secure_cookie`` to ``get_signed_cookie`` to
avoid confusion with other uses of "secure" in cookie attributes
and prefixes. The old name remains as an alias.
""" """
self.require_setting("cookie_secret", "secure cookies") self.require_setting("cookie_secret", "secure cookies")
if value is None: if value is None:
@ -781,12 +829,22 @@ class RequestHandler(object):
min_version=min_version, min_version=min_version,
) )
def get_secure_cookie_key_version( get_secure_cookie = get_signed_cookie
def get_signed_cookie_key_version(
self, name: str, value: Optional[str] = None self, name: str, value: Optional[str] = None
) -> Optional[int]: ) -> Optional[int]:
"""Returns the signing key version of the secure cookie. """Returns the signing key version of the secure cookie.
The version is returned as int. The version is returned as int.
.. versionchanged:: 6.3
Renamed from ``get_secure_cookie_key_version`` to
``set_signed_cookie_key_version`` to avoid confusion with other
uses of "secure" in cookie attributes and prefixes. The old name
remains as an alias.
""" """
self.require_setting("cookie_secret", "secure cookies") self.require_setting("cookie_secret", "secure cookies")
if value is None: if value is None:
@ -795,6 +853,8 @@ class RequestHandler(object):
return None return None
return get_signature_key_version(value) return get_signature_key_version(value)
get_secure_cookie_key_version = get_signed_cookie_key_version
def redirect( def redirect(
self, url: str, permanent: bool = False, status: Optional[int] = None self, url: str, permanent: bool = False, status: Optional[int] = None
) -> None: ) -> None:
@ -1321,7 +1381,7 @@ class RequestHandler(object):
and is cached for future access:: and is cached for future access::
def get_current_user(self): def get_current_user(self):
user_cookie = self.get_secure_cookie("user") user_cookie = self.get_signed_cookie("user")
if user_cookie: if user_cookie:
return json.loads(user_cookie) return json.loads(user_cookie)
return None return None
@ -1331,7 +1391,7 @@ class RequestHandler(object):
@gen.coroutine @gen.coroutine
def prepare(self): def prepare(self):
user_id_cookie = self.get_secure_cookie("user_id") user_id_cookie = self.get_signed_cookie("user_id")
if user_id_cookie: if user_id_cookie:
self.current_user = yield load_user(user_id_cookie) self.current_user = yield load_user(user_id_cookie)
@ -1643,7 +1703,7 @@ class RequestHandler(object):
# Find all weak and strong etag values from If-None-Match header # Find all weak and strong etag values from If-None-Match header
# because RFC 7232 allows multiple etag values in a single header. # because RFC 7232 allows multiple etag values in a single header.
etags = re.findall( etags = re.findall(
br'\*|(?:W/)?"[^"]*"', utf8(self.request.headers.get("If-None-Match", "")) rb'\*|(?:W/)?"[^"]*"', utf8(self.request.headers.get("If-None-Match", ""))
) )
if not computed_etag or not etags: if not computed_etag or not etags:
return False return False
@ -1676,20 +1736,16 @@ class RequestHandler(object):
) )
# If XSRF cookies are turned on, reject form submissions without # If XSRF cookies are turned on, reject form submissions without
# the proper cookie # the proper cookie
if ( if self.request.method not in (
self.request.method
not in (
"GET", "GET",
"HEAD", "HEAD",
"OPTIONS", "OPTIONS",
) ) and self.application.settings.get("xsrf_cookies"):
and self.application.settings.get("xsrf_cookies")
):
self.check_xsrf_cookie() self.check_xsrf_cookie()
result = self.prepare() result = self.prepare()
if result is not None: if result is not None:
result = await result result = await result # type: ignore
if self._prepared_future is not None: if self._prepared_future is not None:
# Tell the Application we've finished with prepare() # Tell the Application we've finished with prepare()
# and are ready for the body to arrive. # and are ready for the body to arrive.
@ -1848,7 +1904,7 @@ def stream_request_body(cls: Type[_RequestHandlerType]) -> Type[_RequestHandlerT
* The regular HTTP method (``post``, ``put``, etc) will be called after * The regular HTTP method (``post``, ``put``, etc) will be called after
the entire body has been read. the entire body has been read.
See the `file receiver demo <https://github.com/tornadoweb/tornado/tree/master/demos/file_upload/>`_ See the `file receiver demo <https://github.com/tornadoweb/tornado/tree/stable/demos/file_upload/>`_
for example usage. for example usage.
""" # noqa: E501 """ # noqa: E501
if not issubclass(cls, RequestHandler): if not issubclass(cls, RequestHandler):
@ -2046,7 +2102,7 @@ class Application(ReversibleRouter):
handlers: Optional[_RuleList] = None, handlers: Optional[_RuleList] = None,
default_host: Optional[str] = None, default_host: Optional[str] = None,
transforms: Optional[List[Type["OutputTransform"]]] = None, transforms: Optional[List[Type["OutputTransform"]]] = None,
**settings: Any **settings: Any,
) -> None: ) -> None:
if transforms is None: if transforms is None:
self.transforms = [] # type: List[Type[OutputTransform]] self.transforms = [] # type: List[Type[OutputTransform]]
@ -2106,7 +2162,7 @@ class Application(ReversibleRouter):
backlog: int = tornado.netutil._DEFAULT_BACKLOG, backlog: int = tornado.netutil._DEFAULT_BACKLOG,
flags: Optional[int] = None, flags: Optional[int] = None,
reuse_port: bool = False, reuse_port: bool = False,
**kwargs: Any **kwargs: Any,
) -> HTTPServer: ) -> HTTPServer:
"""Starts an HTTP server for this application on the given port. """Starts an HTTP server for this application on the given port.
@ -2393,7 +2449,7 @@ class HTTPError(Exception):
status_code: int = 500, status_code: int = 500,
log_message: Optional[str] = None, log_message: Optional[str] = None,
*args: Any, *args: Any,
**kwargs: Any **kwargs: Any,
) -> None: ) -> None:
self.status_code = status_code self.status_code = status_code
self.log_message = log_message self.log_message = log_message
@ -3441,7 +3497,7 @@ def create_signed_value(
# A leading version number in decimal # A leading version number in decimal
# with no leading zeros, followed by a pipe. # with no leading zeros, followed by a pipe.
_signed_value_version_re = re.compile(br"^([1-9][0-9]*)\|(.*)$") _signed_value_version_re = re.compile(rb"^([1-9][0-9]*)\|(.*)$")
def _get_version(value: bytes) -> int: def _get_version(value: bytes) -> int:

View file

@ -23,7 +23,6 @@ import hashlib
import os import os
import sys import sys
import struct import struct
import tornado.escape
import tornado.web import tornado.web
from urllib.parse import urlparse from urllib.parse import urlparse
import zlib import zlib
@ -34,6 +33,7 @@ from tornado import gen, httpclient, httputil
from tornado.ioloop import IOLoop, PeriodicCallback from tornado.ioloop import IOLoop, PeriodicCallback
from tornado.iostream import StreamClosedError, IOStream from tornado.iostream import StreamClosedError, IOStream
from tornado.log import gen_log, app_log from tornado.log import gen_log, app_log
from tornado.netutil import Resolver
from tornado import simple_httpclient from tornado import simple_httpclient
from tornado.queues import Queue from tornado.queues import Queue
from tornado.tcpclient import TCPClient from tornado.tcpclient import TCPClient
@ -822,7 +822,7 @@ class WebSocketProtocol13(WebSocketProtocol):
self._masked_frame = None self._masked_frame = None
self._frame_mask = None # type: Optional[bytes] self._frame_mask = None # type: Optional[bytes]
self._frame_length = None self._frame_length = None
self._fragmented_message_buffer = None # type: Optional[bytes] self._fragmented_message_buffer = None # type: Optional[bytearray]
self._fragmented_message_opcode = None self._fragmented_message_opcode = None
self._waiting = None # type: object self._waiting = None # type: object
self._compression_options = params.compression_options self._compression_options = params.compression_options
@ -1177,10 +1177,10 @@ class WebSocketProtocol13(WebSocketProtocol):
# nothing to continue # nothing to continue
self._abort() self._abort()
return return
self._fragmented_message_buffer += data self._fragmented_message_buffer.extend(data)
if is_final_frame: if is_final_frame:
opcode = self._fragmented_message_opcode opcode = self._fragmented_message_opcode
data = self._fragmented_message_buffer data = bytes(self._fragmented_message_buffer)
self._fragmented_message_buffer = None self._fragmented_message_buffer = None
else: # start of new data message else: # start of new data message
if self._fragmented_message_buffer is not None: if self._fragmented_message_buffer is not None:
@ -1189,7 +1189,7 @@ class WebSocketProtocol13(WebSocketProtocol):
return return
if not is_final_frame: if not is_final_frame:
self._fragmented_message_opcode = opcode self._fragmented_message_opcode = opcode
self._fragmented_message_buffer = data self._fragmented_message_buffer = bytearray(data)
if is_final_frame: if is_final_frame:
handled_future = self._handle_message(opcode, data) handled_future = self._handle_message(opcode, data)
@ -1362,6 +1362,7 @@ class WebSocketClientConnection(simple_httpclient._HTTPConnection):
ping_timeout: Optional[float] = None, ping_timeout: Optional[float] = None,
max_message_size: int = _default_max_message_size, max_message_size: int = _default_max_message_size,
subprotocols: Optional[List[str]] = [], subprotocols: Optional[List[str]] = [],
resolver: Optional[Resolver] = None,
) -> None: ) -> None:
self.connect_future = Future() # type: Future[WebSocketClientConnection] self.connect_future = Future() # type: Future[WebSocketClientConnection]
self.read_queue = Queue(1) # type: Queue[Union[None, str, bytes]] self.read_queue = Queue(1) # type: Queue[Union[None, str, bytes]]
@ -1402,7 +1403,7 @@ class WebSocketClientConnection(simple_httpclient._HTTPConnection):
# Websocket connection is currently unable to follow redirects # Websocket connection is currently unable to follow redirects
request.follow_redirects = False request.follow_redirects = False
self.tcp_client = TCPClient() self.tcp_client = TCPClient(resolver=resolver)
super().__init__( super().__init__(
None, None,
request, request,

View file

@ -27,12 +27,15 @@ container.
""" """
import sys import concurrent.futures
from io import BytesIO from io import BytesIO
import tornado import tornado
import sys
from tornado.concurrent import dummy_executor
from tornado import escape from tornado import escape
from tornado import httputil from tornado import httputil
from tornado.ioloop import IOLoop
from tornado.log import access_log from tornado.log import access_log
from typing import List, Tuple, Optional, Callable, Any, Dict, Text from typing import List, Tuple, Optional, Callable, Any, Dict, Text
@ -54,20 +57,28 @@ def to_wsgi_str(s: bytes) -> str:
class WSGIContainer(object): class WSGIContainer(object):
r"""Makes a WSGI-compatible function runnable on Tornado's HTTP server. r"""Makes a WSGI-compatible application runnable on Tornado's HTTP server.
.. warning:: .. warning::
WSGI is a *synchronous* interface, while Tornado's concurrency model WSGI is a *synchronous* interface, while Tornado's concurrency model
is based on single-threaded asynchronous execution. This means that is based on single-threaded *asynchronous* execution. Many of Tornado's
running a WSGI app with Tornado's `WSGIContainer` is *less scalable* distinguishing features are not available in WSGI mode, including efficient
than running the same app in a multi-threaded WSGI server like long-polling and websockets. The primary purpose of `WSGIContainer` is
``gunicorn`` or ``uwsgi``. Use `WSGIContainer` only when there are to support both WSGI applications and native Tornado ``RequestHandlers`` in
benefits to combining Tornado and WSGI in the same process that a single process. WSGI-only applications are likely to be better off
outweigh the reduced scalability. with a dedicated WSGI server such as ``gunicorn`` or ``uwsgi``.
Wrap a WSGI function in a `WSGIContainer` and pass it to `.HTTPServer` to Wrap a WSGI application in a `WSGIContainer` to make it implement the Tornado
run it. For example:: `.HTTPServer` ``request_callback`` interface. The `WSGIContainer` object can
then be passed to classes from the `tornado.routing` module,
`tornado.web.FallbackHandler`, or to `.HTTPServer` directly.
This class is intended to let other frameworks (Django, Flask, etc)
run on the Tornado HTTP server and I/O loop.
Realistic usage will be more complicated, but the simplest possible example uses a
hand-written WSGI application with `.HTTPServer`::
def simple_app(environ, start_response): def simple_app(environ, start_response):
status = "200 OK" status = "200 OK"
@ -83,18 +94,46 @@ class WSGIContainer(object):
asyncio.run(main()) asyncio.run(main())
This class is intended to let other frameworks (Django, web.py, etc) The recommended pattern is to use the `tornado.routing` module to set up routing
run on the Tornado HTTP server and I/O loop. rules between your WSGI application and, typically, a `tornado.web.Application`.
Alternatively, `tornado.web.Application` can be used as the top-level router
and `tornado.web.FallbackHandler` can embed a `WSGIContainer` within it.
The `tornado.web.FallbackHandler` class is often useful for mixing If the ``executor`` argument is provided, the WSGI application will be executed
Tornado and WSGI apps in the same server. See on that executor. This must be an instance of `concurrent.futures.Executor`,
https://github.com/bdarnell/django-tornado-demo for a complete example. typically a ``ThreadPoolExecutor`` (``ProcessPoolExecutor`` is not supported).
If no ``executor`` is given, the application will run on the event loop thread in
Tornado 6.3; this will change to use an internal thread pool by default in
Tornado 7.0.
.. warning::
By default, the WSGI application is executed on the event loop's thread. This
limits the server to one request at a time (per process), making it less scalable
than most other WSGI servers. It is therefore highly recommended that you pass
a ``ThreadPoolExecutor`` when constructing the `WSGIContainer`, after verifying
that your application is thread-safe. The default will change to use a
``ThreadPoolExecutor`` in Tornado 7.0.
.. versionadded:: 6.3
The ``executor`` parameter.
.. deprecated:: 6.3
The default behavior of running the WSGI application on the event loop thread
is deprecated and will change in Tornado 7.0 to use a thread pool by default.
""" """
def __init__(self, wsgi_application: "WSGIAppType") -> None: def __init__(
self,
wsgi_application: "WSGIAppType",
executor: Optional[concurrent.futures.Executor] = None,
) -> None:
self.wsgi_application = wsgi_application self.wsgi_application = wsgi_application
self.executor = dummy_executor if executor is None else executor
def __call__(self, request: httputil.HTTPServerRequest) -> None: def __call__(self, request: httputil.HTTPServerRequest) -> None:
IOLoop.current().spawn_callback(self.handle_request, request)
async def handle_request(self, request: httputil.HTTPServerRequest) -> None:
data = {} # type: Dict[str, Any] data = {} # type: Dict[str, Any]
response = [] # type: List[bytes] response = [] # type: List[bytes]
@ -113,15 +152,33 @@ class WSGIContainer(object):
data["headers"] = headers data["headers"] = headers
return response.append return response.append
app_response = self.wsgi_application( loop = IOLoop.current()
WSGIContainer.environ(request), start_response app_response = await loop.run_in_executor(
self.executor,
self.wsgi_application,
self.environ(request),
start_response,
) )
try: try:
response.extend(app_response) app_response_iter = iter(app_response)
body = b"".join(response)
def next_chunk() -> Optional[bytes]:
try:
return next(app_response_iter)
except StopIteration:
# StopIteration is special and is not allowed to pass through
# coroutines normally.
return None
while True:
chunk = await loop.run_in_executor(self.executor, next_chunk)
if chunk is None:
break
response.append(chunk)
finally: finally:
if hasattr(app_response, "close"): if hasattr(app_response, "close"):
app_response.close() # type: ignore app_response.close() # type: ignore
body = b"".join(response)
if not data: if not data:
raise Exception("WSGI app did not call start_response") raise Exception("WSGI app did not call start_response")
@ -147,9 +204,12 @@ class WSGIContainer(object):
request.connection.finish() request.connection.finish()
self._log(status_code, request) self._log(status_code, request)
@staticmethod def environ(self, request: httputil.HTTPServerRequest) -> Dict[Text, Any]:
def environ(request: httputil.HTTPServerRequest) -> Dict[Text, Any]: """Converts a `tornado.httputil.HTTPServerRequest` to a WSGI environment.
"""Converts a `tornado.httputil.HTTPServerRequest` to a WSGI environment."""
.. versionchanged:: 6.3
No longer a static method.
"""
hostport = request.host.split(":") hostport = request.host.split(":")
if len(hostport) == 2: if len(hostport) == 2:
host = hostport[0] host = hostport[0]
@ -172,7 +232,7 @@ class WSGIContainer(object):
"wsgi.url_scheme": request.protocol, "wsgi.url_scheme": request.protocol,
"wsgi.input": BytesIO(escape.utf8(request.body)), "wsgi.input": BytesIO(escape.utf8(request.body)),
"wsgi.errors": sys.stderr, "wsgi.errors": sys.stderr,
"wsgi.multithread": False, "wsgi.multithread": self.executor is not dummy_executor,
"wsgi.multiprocess": True, "wsgi.multiprocess": True,
"wsgi.run_once": False, "wsgi.run_once": False,
} }

View file

@ -320,7 +320,7 @@ class BaseHandler(RouteHandler):
def get_current_user(self): def get_current_user(self):
if sickgear.WEB_USERNAME or sickgear.WEB_PASSWORD: if sickgear.WEB_USERNAME or sickgear.WEB_PASSWORD:
return self.get_secure_cookie('sickgear-session-%s' % helpers.md5_for_text(sickgear.WEB_PORT)) return self.get_signed_cookie('sickgear-session-%s' % helpers.md5_for_text(sickgear.WEB_PORT))
return True return True
def get_image(self, image): def get_image(self, image):
@ -401,7 +401,7 @@ class LoginHandler(BaseHandler):
httponly=True) httponly=True)
if sickgear.ENABLE_HTTPS: if sickgear.ENABLE_HTTPS:
params.update(dict(secure=True)) params.update(dict(secure=True))
self.set_secure_cookie('sickgear-session-%s' % helpers.md5_for_text(sickgear.WEB_PORT), self.set_signed_cookie('sickgear-session-%s' % helpers.md5_for_text(sickgear.WEB_PORT),
sickgear.COOKIE_SECRET, **params) sickgear.COOKIE_SECRET, **params)
self.redirect(self.get_argument('next', '/home/')) self.redirect(self.get_argument('next', '/home/'))
else: else: