2014-06-11 04:34:28 -04:00
|
|
|
#!/usr/bin/env python
|
|
|
|
#
|
|
|
|
# Copyright 2009 Facebook
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
|
|
# not use this file except in compliance with the License. You may obtain
|
|
|
|
# a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
|
|
# License for the specific language governing permissions and limitations
|
|
|
|
# under the License.
|
|
|
|
|
|
|
|
"""Utility classes to write to and read from non-blocking files and sockets.
|
|
|
|
|
|
|
|
Contents:
|
|
|
|
|
|
|
|
* `BaseIOStream`: Generic interface for reading and writing.
|
|
|
|
* `IOStream`: Implementation of BaseIOStream using non-blocking sockets.
|
|
|
|
* `SSLIOStream`: SSL-aware version of IOStream.
|
|
|
|
* `PipeIOStream`: Pipe-based IOStream implementation.
|
|
|
|
"""
|
|
|
|
|
|
|
|
from __future__ import absolute_import, division, print_function, with_statement
|
|
|
|
|
|
|
|
import collections
|
|
|
|
import errno
|
|
|
|
import numbers
|
|
|
|
import os
|
|
|
|
import socket
|
|
|
|
import sys
|
|
|
|
import re
|
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
from tornado.concurrent import TracebackFuture
|
2014-06-11 04:34:28 -04:00
|
|
|
from tornado import ioloop
|
|
|
|
from tornado.log import gen_log, app_log
|
|
|
|
from tornado.netutil import ssl_wrap_socket, ssl_match_hostname, SSLCertificateError
|
|
|
|
from tornado import stack_context
|
2014-10-14 00:24:01 -04:00
|
|
|
from tornado.util import errno_from_exception
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
try:
|
|
|
|
from tornado.platform.posix import _set_nonblocking
|
|
|
|
except ImportError:
|
|
|
|
_set_nonblocking = None
|
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
try:
|
|
|
|
import ssl
|
|
|
|
except ImportError:
|
|
|
|
# ssl is not available on Google App Engine
|
|
|
|
ssl = None
|
|
|
|
|
2014-06-11 04:34:28 -04:00
|
|
|
# These errnos indicate that a non-blocking operation must be retried
|
|
|
|
# at a later time. On most platforms they're the same value, but on
|
|
|
|
# some they differ.
|
|
|
|
_ERRNO_WOULDBLOCK = (errno.EWOULDBLOCK, errno.EAGAIN)
|
|
|
|
|
2014-06-29 06:05:33 -04:00
|
|
|
if hasattr(errno, "WSAEWOULDBLOCK"):
|
|
|
|
_ERRNO_WOULDBLOCK += (errno.WSAEWOULDBLOCK,)
|
|
|
|
|
2014-06-11 04:34:28 -04:00
|
|
|
# These errnos indicate that a connection has been abruptly terminated.
|
|
|
|
# They should be caught and handled less noisily than other errors.
|
2014-06-17 00:54:00 -04:00
|
|
|
_ERRNO_CONNRESET = (errno.ECONNRESET, errno.ECONNABORTED, errno.EPIPE,
|
|
|
|
errno.ETIMEDOUT)
|
2014-06-11 04:34:28 -04:00
|
|
|
|
2014-06-29 06:05:33 -04:00
|
|
|
if hasattr(errno, "WSAECONNRESET"):
|
|
|
|
_ERRNO_CONNRESET += (errno.WSAECONNRESET, errno.WSAECONNABORTED, errno.WSAETIMEDOUT)
|
|
|
|
|
|
|
|
# More non-portable errnos:
|
|
|
|
_ERRNO_INPROGRESS = (errno.EINPROGRESS,)
|
|
|
|
|
|
|
|
if hasattr(errno, "WSAEINPROGRESS"):
|
|
|
|
_ERRNO_INPROGRESS += (errno.WSAEINPROGRESS,)
|
2014-06-11 04:34:28 -04:00
|
|
|
|
2014-06-29 06:05:33 -04:00
|
|
|
#######################################################
|
2014-06-11 04:34:28 -04:00
|
|
|
class StreamClosedError(IOError):
|
|
|
|
"""Exception raised by `IOStream` methods when the stream is closed.
|
|
|
|
|
|
|
|
Note that the close callback is scheduled to run *after* other
|
|
|
|
callbacks on the stream (to allow for buffered data to be processed),
|
|
|
|
so you may see this error before you see the close callback.
|
|
|
|
"""
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
class UnsatisfiableReadError(Exception):
|
|
|
|
"""Exception raised when a read cannot be satisfied.
|
|
|
|
|
|
|
|
Raised by ``read_until`` and ``read_until_regex`` with a ``max_bytes``
|
|
|
|
argument.
|
|
|
|
"""
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class StreamBufferFullError(Exception):
|
|
|
|
"""Exception raised by `IOStream` methods when the buffer is full.
|
|
|
|
"""
|
|
|
|
|
|
|
|
|
2014-06-11 04:34:28 -04:00
|
|
|
class BaseIOStream(object):
|
|
|
|
"""A utility class to write to and read from a non-blocking file or socket.
|
|
|
|
|
|
|
|
We support a non-blocking ``write()`` and a family of ``read_*()`` methods.
|
2014-06-17 00:54:00 -04:00
|
|
|
All of the methods take an optional ``callback`` argument and return a
|
|
|
|
`.Future` only if no callback is given. When the operation completes,
|
|
|
|
the callback will be run or the `.Future` will resolve with the data
|
|
|
|
read (or ``None`` for ``write()``). All outstanding ``Futures`` will
|
|
|
|
resolve with a `StreamClosedError` when the stream is closed; users
|
|
|
|
of the callback interface will be notified via
|
|
|
|
`.BaseIOStream.set_close_callback` instead.
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
When a stream is closed due to an error, the IOStream's ``error``
|
|
|
|
attribute contains the exception object.
|
|
|
|
|
|
|
|
Subclasses must implement `fileno`, `close_fd`, `write_to_fd`,
|
|
|
|
`read_from_fd`, and optionally `get_fd_error`.
|
|
|
|
"""
|
|
|
|
def __init__(self, io_loop=None, max_buffer_size=None,
|
2014-06-17 00:54:00 -04:00
|
|
|
read_chunk_size=None, max_write_buffer_size=None):
|
|
|
|
"""`BaseIOStream` constructor.
|
|
|
|
|
|
|
|
:arg io_loop: The `.IOLoop` to use; defaults to `.IOLoop.current`.
|
|
|
|
:arg max_buffer_size: Maximum amount of incoming data to buffer;
|
|
|
|
defaults to 100MB.
|
|
|
|
:arg read_chunk_size: Amount of data to read at one time from the
|
|
|
|
underlying transport; defaults to 64KB.
|
|
|
|
:arg max_write_buffer_size: Amount of outgoing data to buffer;
|
|
|
|
defaults to unlimited.
|
|
|
|
|
|
|
|
.. versionchanged:: 4.0
|
|
|
|
Add the ``max_write_buffer_size`` parameter. Changed default
|
|
|
|
``read_chunk_size`` to 64KB.
|
|
|
|
"""
|
2014-06-11 04:34:28 -04:00
|
|
|
self.io_loop = io_loop or ioloop.IOLoop.current()
|
|
|
|
self.max_buffer_size = max_buffer_size or 104857600
|
2014-06-17 00:54:00 -04:00
|
|
|
# A chunk size that is too close to max_buffer_size can cause
|
|
|
|
# spurious failures.
|
|
|
|
self.read_chunk_size = min(read_chunk_size or 65536,
|
|
|
|
self.max_buffer_size // 2)
|
|
|
|
self.max_write_buffer_size = max_write_buffer_size
|
2014-06-11 04:34:28 -04:00
|
|
|
self.error = None
|
|
|
|
self._read_buffer = collections.deque()
|
|
|
|
self._write_buffer = collections.deque()
|
|
|
|
self._read_buffer_size = 0
|
2014-06-17 00:54:00 -04:00
|
|
|
self._write_buffer_size = 0
|
2014-06-11 04:34:28 -04:00
|
|
|
self._write_buffer_frozen = False
|
|
|
|
self._read_delimiter = None
|
|
|
|
self._read_regex = None
|
2014-06-17 00:54:00 -04:00
|
|
|
self._read_max_bytes = None
|
2014-06-11 04:34:28 -04:00
|
|
|
self._read_bytes = None
|
2014-06-17 00:54:00 -04:00
|
|
|
self._read_partial = False
|
2014-06-11 04:34:28 -04:00
|
|
|
self._read_until_close = False
|
|
|
|
self._read_callback = None
|
2014-06-17 00:54:00 -04:00
|
|
|
self._read_future = None
|
2014-06-11 04:34:28 -04:00
|
|
|
self._streaming_callback = None
|
|
|
|
self._write_callback = None
|
2014-06-17 00:54:00 -04:00
|
|
|
self._write_future = None
|
2014-06-11 04:34:28 -04:00
|
|
|
self._close_callback = None
|
|
|
|
self._connect_callback = None
|
2014-06-17 00:54:00 -04:00
|
|
|
self._connect_future = None
|
2014-06-11 04:34:28 -04:00
|
|
|
self._connecting = False
|
|
|
|
self._state = None
|
|
|
|
self._pending_callbacks = 0
|
|
|
|
self._closed = False
|
|
|
|
|
|
|
|
def fileno(self):
|
|
|
|
"""Returns the file descriptor for this stream."""
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
def close_fd(self):
|
|
|
|
"""Closes the file underlying this stream.
|
|
|
|
|
|
|
|
``close_fd`` is called by `BaseIOStream` and should not be called
|
|
|
|
elsewhere; other users should call `close` instead.
|
|
|
|
"""
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
def write_to_fd(self, data):
|
|
|
|
"""Attempts to write ``data`` to the underlying file.
|
|
|
|
|
|
|
|
Returns the number of bytes written.
|
|
|
|
"""
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
def read_from_fd(self):
|
|
|
|
"""Attempts to read from the underlying file.
|
|
|
|
|
|
|
|
Returns ``None`` if there was nothing to read (the socket
|
|
|
|
returned `~errno.EWOULDBLOCK` or equivalent), otherwise
|
|
|
|
returns the data. When possible, should return no more than
|
|
|
|
``self.read_chunk_size`` bytes at a time.
|
|
|
|
"""
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
def get_fd_error(self):
|
|
|
|
"""Returns information about any error on the underlying file.
|
|
|
|
|
|
|
|
This method is called after the `.IOLoop` has signaled an error on the
|
|
|
|
file descriptor, and should return an Exception (such as `socket.error`
|
|
|
|
with additional information, or None if no such information is
|
|
|
|
available.
|
|
|
|
"""
|
|
|
|
return None
|
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
def read_until_regex(self, regex, callback=None, max_bytes=None):
|
|
|
|
"""Asynchronously read until we have matched the given regex.
|
|
|
|
|
|
|
|
The result includes the data that matches the regex and anything
|
|
|
|
that came before it. If a callback is given, it will be run
|
|
|
|
with the data as an argument; if not, this method returns a
|
|
|
|
`.Future`.
|
|
|
|
|
|
|
|
If ``max_bytes`` is not None, the connection will be closed
|
|
|
|
if more than ``max_bytes`` bytes have been read and the regex is
|
|
|
|
not satisfied.
|
2014-06-11 04:34:28 -04:00
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
.. versionchanged:: 4.0
|
|
|
|
Added the ``max_bytes`` argument. The ``callback`` argument is
|
|
|
|
now optional and a `.Future` will be returned if it is omitted.
|
2014-06-11 04:34:28 -04:00
|
|
|
"""
|
2014-06-17 00:54:00 -04:00
|
|
|
future = self._set_read_callback(callback)
|
2014-06-11 04:34:28 -04:00
|
|
|
self._read_regex = re.compile(regex)
|
2014-06-17 00:54:00 -04:00
|
|
|
self._read_max_bytes = max_bytes
|
|
|
|
try:
|
|
|
|
self._try_inline_read()
|
|
|
|
except UnsatisfiableReadError as e:
|
|
|
|
# Handle this the same way as in _handle_events.
|
|
|
|
gen_log.info("Unsatisfiable read, closing connection: %s" % e)
|
|
|
|
self.close(exc_info=True)
|
|
|
|
return future
|
|
|
|
return future
|
|
|
|
|
|
|
|
def read_until(self, delimiter, callback=None, max_bytes=None):
|
|
|
|
"""Asynchronously read until we have found the given delimiter.
|
2014-06-11 04:34:28 -04:00
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
The result includes all the data read including the delimiter.
|
|
|
|
If a callback is given, it will be run with the data as an argument;
|
|
|
|
if not, this method returns a `.Future`.
|
2014-06-11 04:34:28 -04:00
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
If ``max_bytes`` is not None, the connection will be closed
|
|
|
|
if more than ``max_bytes`` bytes have been read and the delimiter
|
|
|
|
is not found.
|
|
|
|
|
|
|
|
.. versionchanged:: 4.0
|
|
|
|
Added the ``max_bytes`` argument. The ``callback`` argument is
|
|
|
|
now optional and a `.Future` will be returned if it is omitted.
|
2014-06-11 04:34:28 -04:00
|
|
|
"""
|
2014-06-17 00:54:00 -04:00
|
|
|
future = self._set_read_callback(callback)
|
2014-06-11 04:34:28 -04:00
|
|
|
self._read_delimiter = delimiter
|
2014-06-17 00:54:00 -04:00
|
|
|
self._read_max_bytes = max_bytes
|
|
|
|
try:
|
|
|
|
self._try_inline_read()
|
|
|
|
except UnsatisfiableReadError as e:
|
|
|
|
# Handle this the same way as in _handle_events.
|
|
|
|
gen_log.info("Unsatisfiable read, closing connection: %s" % e)
|
|
|
|
self.close(exc_info=True)
|
|
|
|
return future
|
|
|
|
return future
|
2014-06-11 04:34:28 -04:00
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
def read_bytes(self, num_bytes, callback=None, streaming_callback=None,
|
|
|
|
partial=False):
|
|
|
|
"""Asynchronously read a number of bytes.
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
If a ``streaming_callback`` is given, it will be called with chunks
|
2014-06-17 00:54:00 -04:00
|
|
|
of data as they become available, and the final result will be empty.
|
|
|
|
Otherwise, the result is all the data that was read.
|
|
|
|
If a callback is given, it will be run with the data as an argument;
|
|
|
|
if not, this method returns a `.Future`.
|
|
|
|
|
|
|
|
If ``partial`` is true, the callback is run as soon as we have
|
|
|
|
any bytes to return (but never more than ``num_bytes``)
|
|
|
|
|
|
|
|
.. versionchanged:: 4.0
|
|
|
|
Added the ``partial`` argument. The callback argument is now
|
|
|
|
optional and a `.Future` will be returned if it is omitted.
|
2014-06-11 04:34:28 -04:00
|
|
|
"""
|
2014-06-17 00:54:00 -04:00
|
|
|
future = self._set_read_callback(callback)
|
2014-06-11 04:34:28 -04:00
|
|
|
assert isinstance(num_bytes, numbers.Integral)
|
|
|
|
self._read_bytes = num_bytes
|
2014-06-17 00:54:00 -04:00
|
|
|
self._read_partial = partial
|
2014-06-11 04:34:28 -04:00
|
|
|
self._streaming_callback = stack_context.wrap(streaming_callback)
|
|
|
|
self._try_inline_read()
|
2014-06-17 00:54:00 -04:00
|
|
|
return future
|
2014-06-11 04:34:28 -04:00
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
def read_until_close(self, callback=None, streaming_callback=None):
|
|
|
|
"""Asynchronously reads all data from the socket until it is closed.
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
If a ``streaming_callback`` is given, it will be called with chunks
|
2014-06-17 00:54:00 -04:00
|
|
|
of data as they become available, and the final result will be empty.
|
|
|
|
Otherwise, the result is all the data that was read.
|
|
|
|
If a callback is given, it will be run with the data as an argument;
|
|
|
|
if not, this method returns a `.Future`.
|
|
|
|
|
|
|
|
.. versionchanged:: 4.0
|
|
|
|
The callback argument is now optional and a `.Future` will
|
|
|
|
be returned if it is omitted.
|
2014-06-11 04:34:28 -04:00
|
|
|
"""
|
2014-06-17 00:54:00 -04:00
|
|
|
future = self._set_read_callback(callback)
|
2014-06-11 04:34:28 -04:00
|
|
|
self._streaming_callback = stack_context.wrap(streaming_callback)
|
|
|
|
if self.closed():
|
|
|
|
if self._streaming_callback is not None:
|
2014-06-17 00:54:00 -04:00
|
|
|
self._run_read_callback(self._read_buffer_size, True)
|
|
|
|
self._run_read_callback(self._read_buffer_size, False)
|
|
|
|
return future
|
2014-06-11 04:34:28 -04:00
|
|
|
self._read_until_close = True
|
|
|
|
self._try_inline_read()
|
2014-06-17 00:54:00 -04:00
|
|
|
return future
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def write(self, data, callback=None):
|
2014-06-17 00:54:00 -04:00
|
|
|
"""Asynchronously write the given data to this stream.
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
If ``callback`` is given, we call it when all of the buffered write
|
|
|
|
data has been successfully written to the stream. If there was
|
|
|
|
previously buffered write data and an old write callback, that
|
|
|
|
callback is simply overwritten with this new callback.
|
2014-06-17 00:54:00 -04:00
|
|
|
|
|
|
|
If no ``callback`` is given, this method returns a `.Future` that
|
|
|
|
resolves (with a result of ``None``) when the write has been
|
|
|
|
completed. If `write` is called again before that `.Future` has
|
|
|
|
resolved, the previous future will be orphaned and will never resolve.
|
|
|
|
|
|
|
|
.. versionchanged:: 4.0
|
|
|
|
Now returns a `.Future` if no callback is given.
|
2014-06-11 04:34:28 -04:00
|
|
|
"""
|
2014-10-14 00:24:01 -04:00
|
|
|
assert isinstance(data, bytes)
|
2014-06-11 04:34:28 -04:00
|
|
|
self._check_closed()
|
|
|
|
# We use bool(_write_buffer) as a proxy for write_buffer_size>0,
|
|
|
|
# so never put empty strings in the buffer.
|
|
|
|
if data:
|
2014-06-17 00:54:00 -04:00
|
|
|
if (self.max_write_buffer_size is not None and
|
|
|
|
self._write_buffer_size + len(data) > self.max_write_buffer_size):
|
|
|
|
raise StreamBufferFullError("Reached maximum read buffer size")
|
2014-06-11 04:34:28 -04:00
|
|
|
# Break up large contiguous strings before inserting them in the
|
|
|
|
# write buffer, so we don't have to recopy the entire thing
|
|
|
|
# as we slice off pieces to send to the socket.
|
|
|
|
WRITE_BUFFER_CHUNK_SIZE = 128 * 1024
|
2014-06-17 00:54:00 -04:00
|
|
|
for i in range(0, len(data), WRITE_BUFFER_CHUNK_SIZE):
|
|
|
|
self._write_buffer.append(data[i:i + WRITE_BUFFER_CHUNK_SIZE])
|
|
|
|
self._write_buffer_size += len(data)
|
|
|
|
if callback is not None:
|
|
|
|
self._write_callback = stack_context.wrap(callback)
|
|
|
|
future = None
|
|
|
|
else:
|
|
|
|
future = self._write_future = TracebackFuture()
|
2014-06-11 04:34:28 -04:00
|
|
|
if not self._connecting:
|
|
|
|
self._handle_write()
|
|
|
|
if self._write_buffer:
|
|
|
|
self._add_io_state(self.io_loop.WRITE)
|
|
|
|
self._maybe_add_error_listener()
|
2014-06-17 00:54:00 -04:00
|
|
|
return future
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def set_close_callback(self, callback):
|
2014-06-17 00:54:00 -04:00
|
|
|
"""Call the given callback when the stream is closed.
|
|
|
|
|
|
|
|
This is not necessary for applications that use the `.Future`
|
|
|
|
interface; all outstanding ``Futures`` will resolve with a
|
|
|
|
`StreamClosedError` when the stream is closed.
|
|
|
|
"""
|
2014-06-11 04:34:28 -04:00
|
|
|
self._close_callback = stack_context.wrap(callback)
|
2014-06-17 00:54:00 -04:00
|
|
|
self._maybe_add_error_listener()
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def close(self, exc_info=False):
|
|
|
|
"""Close this stream.
|
|
|
|
|
|
|
|
If ``exc_info`` is true, set the ``error`` attribute to the current
|
|
|
|
exception from `sys.exc_info` (or if ``exc_info`` is a tuple,
|
|
|
|
use that instead of `sys.exc_info`).
|
|
|
|
"""
|
|
|
|
if not self.closed():
|
|
|
|
if exc_info:
|
|
|
|
if not isinstance(exc_info, tuple):
|
|
|
|
exc_info = sys.exc_info()
|
|
|
|
if any(exc_info):
|
|
|
|
self.error = exc_info[1]
|
|
|
|
if self._read_until_close:
|
|
|
|
if (self._streaming_callback is not None and
|
|
|
|
self._read_buffer_size):
|
2014-06-17 00:54:00 -04:00
|
|
|
self._run_read_callback(self._read_buffer_size, True)
|
2014-06-11 04:34:28 -04:00
|
|
|
self._read_until_close = False
|
2014-06-17 00:54:00 -04:00
|
|
|
self._run_read_callback(self._read_buffer_size, False)
|
2014-06-11 04:34:28 -04:00
|
|
|
if self._state is not None:
|
|
|
|
self.io_loop.remove_handler(self.fileno())
|
|
|
|
self._state = None
|
|
|
|
self.close_fd()
|
|
|
|
self._closed = True
|
|
|
|
self._maybe_run_close_callback()
|
|
|
|
|
|
|
|
def _maybe_run_close_callback(self):
|
|
|
|
# If there are pending callbacks, don't run the close callback
|
|
|
|
# until they're done (see _maybe_add_error_handler)
|
|
|
|
if self.closed() and self._pending_callbacks == 0:
|
2014-06-17 00:54:00 -04:00
|
|
|
futures = []
|
|
|
|
if self._read_future is not None:
|
|
|
|
futures.append(self._read_future)
|
|
|
|
self._read_future = None
|
|
|
|
if self._write_future is not None:
|
|
|
|
futures.append(self._write_future)
|
|
|
|
self._write_future = None
|
|
|
|
if self._connect_future is not None:
|
|
|
|
futures.append(self._connect_future)
|
|
|
|
self._connect_future = None
|
|
|
|
for future in futures:
|
|
|
|
if (isinstance(self.error, (socket.error, IOError)) and
|
|
|
|
errno_from_exception(self.error) in _ERRNO_CONNRESET):
|
|
|
|
# Treat connection resets as closed connections so
|
|
|
|
# clients only have to catch one kind of exception
|
|
|
|
# to avoid logging.
|
|
|
|
future.set_exception(StreamClosedError())
|
|
|
|
else:
|
|
|
|
future.set_exception(self.error or StreamClosedError())
|
2014-06-11 04:34:28 -04:00
|
|
|
if self._close_callback is not None:
|
|
|
|
cb = self._close_callback
|
|
|
|
self._close_callback = None
|
|
|
|
self._run_callback(cb)
|
|
|
|
# Delete any unfinished callbacks to break up reference cycles.
|
|
|
|
self._read_callback = self._write_callback = None
|
|
|
|
# Clear the buffers so they can be cleared immediately even
|
|
|
|
# if the IOStream object is kept alive by a reference cycle.
|
|
|
|
# TODO: Clear the read buffer too; it currently breaks some tests.
|
|
|
|
self._write_buffer = None
|
|
|
|
|
|
|
|
def reading(self):
|
|
|
|
"""Returns true if we are currently reading from the stream."""
|
2014-06-17 00:54:00 -04:00
|
|
|
return self._read_callback is not None or self._read_future is not None
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def writing(self):
|
|
|
|
"""Returns true if we are currently writing to the stream."""
|
|
|
|
return bool(self._write_buffer)
|
|
|
|
|
|
|
|
def closed(self):
|
|
|
|
"""Returns true if the stream has been closed."""
|
|
|
|
return self._closed
|
|
|
|
|
|
|
|
def set_nodelay(self, value):
|
|
|
|
"""Sets the no-delay flag for this stream.
|
|
|
|
|
|
|
|
By default, data written to TCP streams may be held for a time
|
|
|
|
to make the most efficient use of bandwidth (according to
|
|
|
|
Nagle's algorithm). The no-delay flag requests that data be
|
|
|
|
written as soon as possible, even if doing so would consume
|
|
|
|
additional bandwidth.
|
|
|
|
|
|
|
|
This flag is currently defined only for TCP-based ``IOStreams``.
|
|
|
|
|
|
|
|
.. versionadded:: 3.1
|
|
|
|
"""
|
|
|
|
pass
|
|
|
|
|
|
|
|
def _handle_events(self, fd, events):
|
|
|
|
if self.closed():
|
2014-06-17 00:54:00 -04:00
|
|
|
gen_log.warning("Got events for closed stream %s", fd)
|
2014-06-11 04:34:28 -04:00
|
|
|
return
|
|
|
|
try:
|
2014-06-17 00:54:00 -04:00
|
|
|
if self._connecting:
|
|
|
|
# Most IOLoops will report a write failed connect
|
|
|
|
# with the WRITE event, but SelectIOLoop reports a
|
|
|
|
# READ as well so we must check for connecting before
|
|
|
|
# either.
|
|
|
|
self._handle_connect()
|
|
|
|
if self.closed():
|
|
|
|
return
|
2014-06-11 04:34:28 -04:00
|
|
|
if events & self.io_loop.READ:
|
|
|
|
self._handle_read()
|
|
|
|
if self.closed():
|
|
|
|
return
|
|
|
|
if events & self.io_loop.WRITE:
|
|
|
|
self._handle_write()
|
|
|
|
if self.closed():
|
|
|
|
return
|
|
|
|
if events & self.io_loop.ERROR:
|
|
|
|
self.error = self.get_fd_error()
|
|
|
|
# We may have queued up a user callback in _handle_read or
|
|
|
|
# _handle_write, so don't close the IOStream until those
|
|
|
|
# callbacks have had a chance to run.
|
|
|
|
self.io_loop.add_callback(self.close)
|
|
|
|
return
|
|
|
|
state = self.io_loop.ERROR
|
|
|
|
if self.reading():
|
|
|
|
state |= self.io_loop.READ
|
|
|
|
if self.writing():
|
|
|
|
state |= self.io_loop.WRITE
|
2014-06-17 00:54:00 -04:00
|
|
|
if state == self.io_loop.ERROR and self._read_buffer_size == 0:
|
|
|
|
# If the connection is idle, listen for reads too so
|
|
|
|
# we can tell if the connection is closed. If there is
|
|
|
|
# data in the read buffer we won't run the close callback
|
|
|
|
# yet anyway, so we don't need to listen in this case.
|
2014-06-11 04:34:28 -04:00
|
|
|
state |= self.io_loop.READ
|
|
|
|
if state != self._state:
|
|
|
|
assert self._state is not None, \
|
|
|
|
"shouldn't happen: _handle_events without self._state"
|
|
|
|
self._state = state
|
|
|
|
self.io_loop.update_handler(self.fileno(), self._state)
|
2014-06-17 00:54:00 -04:00
|
|
|
except UnsatisfiableReadError as e:
|
|
|
|
gen_log.info("Unsatisfiable read, closing connection: %s" % e)
|
|
|
|
self.close(exc_info=True)
|
2014-06-11 04:34:28 -04:00
|
|
|
except Exception:
|
|
|
|
gen_log.error("Uncaught exception, closing connection.",
|
|
|
|
exc_info=True)
|
|
|
|
self.close(exc_info=True)
|
|
|
|
raise
|
|
|
|
|
|
|
|
def _run_callback(self, callback, *args):
|
|
|
|
def wrapper():
|
|
|
|
self._pending_callbacks -= 1
|
|
|
|
try:
|
2014-10-14 00:24:01 -04:00
|
|
|
return callback(*args)
|
2014-06-11 04:34:28 -04:00
|
|
|
except Exception:
|
|
|
|
app_log.error("Uncaught exception, closing connection.",
|
|
|
|
exc_info=True)
|
|
|
|
# Close the socket on an uncaught exception from a user callback
|
|
|
|
# (It would eventually get closed when the socket object is
|
|
|
|
# gc'd, but we don't want to rely on gc happening before we
|
|
|
|
# run out of file descriptors)
|
|
|
|
self.close(exc_info=True)
|
|
|
|
# Re-raise the exception so that IOLoop.handle_callback_exception
|
|
|
|
# can see it and log the error
|
|
|
|
raise
|
2014-10-14 00:24:01 -04:00
|
|
|
finally:
|
|
|
|
self._maybe_add_error_listener()
|
2014-06-11 04:34:28 -04:00
|
|
|
# We schedule callbacks to be run on the next IOLoop iteration
|
|
|
|
# rather than running them directly for several reasons:
|
|
|
|
# * Prevents unbounded stack growth when a callback calls an
|
|
|
|
# IOLoop operation that immediately runs another callback
|
|
|
|
# * Provides a predictable execution context for e.g.
|
|
|
|
# non-reentrant mutexes
|
|
|
|
# * Ensures that the try/except in wrapper() is run outside
|
|
|
|
# of the application's StackContexts
|
|
|
|
with stack_context.NullContext():
|
|
|
|
# stack_context was already captured in callback, we don't need to
|
|
|
|
# capture it again for IOStream's wrapper. This is especially
|
|
|
|
# important if the callback was pre-wrapped before entry to
|
|
|
|
# IOStream (as in HTTPConnection._header_callback), as we could
|
|
|
|
# capture and leak the wrong context here.
|
|
|
|
self._pending_callbacks += 1
|
|
|
|
self.io_loop.add_callback(wrapper)
|
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
def _read_to_buffer_loop(self):
|
|
|
|
# This method is called from _handle_read and _try_inline_read.
|
2014-06-11 04:34:28 -04:00
|
|
|
try:
|
2014-06-17 00:54:00 -04:00
|
|
|
if self._read_bytes is not None:
|
|
|
|
target_bytes = self._read_bytes
|
|
|
|
elif self._read_max_bytes is not None:
|
|
|
|
target_bytes = self._read_max_bytes
|
|
|
|
elif self.reading():
|
|
|
|
# For read_until without max_bytes, or
|
|
|
|
# read_until_close, read as much as we can before
|
|
|
|
# scanning for the delimiter.
|
|
|
|
target_bytes = None
|
|
|
|
else:
|
|
|
|
target_bytes = 0
|
|
|
|
next_find_pos = 0
|
|
|
|
# Pretend to have a pending callback so that an EOF in
|
|
|
|
# _read_to_buffer doesn't trigger an immediate close
|
|
|
|
# callback. At the end of this method we'll either
|
2014-10-14 00:24:01 -04:00
|
|
|
# establish a real pending callback via
|
2014-06-17 00:54:00 -04:00
|
|
|
# _read_from_buffer or run the close callback.
|
|
|
|
#
|
|
|
|
# We need two try statements here so that
|
|
|
|
# pending_callbacks is decremented before the `except`
|
|
|
|
# clause below (which calls `close` and does need to
|
|
|
|
# trigger the callback)
|
|
|
|
self._pending_callbacks += 1
|
|
|
|
while not self.closed():
|
|
|
|
# Read from the socket until we get EWOULDBLOCK or equivalent.
|
|
|
|
# SSL sockets do some internal buffering, and if the data is
|
|
|
|
# sitting in the SSL object's buffer select() and friends
|
|
|
|
# can't see it; the only way to find out if it's there is to
|
|
|
|
# try to read it.
|
|
|
|
if self._read_to_buffer() == 0:
|
|
|
|
break
|
|
|
|
|
|
|
|
self._run_streaming_callback()
|
|
|
|
|
|
|
|
# If we've read all the bytes we can use, break out of
|
|
|
|
# this loop. We can't just call read_from_buffer here
|
|
|
|
# because of subtle interactions with the
|
|
|
|
# pending_callback and error_listener mechanisms.
|
2014-06-11 04:34:28 -04:00
|
|
|
#
|
2014-06-17 00:54:00 -04:00
|
|
|
# If we've reached target_bytes, we know we're done.
|
|
|
|
if (target_bytes is not None and
|
|
|
|
self._read_buffer_size >= target_bytes):
|
|
|
|
break
|
|
|
|
|
|
|
|
# Otherwise, we need to call the more expensive find_read_pos.
|
|
|
|
# It's inefficient to do this on every read, so instead
|
|
|
|
# do it on the first read and whenever the read buffer
|
|
|
|
# size has doubled.
|
|
|
|
if self._read_buffer_size >= next_find_pos:
|
|
|
|
pos = self._find_read_pos()
|
|
|
|
if pos is not None:
|
|
|
|
return pos
|
|
|
|
next_find_pos = self._read_buffer_size * 2
|
|
|
|
return self._find_read_pos()
|
|
|
|
finally:
|
|
|
|
self._pending_callbacks -= 1
|
|
|
|
|
|
|
|
def _handle_read(self):
|
|
|
|
try:
|
|
|
|
pos = self._read_to_buffer_loop()
|
|
|
|
except UnsatisfiableReadError:
|
|
|
|
raise
|
2014-06-11 04:34:28 -04:00
|
|
|
except Exception:
|
|
|
|
gen_log.warning("error on read", exc_info=True)
|
|
|
|
self.close(exc_info=True)
|
|
|
|
return
|
2014-06-17 00:54:00 -04:00
|
|
|
if pos is not None:
|
|
|
|
self._read_from_buffer(pos)
|
2014-06-11 04:34:28 -04:00
|
|
|
return
|
|
|
|
else:
|
|
|
|
self._maybe_run_close_callback()
|
|
|
|
|
|
|
|
def _set_read_callback(self, callback):
|
2014-06-17 00:54:00 -04:00
|
|
|
assert self._read_callback is None, "Already reading"
|
|
|
|
assert self._read_future is None, "Already reading"
|
|
|
|
if callback is not None:
|
|
|
|
self._read_callback = stack_context.wrap(callback)
|
|
|
|
else:
|
|
|
|
self._read_future = TracebackFuture()
|
|
|
|
return self._read_future
|
|
|
|
|
|
|
|
def _run_read_callback(self, size, streaming):
|
|
|
|
if streaming:
|
|
|
|
callback = self._streaming_callback
|
|
|
|
else:
|
|
|
|
callback = self._read_callback
|
|
|
|
self._read_callback = self._streaming_callback = None
|
|
|
|
if self._read_future is not None:
|
|
|
|
assert callback is None
|
|
|
|
future = self._read_future
|
|
|
|
self._read_future = None
|
|
|
|
future.set_result(self._consume(size))
|
|
|
|
if callback is not None:
|
|
|
|
assert self._read_future is None
|
|
|
|
self._run_callback(callback, self._consume(size))
|
|
|
|
else:
|
|
|
|
# If we scheduled a callback, we will add the error listener
|
|
|
|
# afterwards. If we didn't, we have to do it now.
|
|
|
|
self._maybe_add_error_listener()
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def _try_inline_read(self):
|
|
|
|
"""Attempt to complete the current read operation from buffered data.
|
|
|
|
|
|
|
|
If the read can be completed without blocking, schedules the
|
|
|
|
read callback on the next IOLoop iteration; otherwise starts
|
|
|
|
listening for reads on the socket.
|
|
|
|
"""
|
|
|
|
# See if we've already got the data from a previous read
|
2014-06-17 00:54:00 -04:00
|
|
|
self._run_streaming_callback()
|
|
|
|
pos = self._find_read_pos()
|
|
|
|
if pos is not None:
|
|
|
|
self._read_from_buffer(pos)
|
2014-06-11 04:34:28 -04:00
|
|
|
return
|
|
|
|
self._check_closed()
|
|
|
|
try:
|
2014-06-17 00:54:00 -04:00
|
|
|
pos = self._read_to_buffer_loop()
|
2014-06-11 04:34:28 -04:00
|
|
|
except Exception:
|
|
|
|
# If there was an in _read_to_buffer, we called close() already,
|
|
|
|
# but couldn't run the close callback because of _pending_callbacks.
|
|
|
|
# Before we escape from this function, run the close callback if
|
|
|
|
# applicable.
|
|
|
|
self._maybe_run_close_callback()
|
|
|
|
raise
|
2014-06-17 00:54:00 -04:00
|
|
|
if pos is not None:
|
|
|
|
self._read_from_buffer(pos)
|
2014-06-11 04:34:28 -04:00
|
|
|
return
|
2014-06-17 00:54:00 -04:00
|
|
|
# We couldn't satisfy the read inline, so either close the stream
|
|
|
|
# or listen for new data.
|
|
|
|
if self.closed():
|
|
|
|
self._maybe_run_close_callback()
|
|
|
|
else:
|
|
|
|
self._add_io_state(ioloop.IOLoop.READ)
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def _read_to_buffer(self):
|
|
|
|
"""Reads from the socket and appends the result to the read buffer.
|
|
|
|
|
|
|
|
Returns the number of bytes read. Returns 0 if there is nothing
|
|
|
|
to read (i.e. the read returns EWOULDBLOCK or equivalent). On
|
|
|
|
error closes the socket and raises an exception.
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
chunk = self.read_from_fd()
|
|
|
|
except (socket.error, IOError, OSError) as e:
|
|
|
|
# ssl.SSLError is a subclass of socket.error
|
|
|
|
if e.args[0] in _ERRNO_CONNRESET:
|
|
|
|
# Treat ECONNRESET as a connection close rather than
|
|
|
|
# an error to minimize log spam (the exception will
|
|
|
|
# be available on self.error for apps that care).
|
|
|
|
self.close(exc_info=True)
|
|
|
|
return
|
|
|
|
self.close(exc_info=True)
|
|
|
|
raise
|
|
|
|
if chunk is None:
|
|
|
|
return 0
|
|
|
|
self._read_buffer.append(chunk)
|
|
|
|
self._read_buffer_size += len(chunk)
|
2014-06-17 00:54:00 -04:00
|
|
|
if self._read_buffer_size > self.max_buffer_size:
|
2014-06-11 04:34:28 -04:00
|
|
|
gen_log.error("Reached maximum read buffer size")
|
|
|
|
self.close()
|
2014-06-17 00:54:00 -04:00
|
|
|
raise StreamBufferFullError("Reached maximum read buffer size")
|
2014-06-11 04:34:28 -04:00
|
|
|
return len(chunk)
|
|
|
|
|
2014-06-17 00:54:00 -04:00
|
|
|
def _run_streaming_callback(self):
|
2014-06-11 04:34:28 -04:00
|
|
|
if self._streaming_callback is not None and self._read_buffer_size:
|
|
|
|
bytes_to_consume = self._read_buffer_size
|
|
|
|
if self._read_bytes is not None:
|
|
|
|
bytes_to_consume = min(self._read_bytes, bytes_to_consume)
|
|
|
|
self._read_bytes -= bytes_to_consume
|
2014-06-17 00:54:00 -04:00
|
|
|
self._run_read_callback(bytes_to_consume, True)
|
|
|
|
|
|
|
|
def _read_from_buffer(self, pos):
|
|
|
|
"""Attempts to complete the currently-pending read from the buffer.
|
|
|
|
|
|
|
|
The argument is either a position in the read buffer or None,
|
|
|
|
as returned by _find_read_pos.
|
|
|
|
"""
|
|
|
|
self._read_bytes = self._read_delimiter = self._read_regex = None
|
|
|
|
self._read_partial = False
|
|
|
|
self._run_read_callback(pos, False)
|
|
|
|
|
|
|
|
def _find_read_pos(self):
|
|
|
|
"""Attempts to find a position in the read buffer that satisfies
|
|
|
|
the currently-pending read.
|
|
|
|
|
|
|
|
Returns a position in the buffer if the current read can be satisfied,
|
|
|
|
or None if it cannot.
|
|
|
|
"""
|
|
|
|
if (self._read_bytes is not None and
|
|
|
|
(self._read_buffer_size >= self._read_bytes or
|
|
|
|
(self._read_partial and self._read_buffer_size > 0))):
|
|
|
|
num_bytes = min(self._read_bytes, self._read_buffer_size)
|
|
|
|
return num_bytes
|
2014-06-11 04:34:28 -04:00
|
|
|
elif self._read_delimiter is not None:
|
|
|
|
# Multi-byte delimiters (e.g. '\r\n') may straddle two
|
|
|
|
# chunks in the read buffer, so we can't easily find them
|
|
|
|
# without collapsing the buffer. However, since protocols
|
|
|
|
# using delimited reads (as opposed to reads of a known
|
|
|
|
# length) tend to be "line" oriented, the delimiter is likely
|
|
|
|
# to be in the first few chunks. Merge the buffer gradually
|
|
|
|
# since large merges are relatively expensive and get undone in
|
2014-06-17 00:54:00 -04:00
|
|
|
# _consume().
|
2014-06-11 04:34:28 -04:00
|
|
|
if self._read_buffer:
|
|
|
|
while True:
|
|
|
|
loc = self._read_buffer[0].find(self._read_delimiter)
|
|
|
|
if loc != -1:
|
|
|
|
delimiter_len = len(self._read_delimiter)
|
2014-06-17 00:54:00 -04:00
|
|
|
self._check_max_bytes(self._read_delimiter,
|
|
|
|
loc + delimiter_len)
|
|
|
|
return loc + delimiter_len
|
2014-06-11 04:34:28 -04:00
|
|
|
if len(self._read_buffer) == 1:
|
|
|
|
break
|
|
|
|
_double_prefix(self._read_buffer)
|
2014-06-17 00:54:00 -04:00
|
|
|
self._check_max_bytes(self._read_delimiter,
|
|
|
|
len(self._read_buffer[0]))
|
2014-06-11 04:34:28 -04:00
|
|
|
elif self._read_regex is not None:
|
|
|
|
if self._read_buffer:
|
|
|
|
while True:
|
|
|
|
m = self._read_regex.search(self._read_buffer[0])
|
|
|
|
if m is not None:
|
2014-06-17 00:54:00 -04:00
|
|
|
self._check_max_bytes(self._read_regex, m.end())
|
|
|
|
return m.end()
|
2014-06-11 04:34:28 -04:00
|
|
|
if len(self._read_buffer) == 1:
|
|
|
|
break
|
|
|
|
_double_prefix(self._read_buffer)
|
2014-06-17 00:54:00 -04:00
|
|
|
self._check_max_bytes(self._read_regex,
|
|
|
|
len(self._read_buffer[0]))
|
|
|
|
return None
|
|
|
|
|
|
|
|
def _check_max_bytes(self, delimiter, size):
|
|
|
|
if (self._read_max_bytes is not None and
|
|
|
|
size > self._read_max_bytes):
|
|
|
|
raise UnsatisfiableReadError(
|
|
|
|
"delimiter %r not found within %d bytes" % (
|
|
|
|
delimiter, self._read_max_bytes))
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def _handle_write(self):
|
|
|
|
while self._write_buffer:
|
|
|
|
try:
|
|
|
|
if not self._write_buffer_frozen:
|
|
|
|
# On windows, socket.send blows up if given a
|
|
|
|
# write buffer that's too large, instead of just
|
|
|
|
# returning the number of bytes it was able to
|
|
|
|
# process. Therefore we must not call socket.send
|
|
|
|
# with more than 128KB at a time.
|
|
|
|
_merge_prefix(self._write_buffer, 128 * 1024)
|
|
|
|
num_bytes = self.write_to_fd(self._write_buffer[0])
|
|
|
|
if num_bytes == 0:
|
|
|
|
# With OpenSSL, if we couldn't write the entire buffer,
|
|
|
|
# the very same string object must be used on the
|
|
|
|
# next call to send. Therefore we suppress
|
|
|
|
# merging the write buffer after an incomplete send.
|
|
|
|
# A cleaner solution would be to set
|
|
|
|
# SSL_MODE_ACCEPT_MOVING_WRITE_BUFFER, but this is
|
|
|
|
# not yet accessible from python
|
|
|
|
# (http://bugs.python.org/issue8240)
|
|
|
|
self._write_buffer_frozen = True
|
|
|
|
break
|
|
|
|
self._write_buffer_frozen = False
|
|
|
|
_merge_prefix(self._write_buffer, num_bytes)
|
|
|
|
self._write_buffer.popleft()
|
2014-06-17 00:54:00 -04:00
|
|
|
self._write_buffer_size -= num_bytes
|
2014-06-11 04:34:28 -04:00
|
|
|
except (socket.error, IOError, OSError) as e:
|
|
|
|
if e.args[0] in _ERRNO_WOULDBLOCK:
|
|
|
|
self._write_buffer_frozen = True
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
if e.args[0] not in _ERRNO_CONNRESET:
|
|
|
|
# Broken pipe errors are usually caused by connection
|
|
|
|
# reset, and its better to not log EPIPE errors to
|
|
|
|
# minimize log spam
|
2014-06-17 00:54:00 -04:00
|
|
|
gen_log.warning("Write error on %s: %s",
|
2014-06-11 04:34:28 -04:00
|
|
|
self.fileno(), e)
|
|
|
|
self.close(exc_info=True)
|
|
|
|
return
|
2014-06-17 00:54:00 -04:00
|
|
|
if not self._write_buffer:
|
|
|
|
if self._write_callback:
|
|
|
|
callback = self._write_callback
|
|
|
|
self._write_callback = None
|
|
|
|
self._run_callback(callback)
|
|
|
|
if self._write_future:
|
|
|
|
future = self._write_future
|
|
|
|
self._write_future = None
|
|
|
|
future.set_result(None)
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def _consume(self, loc):
|
|
|
|
if loc == 0:
|
|
|
|
return b""
|
|
|
|
_merge_prefix(self._read_buffer, loc)
|
|
|
|
self._read_buffer_size -= loc
|
|
|
|
return self._read_buffer.popleft()
|
|
|
|
|
|
|
|
def _check_closed(self):
|
|
|
|
if self.closed():
|
|
|
|
raise StreamClosedError("Stream is closed")
|
|
|
|
|
|
|
|
def _maybe_add_error_listener(self):
|
2014-06-17 00:54:00 -04:00
|
|
|
# This method is part of an optimization: to detect a connection that
|
|
|
|
# is closed when we're not actively reading or writing, we must listen
|
|
|
|
# for read events. However, it is inefficient to do this when the
|
|
|
|
# connection is first established because we are going to read or write
|
|
|
|
# immediately anyway. Instead, we insert checks at various times to
|
|
|
|
# see if the connection is idle and add the read listener then.
|
|
|
|
if self._pending_callbacks != 0:
|
|
|
|
return
|
|
|
|
if self._state is None or self._state == ioloop.IOLoop.ERROR:
|
2014-06-11 04:34:28 -04:00
|
|
|
if self.closed():
|
|
|
|
self._maybe_run_close_callback()
|
2014-06-17 00:54:00 -04:00
|
|
|
elif (self._read_buffer_size == 0 and
|
|
|
|
self._close_callback is not None):
|
2014-06-11 04:34:28 -04:00
|
|
|
self._add_io_state(ioloop.IOLoop.READ)
|
|
|
|
|
|
|
|
def _add_io_state(self, state):
|
|
|
|
"""Adds `state` (IOLoop.{READ,WRITE} flags) to our event handler.
|
|
|
|
|
|
|
|
Implementation notes: Reads and writes have a fast path and a
|
|
|
|
slow path. The fast path reads synchronously from socket
|
|
|
|
buffers, while the slow path uses `_add_io_state` to schedule
|
|
|
|
an IOLoop callback. Note that in both cases, the callback is
|
|
|
|
run asynchronously with `_run_callback`.
|
|
|
|
|
|
|
|
To detect closed connections, we must have called
|
|
|
|
`_add_io_state` at some point, but we want to delay this as
|
|
|
|
much as possible so we don't have to set an `IOLoop.ERROR`
|
|
|
|
listener that will be overwritten by the next slow-path
|
|
|
|
operation. As long as there are callbacks scheduled for
|
|
|
|
fast-path ops, those callbacks may do more reads.
|
|
|
|
If a sequence of fast-path ops do not end in a slow-path op,
|
|
|
|
(e.g. for an @asynchronous long-poll request), we must add
|
|
|
|
the error handler. This is done in `_run_callback` and `write`
|
|
|
|
(since the write callback is optional so we can have a
|
|
|
|
fast-path write with no `_run_callback`)
|
|
|
|
"""
|
|
|
|
if self.closed():
|
|
|
|
# connection has been closed, so there can be no future events
|
|
|
|
return
|
|
|
|
if self._state is None:
|
|
|
|
self._state = ioloop.IOLoop.ERROR | state
|
|
|
|
with stack_context.NullContext():
|
|
|
|
self.io_loop.add_handler(
|
|
|
|
self.fileno(), self._handle_events, self._state)
|
|
|
|
elif not self._state & state:
|
|
|
|
self._state = self._state | state
|
|
|
|
self.io_loop.update_handler(self.fileno(), self._state)
|
|
|
|
|
|
|
|
|
|
|
|
class IOStream(BaseIOStream):
|
|
|
|
r"""Socket-based `IOStream` implementation.
|
|
|
|
|
|
|
|
This class supports the read and write methods from `BaseIOStream`
|
|
|
|
plus a `connect` method.
|
|
|
|
|
|
|
|
The ``socket`` parameter may either be connected or unconnected.
|
|
|
|
For server operations the socket is the result of calling
|
|
|
|
`socket.accept <socket.socket.accept>`. For client operations the
|
|
|
|
socket is created with `socket.socket`, and may either be
|
|
|
|
connected before passing it to the `IOStream` or connected with
|
|
|
|
`IOStream.connect`.
|
|
|
|
|
|
|
|
A very simple (and broken) HTTP client using this class::
|
|
|
|
|
|
|
|
import tornado.ioloop
|
|
|
|
import tornado.iostream
|
|
|
|
import socket
|
|
|
|
|
|
|
|
def send_request():
|
|
|
|
stream.write(b"GET / HTTP/1.0\r\nHost: friendfeed.com\r\n\r\n")
|
|
|
|
stream.read_until(b"\r\n\r\n", on_headers)
|
|
|
|
|
|
|
|
def on_headers(data):
|
|
|
|
headers = {}
|
|
|
|
for line in data.split(b"\r\n"):
|
|
|
|
parts = line.split(b":")
|
|
|
|
if len(parts) == 2:
|
|
|
|
headers[parts[0].strip()] = parts[1].strip()
|
|
|
|
stream.read_bytes(int(headers[b"Content-Length"]), on_body)
|
|
|
|
|
|
|
|
def on_body(data):
|
|
|
|
print data
|
|
|
|
stream.close()
|
|
|
|
tornado.ioloop.IOLoop.instance().stop()
|
|
|
|
|
|
|
|
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM, 0)
|
|
|
|
stream = tornado.iostream.IOStream(s)
|
|
|
|
stream.connect(("friendfeed.com", 80), send_request)
|
|
|
|
tornado.ioloop.IOLoop.instance().start()
|
|
|
|
"""
|
|
|
|
def __init__(self, socket, *args, **kwargs):
|
|
|
|
self.socket = socket
|
|
|
|
self.socket.setblocking(False)
|
|
|
|
super(IOStream, self).__init__(*args, **kwargs)
|
|
|
|
|
|
|
|
def fileno(self):
|
2014-06-17 00:54:00 -04:00
|
|
|
return self.socket
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def close_fd(self):
|
|
|
|
self.socket.close()
|
|
|
|
self.socket = None
|
|
|
|
|
|
|
|
def get_fd_error(self):
|
|
|
|
errno = self.socket.getsockopt(socket.SOL_SOCKET,
|
|
|
|
socket.SO_ERROR)
|
|
|
|
return socket.error(errno, os.strerror(errno))
|
|
|
|
|
|
|
|
def read_from_fd(self):
|
|
|
|
try:
|
|
|
|
chunk = self.socket.recv(self.read_chunk_size)
|
|
|
|
except socket.error as e:
|
|
|
|
if e.args[0] in _ERRNO_WOULDBLOCK:
|
|
|
|
return None
|
|
|
|
else:
|
|
|
|
raise
|
|
|
|
if not chunk:
|
|
|
|
self.close()
|
|
|
|
return None
|
|
|
|
return chunk
|
|
|
|
|
|
|
|
def write_to_fd(self, data):
|
|
|
|
return self.socket.send(data)
|
|
|
|
|
|
|
|
def connect(self, address, callback=None, server_hostname=None):
|
|
|
|
"""Connects the socket to a remote address without blocking.
|
|
|
|
|
|
|
|
May only be called if the socket passed to the constructor was
|
|
|
|
not previously connected. The address parameter is in the
|
2014-06-17 00:54:00 -04:00
|
|
|
same format as for `socket.connect <socket.socket.connect>` for
|
|
|
|
the type of socket passed to the IOStream constructor,
|
|
|
|
e.g. an ``(ip, port)`` tuple. Hostnames are accepted here,
|
|
|
|
but will be resolved synchronously and block the IOLoop.
|
|
|
|
If you have a hostname instead of an IP address, the `.TCPClient`
|
|
|
|
class is recommended instead of calling this method directly.
|
|
|
|
`.TCPClient` will do asynchronous DNS resolution and handle
|
|
|
|
both IPv4 and IPv6.
|
|
|
|
|
|
|
|
If ``callback`` is specified, it will be called with no
|
|
|
|
arguments when the connection is completed; if not this method
|
|
|
|
returns a `.Future` (whose result after a successful
|
|
|
|
connection will be the stream itself).
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
If specified, the ``server_hostname`` parameter will be used
|
|
|
|
in SSL connections for certificate validation (if requested in
|
|
|
|
the ``ssl_options``) and SNI (if supported; requires
|
|
|
|
Python 3.2+).
|
|
|
|
|
|
|
|
Note that it is safe to call `IOStream.write
|
|
|
|
<BaseIOStream.write>` while the connection is pending, in
|
|
|
|
which case the data will be written as soon as the connection
|
|
|
|
is ready. Calling `IOStream` read methods before the socket is
|
|
|
|
connected works on some platforms but is non-portable.
|
2014-06-17 00:54:00 -04:00
|
|
|
|
|
|
|
.. versionchanged:: 4.0
|
|
|
|
If no callback is given, returns a `.Future`.
|
|
|
|
|
2014-06-11 04:34:28 -04:00
|
|
|
"""
|
|
|
|
self._connecting = True
|
2014-10-14 00:24:01 -04:00
|
|
|
if callback is not None:
|
|
|
|
self._connect_callback = stack_context.wrap(callback)
|
|
|
|
future = None
|
|
|
|
else:
|
|
|
|
future = self._connect_future = TracebackFuture()
|
2014-06-11 04:34:28 -04:00
|
|
|
try:
|
|
|
|
self.socket.connect(address)
|
|
|
|
except socket.error as e:
|
|
|
|
# In non-blocking mode we expect connect() to raise an
|
|
|
|
# exception with EINPROGRESS or EWOULDBLOCK.
|
|
|
|
#
|
|
|
|
# On freebsd, other errors such as ECONNREFUSED may be
|
|
|
|
# returned immediately when attempting to connect to
|
|
|
|
# localhost, so handle them the same way as an error
|
|
|
|
# reported later in _handle_connect.
|
2014-06-29 06:05:33 -04:00
|
|
|
if (errno_from_exception(e) not in _ERRNO_INPROGRESS and
|
2014-06-17 00:54:00 -04:00
|
|
|
errno_from_exception(e) not in _ERRNO_WOULDBLOCK):
|
|
|
|
gen_log.warning("Connect error on fd %s: %s",
|
2014-06-11 04:34:28 -04:00
|
|
|
self.socket.fileno(), e)
|
|
|
|
self.close(exc_info=True)
|
2014-10-14 00:24:01 -04:00
|
|
|
return future
|
2014-06-11 04:34:28 -04:00
|
|
|
self._add_io_state(self.io_loop.WRITE)
|
2014-06-17 00:54:00 -04:00
|
|
|
return future
|
|
|
|
|
|
|
|
def start_tls(self, server_side, ssl_options=None, server_hostname=None):
|
|
|
|
"""Convert this `IOStream` to an `SSLIOStream`.
|
|
|
|
|
|
|
|
This enables protocols that begin in clear-text mode and
|
|
|
|
switch to SSL after some initial negotiation (such as the
|
|
|
|
``STARTTLS`` extension to SMTP and IMAP).
|
|
|
|
|
|
|
|
This method cannot be used if there are outstanding reads
|
|
|
|
or writes on the stream, or if there is any data in the
|
|
|
|
IOStream's buffer (data in the operating system's socket
|
|
|
|
buffer is allowed). This means it must generally be used
|
|
|
|
immediately after reading or writing the last clear-text
|
|
|
|
data. It can also be used immediately after connecting,
|
|
|
|
before any reads or writes.
|
|
|
|
|
|
|
|
The ``ssl_options`` argument may be either a dictionary
|
|
|
|
of options or an `ssl.SSLContext`. If a ``server_hostname``
|
|
|
|
is given, it will be used for certificate verification
|
|
|
|
(as configured in the ``ssl_options``).
|
|
|
|
|
|
|
|
This method returns a `.Future` whose result is the new
|
|
|
|
`SSLIOStream`. After this method has been called,
|
|
|
|
any other operation on the original stream is undefined.
|
|
|
|
|
|
|
|
If a close callback is defined on this stream, it will be
|
|
|
|
transferred to the new stream.
|
|
|
|
|
|
|
|
.. versionadded:: 4.0
|
|
|
|
"""
|
|
|
|
if (self._read_callback or self._read_future or
|
|
|
|
self._write_callback or self._write_future or
|
|
|
|
self._connect_callback or self._connect_future or
|
|
|
|
self._pending_callbacks or self._closed or
|
|
|
|
self._read_buffer or self._write_buffer):
|
|
|
|
raise ValueError("IOStream is not idle; cannot convert to SSL")
|
|
|
|
if ssl_options is None:
|
|
|
|
ssl_options = {}
|
|
|
|
|
|
|
|
socket = self.socket
|
|
|
|
self.io_loop.remove_handler(socket)
|
|
|
|
self.socket = None
|
|
|
|
socket = ssl_wrap_socket(socket, ssl_options, server_side=server_side,
|
|
|
|
do_handshake_on_connect=False)
|
|
|
|
orig_close_callback = self._close_callback
|
|
|
|
self._close_callback = None
|
|
|
|
|
|
|
|
future = TracebackFuture()
|
|
|
|
ssl_stream = SSLIOStream(socket, ssl_options=ssl_options,
|
|
|
|
io_loop=self.io_loop)
|
|
|
|
# Wrap the original close callback so we can fail our Future as well.
|
|
|
|
# If we had an "unwrap" counterpart to this method we would need
|
|
|
|
# to restore the original callback after our Future resolves
|
|
|
|
# so that repeated wrap/unwrap calls don't build up layers.
|
|
|
|
def close_callback():
|
|
|
|
if not future.done():
|
|
|
|
future.set_exception(ssl_stream.error or StreamClosedError())
|
|
|
|
if orig_close_callback is not None:
|
|
|
|
orig_close_callback()
|
|
|
|
ssl_stream.set_close_callback(close_callback)
|
|
|
|
ssl_stream._ssl_connect_callback = lambda: future.set_result(ssl_stream)
|
|
|
|
ssl_stream.max_buffer_size = self.max_buffer_size
|
|
|
|
ssl_stream.read_chunk_size = self.read_chunk_size
|
|
|
|
return future
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def _handle_connect(self):
|
|
|
|
err = self.socket.getsockopt(socket.SOL_SOCKET, socket.SO_ERROR)
|
|
|
|
if err != 0:
|
|
|
|
self.error = socket.error(err, os.strerror(err))
|
|
|
|
# IOLoop implementations may vary: some of them return
|
|
|
|
# an error state before the socket becomes writable, so
|
|
|
|
# in that case a connection failure would be handled by the
|
|
|
|
# error path in _handle_events instead of here.
|
2014-06-17 00:54:00 -04:00
|
|
|
if self._connect_future is None:
|
|
|
|
gen_log.warning("Connect error on fd %s: %s",
|
|
|
|
self.socket.fileno(), errno.errorcode[err])
|
2014-06-11 04:34:28 -04:00
|
|
|
self.close()
|
|
|
|
return
|
|
|
|
if self._connect_callback is not None:
|
|
|
|
callback = self._connect_callback
|
|
|
|
self._connect_callback = None
|
|
|
|
self._run_callback(callback)
|
2014-06-17 00:54:00 -04:00
|
|
|
if self._connect_future is not None:
|
|
|
|
future = self._connect_future
|
|
|
|
self._connect_future = None
|
|
|
|
future.set_result(self)
|
2014-06-11 04:34:28 -04:00
|
|
|
self._connecting = False
|
|
|
|
|
|
|
|
def set_nodelay(self, value):
|
|
|
|
if (self.socket is not None and
|
|
|
|
self.socket.family in (socket.AF_INET, socket.AF_INET6)):
|
|
|
|
try:
|
|
|
|
self.socket.setsockopt(socket.IPPROTO_TCP,
|
|
|
|
socket.TCP_NODELAY, 1 if value else 0)
|
|
|
|
except socket.error as e:
|
|
|
|
# Sometimes setsockopt will fail if the socket is closed
|
|
|
|
# at the wrong time. This can happen with HTTPServer
|
|
|
|
# resetting the value to false between requests.
|
|
|
|
if e.errno not in (errno.EINVAL, errno.ECONNRESET):
|
|
|
|
raise
|
|
|
|
|
|
|
|
|
|
|
|
class SSLIOStream(IOStream):
|
|
|
|
"""A utility class to write to and read from a non-blocking SSL socket.
|
|
|
|
|
|
|
|
If the socket passed to the constructor is already connected,
|
|
|
|
it should be wrapped with::
|
|
|
|
|
|
|
|
ssl.wrap_socket(sock, do_handshake_on_connect=False, **kwargs)
|
|
|
|
|
|
|
|
before constructing the `SSLIOStream`. Unconnected sockets will be
|
|
|
|
wrapped when `IOStream.connect` is finished.
|
|
|
|
"""
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
|
|
"""The ``ssl_options`` keyword argument may either be a dictionary
|
|
|
|
of keywords arguments for `ssl.wrap_socket`, or an `ssl.SSLContext`
|
|
|
|
object.
|
|
|
|
"""
|
|
|
|
self._ssl_options = kwargs.pop('ssl_options', {})
|
|
|
|
super(SSLIOStream, self).__init__(*args, **kwargs)
|
|
|
|
self._ssl_accepting = True
|
|
|
|
self._handshake_reading = False
|
|
|
|
self._handshake_writing = False
|
|
|
|
self._ssl_connect_callback = None
|
|
|
|
self._server_hostname = None
|
|
|
|
|
|
|
|
# If the socket is already connected, attempt to start the handshake.
|
|
|
|
try:
|
|
|
|
self.socket.getpeername()
|
|
|
|
except socket.error:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
# Indirectly start the handshake, which will run on the next
|
|
|
|
# IOLoop iteration and then the real IO state will be set in
|
|
|
|
# _handle_events.
|
|
|
|
self._add_io_state(self.io_loop.WRITE)
|
|
|
|
|
|
|
|
def reading(self):
|
|
|
|
return self._handshake_reading or super(SSLIOStream, self).reading()
|
|
|
|
|
|
|
|
def writing(self):
|
|
|
|
return self._handshake_writing or super(SSLIOStream, self).writing()
|
|
|
|
|
|
|
|
def _do_ssl_handshake(self):
|
|
|
|
# Based on code from test_ssl.py in the python stdlib
|
|
|
|
try:
|
|
|
|
self._handshake_reading = False
|
|
|
|
self._handshake_writing = False
|
|
|
|
self.socket.do_handshake()
|
|
|
|
except ssl.SSLError as err:
|
|
|
|
if err.args[0] == ssl.SSL_ERROR_WANT_READ:
|
|
|
|
self._handshake_reading = True
|
|
|
|
return
|
|
|
|
elif err.args[0] == ssl.SSL_ERROR_WANT_WRITE:
|
|
|
|
self._handshake_writing = True
|
|
|
|
return
|
|
|
|
elif err.args[0] in (ssl.SSL_ERROR_EOF,
|
|
|
|
ssl.SSL_ERROR_ZERO_RETURN):
|
|
|
|
return self.close(exc_info=True)
|
|
|
|
elif err.args[0] == ssl.SSL_ERROR_SSL:
|
|
|
|
try:
|
|
|
|
peer = self.socket.getpeername()
|
|
|
|
except Exception:
|
|
|
|
peer = '(not connected)'
|
2014-06-17 00:54:00 -04:00
|
|
|
gen_log.warning("SSL Error on %s %s: %s",
|
2014-06-11 04:34:28 -04:00
|
|
|
self.socket.fileno(), peer, err)
|
|
|
|
return self.close(exc_info=True)
|
|
|
|
raise
|
|
|
|
except socket.error as err:
|
2014-10-14 00:24:01 -04:00
|
|
|
# Some port scans (e.g. nmap in -sT mode) have been known
|
|
|
|
# to cause do_handshake to raise EBADF, so make that error
|
|
|
|
# quiet as well.
|
|
|
|
# https://groups.google.com/forum/?fromgroups#!topic/python-tornado/ApucKJat1_0
|
|
|
|
if (err.args[0] in _ERRNO_CONNRESET or
|
|
|
|
err.args[0] == errno.EBADF):
|
2014-06-11 04:34:28 -04:00
|
|
|
return self.close(exc_info=True)
|
2014-10-14 00:24:01 -04:00
|
|
|
raise
|
2014-06-11 04:34:28 -04:00
|
|
|
except AttributeError:
|
|
|
|
# On Linux, if the connection was reset before the call to
|
|
|
|
# wrap_socket, do_handshake will fail with an
|
|
|
|
# AttributeError.
|
|
|
|
return self.close(exc_info=True)
|
|
|
|
else:
|
|
|
|
self._ssl_accepting = False
|
|
|
|
if not self._verify_cert(self.socket.getpeercert()):
|
|
|
|
self.close()
|
|
|
|
return
|
|
|
|
if self._ssl_connect_callback is not None:
|
|
|
|
callback = self._ssl_connect_callback
|
|
|
|
self._ssl_connect_callback = None
|
|
|
|
self._run_callback(callback)
|
|
|
|
|
|
|
|
def _verify_cert(self, peercert):
|
|
|
|
"""Returns True if peercert is valid according to the configured
|
|
|
|
validation mode and hostname.
|
|
|
|
|
|
|
|
The ssl handshake already tested the certificate for a valid
|
|
|
|
CA signature; the only thing that remains is to check
|
|
|
|
the hostname.
|
|
|
|
"""
|
|
|
|
if isinstance(self._ssl_options, dict):
|
|
|
|
verify_mode = self._ssl_options.get('cert_reqs', ssl.CERT_NONE)
|
|
|
|
elif isinstance(self._ssl_options, ssl.SSLContext):
|
|
|
|
verify_mode = self._ssl_options.verify_mode
|
|
|
|
assert verify_mode in (ssl.CERT_NONE, ssl.CERT_REQUIRED, ssl.CERT_OPTIONAL)
|
|
|
|
if verify_mode == ssl.CERT_NONE or self._server_hostname is None:
|
|
|
|
return True
|
|
|
|
cert = self.socket.getpeercert()
|
|
|
|
if cert is None and verify_mode == ssl.CERT_REQUIRED:
|
|
|
|
gen_log.warning("No SSL certificate given")
|
|
|
|
return False
|
|
|
|
try:
|
|
|
|
ssl_match_hostname(peercert, self._server_hostname)
|
|
|
|
except SSLCertificateError:
|
|
|
|
gen_log.warning("Invalid SSL certificate", exc_info=True)
|
|
|
|
return False
|
|
|
|
else:
|
|
|
|
return True
|
|
|
|
|
|
|
|
def _handle_read(self):
|
|
|
|
if self._ssl_accepting:
|
|
|
|
self._do_ssl_handshake()
|
|
|
|
return
|
|
|
|
super(SSLIOStream, self)._handle_read()
|
|
|
|
|
|
|
|
def _handle_write(self):
|
|
|
|
if self._ssl_accepting:
|
|
|
|
self._do_ssl_handshake()
|
|
|
|
return
|
|
|
|
super(SSLIOStream, self)._handle_write()
|
|
|
|
|
|
|
|
def connect(self, address, callback=None, server_hostname=None):
|
|
|
|
# Save the user's callback and run it after the ssl handshake
|
|
|
|
# has completed.
|
|
|
|
self._ssl_connect_callback = stack_context.wrap(callback)
|
|
|
|
self._server_hostname = server_hostname
|
2014-06-17 00:54:00 -04:00
|
|
|
# Note: Since we don't pass our callback argument along to
|
|
|
|
# super.connect(), this will always return a Future.
|
|
|
|
# This is harmless, but a bit less efficient than it could be.
|
|
|
|
return super(SSLIOStream, self).connect(address, callback=None)
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def _handle_connect(self):
|
2014-06-17 00:54:00 -04:00
|
|
|
# Call the superclass method to check for errors.
|
|
|
|
super(SSLIOStream, self)._handle_connect()
|
|
|
|
if self.closed():
|
|
|
|
return
|
2014-06-11 04:34:28 -04:00
|
|
|
# When the connection is complete, wrap the socket for SSL
|
|
|
|
# traffic. Note that we do this by overriding _handle_connect
|
|
|
|
# instead of by passing a callback to super().connect because
|
|
|
|
# user callbacks are enqueued asynchronously on the IOLoop,
|
|
|
|
# but since _handle_events calls _handle_connect immediately
|
|
|
|
# followed by _handle_write we need this to be synchronous.
|
2014-06-17 00:54:00 -04:00
|
|
|
#
|
|
|
|
# The IOLoop will get confused if we swap out self.socket while the
|
|
|
|
# fd is registered, so remove it now and re-register after
|
|
|
|
# wrap_socket().
|
|
|
|
self.io_loop.remove_handler(self.socket)
|
|
|
|
old_state = self._state
|
|
|
|
self._state = None
|
2014-06-11 04:34:28 -04:00
|
|
|
self.socket = ssl_wrap_socket(self.socket, self._ssl_options,
|
|
|
|
server_hostname=self._server_hostname,
|
|
|
|
do_handshake_on_connect=False)
|
2014-06-17 00:54:00 -04:00
|
|
|
self._add_io_state(old_state)
|
2014-06-11 04:34:28 -04:00
|
|
|
|
|
|
|
def read_from_fd(self):
|
|
|
|
if self._ssl_accepting:
|
|
|
|
# If the handshake hasn't finished yet, there can't be anything
|
|
|
|
# to read (attempting to read may or may not raise an exception
|
|
|
|
# depending on the SSL version)
|
|
|
|
return None
|
|
|
|
try:
|
|
|
|
# SSLSocket objects have both a read() and recv() method,
|
|
|
|
# while regular sockets only have recv().
|
|
|
|
# The recv() method blocks (at least in python 2.6) if it is
|
|
|
|
# called when there is nothing to read, so we have to use
|
|
|
|
# read() instead.
|
|
|
|
chunk = self.socket.read(self.read_chunk_size)
|
|
|
|
except ssl.SSLError as e:
|
|
|
|
# SSLError is a subclass of socket.error, so this except
|
|
|
|
# block must come first.
|
|
|
|
if e.args[0] == ssl.SSL_ERROR_WANT_READ:
|
|
|
|
return None
|
|
|
|
else:
|
|
|
|
raise
|
|
|
|
except socket.error as e:
|
|
|
|
if e.args[0] in _ERRNO_WOULDBLOCK:
|
|
|
|
return None
|
|
|
|
else:
|
|
|
|
raise
|
|
|
|
if not chunk:
|
|
|
|
self.close()
|
|
|
|
return None
|
|
|
|
return chunk
|
|
|
|
|
|
|
|
|
|
|
|
class PipeIOStream(BaseIOStream):
|
|
|
|
"""Pipe-based `IOStream` implementation.
|
|
|
|
|
|
|
|
The constructor takes an integer file descriptor (such as one returned
|
|
|
|
by `os.pipe`) rather than an open file object. Pipes are generally
|
|
|
|
one-way, so a `PipeIOStream` can be used for reading or writing but not
|
|
|
|
both.
|
|
|
|
"""
|
|
|
|
def __init__(self, fd, *args, **kwargs):
|
|
|
|
self.fd = fd
|
|
|
|
_set_nonblocking(fd)
|
|
|
|
super(PipeIOStream, self).__init__(*args, **kwargs)
|
|
|
|
|
|
|
|
def fileno(self):
|
|
|
|
return self.fd
|
|
|
|
|
|
|
|
def close_fd(self):
|
|
|
|
os.close(self.fd)
|
|
|
|
|
|
|
|
def write_to_fd(self, data):
|
|
|
|
return os.write(self.fd, data)
|
|
|
|
|
|
|
|
def read_from_fd(self):
|
|
|
|
try:
|
|
|
|
chunk = os.read(self.fd, self.read_chunk_size)
|
|
|
|
except (IOError, OSError) as e:
|
2014-06-17 00:54:00 -04:00
|
|
|
if errno_from_exception(e) in _ERRNO_WOULDBLOCK:
|
2014-06-11 04:34:28 -04:00
|
|
|
return None
|
2014-06-17 00:54:00 -04:00
|
|
|
elif errno_from_exception(e) == errno.EBADF:
|
2014-06-11 04:34:28 -04:00
|
|
|
# If the writing half of a pipe is closed, select will
|
|
|
|
# report it as readable but reads will fail with EBADF.
|
|
|
|
self.close(exc_info=True)
|
|
|
|
return None
|
|
|
|
else:
|
|
|
|
raise
|
|
|
|
if not chunk:
|
|
|
|
self.close()
|
|
|
|
return None
|
|
|
|
return chunk
|
|
|
|
|
|
|
|
|
|
|
|
def _double_prefix(deque):
|
|
|
|
"""Grow by doubling, but don't split the second chunk just because the
|
|
|
|
first one is small.
|
|
|
|
"""
|
|
|
|
new_len = max(len(deque[0]) * 2,
|
|
|
|
(len(deque[0]) + len(deque[1])))
|
|
|
|
_merge_prefix(deque, new_len)
|
|
|
|
|
|
|
|
|
|
|
|
def _merge_prefix(deque, size):
|
|
|
|
"""Replace the first entries in a deque of strings with a single
|
|
|
|
string of up to size bytes.
|
|
|
|
|
|
|
|
>>> d = collections.deque(['abc', 'de', 'fghi', 'j'])
|
|
|
|
>>> _merge_prefix(d, 5); print(d)
|
|
|
|
deque(['abcde', 'fghi', 'j'])
|
|
|
|
|
|
|
|
Strings will be split as necessary to reach the desired size.
|
|
|
|
>>> _merge_prefix(d, 7); print(d)
|
|
|
|
deque(['abcdefg', 'hi', 'j'])
|
|
|
|
|
|
|
|
>>> _merge_prefix(d, 3); print(d)
|
|
|
|
deque(['abc', 'defg', 'hi', 'j'])
|
|
|
|
|
|
|
|
>>> _merge_prefix(d, 100); print(d)
|
|
|
|
deque(['abcdefghij'])
|
|
|
|
"""
|
|
|
|
if len(deque) == 1 and len(deque[0]) <= size:
|
|
|
|
return
|
|
|
|
prefix = []
|
|
|
|
remaining = size
|
|
|
|
while deque and remaining > 0:
|
|
|
|
chunk = deque.popleft()
|
|
|
|
if len(chunk) > remaining:
|
|
|
|
deque.appendleft(chunk[remaining:])
|
|
|
|
chunk = chunk[:remaining]
|
|
|
|
prefix.append(chunk)
|
|
|
|
remaining -= len(chunk)
|
|
|
|
# This data structure normally just contains byte strings, but
|
|
|
|
# the unittest gets messy if it doesn't use the default str() type,
|
|
|
|
# so do the merge based on the type of data that's actually present.
|
|
|
|
if prefix:
|
|
|
|
deque.appendleft(type(prefix[0])().join(prefix))
|
|
|
|
if not deque:
|
|
|
|
deque.appendleft(b"")
|
|
|
|
|
|
|
|
|
|
|
|
def doctests():
|
|
|
|
import doctest
|
|
|
|
return doctest.DocTestSuite()
|