2014-03-10 21:36:47 +00:00
|
|
|
from __future__ import absolute_import
|
2014-01-30 17:56:23 +00:00
|
|
|
import Cookie, urllib, urlparse, time, copy
|
2014-01-30 04:00:13 +00:00
|
|
|
from email.utils import parsedate_tz, formatdate, mktime_tz
|
2014-03-09 20:13:08 +00:00
|
|
|
from netlib import http, tcp, http_status
|
2014-03-10 21:36:47 +00:00
|
|
|
import netlib.utils
|
2014-01-30 17:56:23 +00:00
|
|
|
from netlib.odict import ODict, ODictCaseless
|
2014-03-10 20:57:50 +00:00
|
|
|
from .primitives import KILL, ProtocolHandler, TemporaryServerChangeMixin, Flow, Error
|
|
|
|
from ..proxy.connection import ServerConnection
|
|
|
|
from .. import encoding, utils, filt, controller, stateobject, proxy
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
HDR_FORM_URLENCODED = "application/x-www-form-urlencoded"
|
|
|
|
CONTENT_MISSING = 0
|
|
|
|
|
|
|
|
|
|
|
|
def get_line(fp):
|
|
|
|
"""
|
|
|
|
Get a line, possibly preceded by a blank.
|
|
|
|
"""
|
|
|
|
line = fp.readline()
|
|
|
|
if line == "\r\n" or line == "\n": # Possible leftover from previous message
|
|
|
|
line = fp.readline()
|
|
|
|
if line == "":
|
|
|
|
raise tcp.NetLibDisconnect
|
|
|
|
return line
|
|
|
|
|
|
|
|
|
|
|
|
class decoded(object):
|
|
|
|
"""
|
|
|
|
A context manager that decodes a request or response, and then
|
|
|
|
re-encodes it with the same encoding after execution of the block.
|
|
|
|
|
|
|
|
Example:
|
|
|
|
with decoded(request):
|
|
|
|
request.content = request.content.replace("foo", "bar")
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, o):
|
|
|
|
self.o = o
|
|
|
|
ce = o.headers.get_first("content-encoding")
|
|
|
|
if ce in encoding.ENCODINGS:
|
|
|
|
self.ce = ce
|
|
|
|
else:
|
|
|
|
self.ce = None
|
|
|
|
|
|
|
|
def __enter__(self):
|
|
|
|
if self.ce:
|
|
|
|
self.o.decode()
|
|
|
|
|
|
|
|
def __exit__(self, type, value, tb):
|
|
|
|
if self.ce:
|
|
|
|
self.o.encode(self.ce)
|
|
|
|
|
2014-01-30 05:03:41 +00:00
|
|
|
|
2014-01-30 17:56:23 +00:00
|
|
|
class HTTPMessage(stateobject.SimpleStateObject):
|
2014-05-15 16:16:42 +00:00
|
|
|
def __init__(self, httpversion, headers, content, timestamp_start=None,
|
|
|
|
timestamp_end=None):
|
2014-02-05 19:26:47 +00:00
|
|
|
self.httpversion = httpversion
|
|
|
|
self.headers = headers
|
2014-02-07 02:56:57 +00:00
|
|
|
"""@type: ODictCaseless"""
|
2014-02-05 19:26:47 +00:00
|
|
|
self.content = content
|
|
|
|
self.timestamp_start = timestamp_start
|
|
|
|
self.timestamp_end = timestamp_end
|
|
|
|
|
2014-02-05 13:33:17 +00:00
|
|
|
self.flow = None # will usually be set by the flow backref mixin
|
2014-01-31 03:45:39 +00:00
|
|
|
"""@type: HTTPFlow"""
|
2014-01-30 04:00:13 +00:00
|
|
|
|
2014-02-05 19:26:47 +00:00
|
|
|
_stateobject_attributes = dict(
|
|
|
|
httpversion=tuple,
|
|
|
|
headers=ODictCaseless,
|
|
|
|
content=str,
|
|
|
|
timestamp_start=float,
|
|
|
|
timestamp_end=float
|
|
|
|
)
|
|
|
|
|
2014-01-30 04:00:13 +00:00
|
|
|
def get_decoded_content(self):
|
|
|
|
"""
|
|
|
|
Returns the decoded content based on the current Content-Encoding header.
|
|
|
|
Doesn't change the message iteself or its headers.
|
|
|
|
"""
|
|
|
|
ce = self.headers.get_first("content-encoding")
|
|
|
|
if not self.content or ce not in encoding.ENCODINGS:
|
|
|
|
return self.content
|
|
|
|
return encoding.decode(ce, self.content)
|
|
|
|
|
|
|
|
def decode(self):
|
|
|
|
"""
|
|
|
|
Decodes content based on the current Content-Encoding header, then
|
|
|
|
removes the header. If there is no Content-Encoding header, no
|
|
|
|
action is taken.
|
|
|
|
|
|
|
|
Returns True if decoding succeeded, False otherwise.
|
|
|
|
"""
|
|
|
|
ce = self.headers.get_first("content-encoding")
|
|
|
|
if not self.content or ce not in encoding.ENCODINGS:
|
|
|
|
return False
|
|
|
|
data = encoding.decode(ce, self.content)
|
|
|
|
if data is None:
|
|
|
|
return False
|
|
|
|
self.content = data
|
|
|
|
del self.headers["content-encoding"]
|
|
|
|
return True
|
|
|
|
|
|
|
|
def encode(self, e):
|
|
|
|
"""
|
|
|
|
Encodes content with the encoding e, where e is "gzip", "deflate"
|
|
|
|
or "identity".
|
|
|
|
"""
|
|
|
|
# FIXME: Error if there's an existing encoding header?
|
|
|
|
self.content = encoding.encode(e, self.content)
|
|
|
|
self.headers["content-encoding"] = [e]
|
|
|
|
|
|
|
|
def size(self, **kwargs):
|
|
|
|
"""
|
|
|
|
Size in bytes of a fully rendered message, including headers and
|
|
|
|
HTTP lead-in.
|
|
|
|
"""
|
|
|
|
hl = len(self._assemble_head(**kwargs))
|
|
|
|
if self.content:
|
|
|
|
return hl + len(self.content)
|
|
|
|
else:
|
|
|
|
return hl
|
|
|
|
|
|
|
|
def copy(self):
|
|
|
|
c = copy.copy(self)
|
|
|
|
c.headers = self.headers.copy()
|
|
|
|
return c
|
|
|
|
|
|
|
|
def replace(self, pattern, repl, *args, **kwargs):
|
|
|
|
"""
|
|
|
|
Replaces a regular expression pattern with repl in both the headers
|
|
|
|
and the body of the message. Encoded content will be decoded
|
|
|
|
before replacement, and re-encoded afterwards.
|
|
|
|
|
|
|
|
Returns the number of replacements made.
|
|
|
|
"""
|
|
|
|
with decoded(self):
|
|
|
|
self.content, c = utils.safe_subn(pattern, repl, self.content, *args, **kwargs)
|
|
|
|
c += self.headers.replace(pattern, repl, *args, **kwargs)
|
|
|
|
return c
|
|
|
|
|
|
|
|
@classmethod
|
2014-07-14 15:26:22 +00:00
|
|
|
def from_stream(cls, rfile, include_body=True, body_size_limit=None):
|
2014-01-30 04:00:13 +00:00
|
|
|
"""
|
|
|
|
Parse an HTTP message from a file stream
|
|
|
|
"""
|
2014-02-07 02:56:57 +00:00
|
|
|
raise NotImplementedError # pragma: nocover
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
def _assemble_first_line(self):
|
|
|
|
"""
|
|
|
|
Returns the assembled request/response line
|
|
|
|
"""
|
2014-02-07 02:56:57 +00:00
|
|
|
raise NotImplementedError # pragma: nocover
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
def _assemble_headers(self):
|
|
|
|
"""
|
|
|
|
Returns the assembled headers
|
|
|
|
"""
|
2014-02-07 02:56:57 +00:00
|
|
|
raise NotImplementedError # pragma: nocover
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
def _assemble_head(self):
|
|
|
|
"""
|
|
|
|
Returns the assembled request/response line plus headers
|
|
|
|
"""
|
2014-02-07 02:56:57 +00:00
|
|
|
raise NotImplementedError # pragma: nocover
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
def _assemble(self):
|
|
|
|
"""
|
|
|
|
Returns the assembled request/response
|
|
|
|
"""
|
2014-02-07 02:56:57 +00:00
|
|
|
raise NotImplementedError # pragma: nocover
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
|
|
|
|
class HTTPRequest(HTTPMessage):
|
|
|
|
"""
|
|
|
|
An HTTP request.
|
|
|
|
|
|
|
|
Exposes the following attributes:
|
|
|
|
|
|
|
|
flow: Flow object the request belongs to
|
|
|
|
|
|
|
|
headers: ODictCaseless object
|
|
|
|
|
|
|
|
content: Content of the request, None, or CONTENT_MISSING if there
|
|
|
|
is content associated, but not present. CONTENT_MISSING evaluates
|
|
|
|
to False to make checking for the presence of content natural.
|
|
|
|
|
|
|
|
form_in: The request form which mitmproxy has received. The following values are possible:
|
2014-03-08 14:47:09 +00:00
|
|
|
- relative (GET /index.html, OPTIONS *) (covers origin form and asterisk form)
|
2014-01-30 04:00:13 +00:00
|
|
|
- absolute (GET http://example.com:80/index.html)
|
|
|
|
- authority-form (CONNECT example.com:443)
|
|
|
|
Details: http://tools.ietf.org/html/draft-ietf-httpbis-p1-messaging-25#section-5.3
|
|
|
|
|
|
|
|
form_out: The request form which mitmproxy has send out to the destination
|
|
|
|
|
|
|
|
method: HTTP method
|
|
|
|
|
|
|
|
scheme: URL scheme (http/https) (absolute-form only)
|
|
|
|
|
|
|
|
host: Host portion of the URL (absolute-form and authority-form only)
|
|
|
|
|
|
|
|
port: Destination port (absolute-form and authority-form only)
|
|
|
|
|
|
|
|
path: Path portion of the URL (not present in authority-form)
|
|
|
|
|
|
|
|
httpversion: HTTP version tuple
|
|
|
|
|
|
|
|
timestamp_start: Timestamp indicating when request transmission started
|
|
|
|
|
|
|
|
timestamp_end: Timestamp indicating when request transmission ended
|
|
|
|
"""
|
2014-05-15 16:16:42 +00:00
|
|
|
|
|
|
|
def __init__(self, form_in, method, scheme, host, port, path, httpversion, headers,
|
|
|
|
content, timestamp_start=None, timestamp_end=None, form_out=None):
|
2014-01-30 04:00:13 +00:00
|
|
|
assert isinstance(headers, ODictCaseless) or not headers
|
2014-05-15 16:16:42 +00:00
|
|
|
HTTPMessage.__init__(self, httpversion, headers, content, timestamp_start,
|
|
|
|
timestamp_end)
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
self.form_in = form_in
|
|
|
|
self.method = method
|
|
|
|
self.scheme = scheme
|
|
|
|
self.host = host
|
|
|
|
self.port = port
|
|
|
|
self.path = path
|
|
|
|
self.httpversion = httpversion
|
|
|
|
self.form_out = form_out or form_in
|
|
|
|
|
|
|
|
# Have this request's cookies been modified by sticky cookies or auth?
|
|
|
|
self.stickycookie = False
|
|
|
|
self.stickyauth = False
|
|
|
|
# Is this request replayed?
|
|
|
|
self.is_replay = False
|
|
|
|
|
2014-02-05 19:26:47 +00:00
|
|
|
_stateobject_attributes = HTTPMessage._stateobject_attributes.copy()
|
|
|
|
_stateobject_attributes.update(
|
2014-01-30 04:00:13 +00:00
|
|
|
form_in=str,
|
|
|
|
method=str,
|
|
|
|
scheme=str,
|
|
|
|
host=str,
|
|
|
|
port=int,
|
|
|
|
path=str,
|
|
|
|
form_out=str
|
|
|
|
)
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def _from_state(cls, state):
|
|
|
|
f = cls(None, None, None, None, None, None, None, None, None, None, None)
|
|
|
|
f._load_state(state)
|
|
|
|
return f
|
|
|
|
|
|
|
|
@classmethod
|
2014-07-14 15:26:22 +00:00
|
|
|
def from_stream(cls, rfile, include_body=True, body_size_limit=None):
|
2014-01-30 04:00:13 +00:00
|
|
|
"""
|
|
|
|
Parse an HTTP request from a file stream
|
|
|
|
"""
|
|
|
|
httpversion, host, port, scheme, method, path, headers, content, timestamp_start, timestamp_end \
|
|
|
|
= None, None, None, None, None, None, None, None, None, None
|
|
|
|
|
2014-02-07 02:56:57 +00:00
|
|
|
if hasattr(rfile, "reset_timestamps"):
|
|
|
|
rfile.reset_timestamps()
|
|
|
|
|
2014-01-30 04:00:13 +00:00
|
|
|
request_line = get_line(rfile)
|
2014-02-07 02:56:57 +00:00
|
|
|
|
|
|
|
if hasattr(rfile, "first_byte_timestamp"):
|
|
|
|
timestamp_start = rfile.first_byte_timestamp
|
|
|
|
else:
|
|
|
|
timestamp_start = utils.timestamp()
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
request_line_parts = http.parse_init(request_line)
|
|
|
|
if not request_line_parts:
|
|
|
|
raise http.HttpError(400, "Bad HTTP request line: %s" % repr(request_line))
|
|
|
|
method, path, httpversion = request_line_parts
|
|
|
|
|
2014-03-08 14:47:09 +00:00
|
|
|
if path == '*' or path.startswith("/"):
|
|
|
|
form_in = "relative"
|
2014-01-30 04:00:13 +00:00
|
|
|
if not netlib.utils.isascii(path):
|
|
|
|
raise http.HttpError(400, "Bad HTTP request line: %s" % repr(request_line))
|
|
|
|
elif method.upper() == 'CONNECT':
|
|
|
|
form_in = "authority"
|
|
|
|
r = http.parse_init_connect(request_line)
|
|
|
|
if not r:
|
|
|
|
raise http.HttpError(400, "Bad HTTP request line: %s" % repr(request_line))
|
|
|
|
host, port, _ = r
|
|
|
|
path = None
|
|
|
|
else:
|
|
|
|
form_in = "absolute"
|
|
|
|
r = http.parse_init_proxy(request_line)
|
|
|
|
if not r:
|
|
|
|
raise http.HttpError(400, "Bad HTTP request line: %s" % repr(request_line))
|
|
|
|
_, scheme, host, port, path, _ = r
|
|
|
|
|
|
|
|
headers = http.read_headers(rfile)
|
|
|
|
if headers is None:
|
|
|
|
raise http.HttpError(400, "Invalid headers")
|
|
|
|
|
2014-07-14 15:26:22 +00:00
|
|
|
if include_body:
|
2014-01-30 04:00:13 +00:00
|
|
|
content = http.read_http_body(rfile, headers, body_size_limit, True)
|
|
|
|
timestamp_end = utils.timestamp()
|
|
|
|
|
2014-05-15 16:16:42 +00:00
|
|
|
return HTTPRequest(form_in, method, scheme, host, port, path, httpversion, headers,
|
|
|
|
content, timestamp_start, timestamp_end)
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
def _assemble_first_line(self, form=None):
|
|
|
|
form = form or self.form_out
|
|
|
|
|
2014-03-08 14:47:09 +00:00
|
|
|
if form == "relative":
|
|
|
|
path = self.path if self.method != "OPTIONS" else "*"
|
2014-05-15 16:16:42 +00:00
|
|
|
request_line = '%s %s HTTP/%s.%s' % \
|
|
|
|
(self.method, path, self.httpversion[0], self.httpversion[1])
|
2014-01-30 04:00:13 +00:00
|
|
|
elif form == "authority":
|
|
|
|
request_line = '%s %s:%s HTTP/%s.%s' % (self.method, self.host, self.port,
|
|
|
|
self.httpversion[0], self.httpversion[1])
|
|
|
|
elif form == "absolute":
|
|
|
|
request_line = '%s %s://%s:%s%s HTTP/%s.%s' % \
|
|
|
|
(self.method, self.scheme, self.host, self.port, self.path,
|
|
|
|
self.httpversion[0], self.httpversion[1])
|
|
|
|
else:
|
|
|
|
raise http.HttpError(400, "Invalid request form")
|
|
|
|
return request_line
|
|
|
|
|
|
|
|
def _assemble_headers(self):
|
|
|
|
headers = self.headers.copy()
|
|
|
|
utils.del_all(
|
|
|
|
headers,
|
|
|
|
[
|
|
|
|
'Proxy-Connection',
|
|
|
|
'Keep-Alive',
|
|
|
|
'Connection',
|
|
|
|
'Transfer-Encoding'
|
|
|
|
]
|
|
|
|
)
|
|
|
|
if not 'host' in headers:
|
2014-01-31 03:45:39 +00:00
|
|
|
headers["Host"] = [utils.hostport(self.scheme,
|
|
|
|
self.host or self.flow.server_conn.address.host,
|
|
|
|
self.port or self.flow.server_conn.address.port)]
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
if self.content:
|
|
|
|
headers["Content-Length"] = [str(len(self.content))]
|
|
|
|
elif 'Transfer-Encoding' in self.headers: # content-length for e.g. chuncked transfer-encoding with no content
|
|
|
|
headers["Content-Length"] = ["0"]
|
|
|
|
|
|
|
|
return str(headers)
|
|
|
|
|
|
|
|
def _assemble_head(self, form=None):
|
|
|
|
return "%s\r\n%s\r\n" % (self._assemble_first_line(form), self._assemble_headers())
|
|
|
|
|
|
|
|
def _assemble(self, form=None):
|
|
|
|
"""
|
|
|
|
Assembles the request for transmission to the server. We make some
|
|
|
|
modifications to make sure interception works properly.
|
|
|
|
|
|
|
|
Raises an Exception if the request cannot be assembled.
|
|
|
|
"""
|
|
|
|
if self.content == CONTENT_MISSING:
|
2014-03-10 20:57:50 +00:00
|
|
|
raise proxy.ProxyError(502, "Cannot assemble flow with CONTENT_MISSING")
|
2014-01-30 04:00:13 +00:00
|
|
|
head = self._assemble_head(form)
|
|
|
|
if self.content:
|
|
|
|
return head + self.content
|
|
|
|
else:
|
|
|
|
return head
|
|
|
|
|
|
|
|
def __hash__(self):
|
|
|
|
return id(self)
|
|
|
|
|
|
|
|
def anticache(self):
|
|
|
|
"""
|
|
|
|
Modifies this request to remove headers that might produce a cached
|
|
|
|
response. That is, we remove ETags and If-Modified-Since headers.
|
|
|
|
"""
|
|
|
|
delheaders = [
|
|
|
|
"if-modified-since",
|
|
|
|
"if-none-match",
|
|
|
|
]
|
|
|
|
for i in delheaders:
|
|
|
|
del self.headers[i]
|
|
|
|
|
|
|
|
def anticomp(self):
|
|
|
|
"""
|
|
|
|
Modifies this request to remove headers that will compress the
|
|
|
|
resource's data.
|
|
|
|
"""
|
|
|
|
self.headers["accept-encoding"] = ["identity"]
|
|
|
|
|
|
|
|
def constrain_encoding(self):
|
|
|
|
"""
|
|
|
|
Limits the permissible Accept-Encoding values, based on what we can
|
|
|
|
decode appropriately.
|
|
|
|
"""
|
|
|
|
if self.headers["accept-encoding"]:
|
|
|
|
self.headers["accept-encoding"] = [', '.join(
|
|
|
|
e for e in encoding.ENCODINGS if e in self.headers["accept-encoding"][0]
|
|
|
|
)]
|
|
|
|
|
|
|
|
def get_form_urlencoded(self):
|
|
|
|
"""
|
|
|
|
Retrieves the URL-encoded form data, returning an ODict object.
|
|
|
|
Returns an empty ODict if there is no data or the content-type
|
|
|
|
indicates non-form data.
|
|
|
|
"""
|
|
|
|
if self.content and self.headers.in_any("content-type", HDR_FORM_URLENCODED, True):
|
|
|
|
return ODict(utils.urldecode(self.content))
|
|
|
|
return ODict([])
|
|
|
|
|
|
|
|
def set_form_urlencoded(self, odict):
|
|
|
|
"""
|
|
|
|
Sets the body to the URL-encoded form data, and adds the
|
|
|
|
appropriate content-type header. Note that this will destory the
|
|
|
|
existing body if there is one.
|
|
|
|
"""
|
|
|
|
# FIXME: If there's an existing content-type header indicating a
|
|
|
|
# url-encoded form, leave it alone.
|
|
|
|
self.headers["Content-Type"] = [HDR_FORM_URLENCODED]
|
|
|
|
self.content = utils.urlencode(odict.lst)
|
|
|
|
|
|
|
|
def get_path_components(self):
|
|
|
|
"""
|
|
|
|
Returns the path components of the URL as a list of strings.
|
|
|
|
|
|
|
|
Components are unquoted.
|
|
|
|
"""
|
|
|
|
_, _, path, _, _, _ = urlparse.urlparse(self.get_url())
|
|
|
|
return [urllib.unquote(i) for i in path.split("/") if i]
|
|
|
|
|
|
|
|
def set_path_components(self, lst):
|
|
|
|
"""
|
|
|
|
Takes a list of strings, and sets the path component of the URL.
|
|
|
|
|
|
|
|
Components are quoted.
|
|
|
|
"""
|
|
|
|
lst = [urllib.quote(i, safe="") for i in lst]
|
|
|
|
path = "/" + "/".join(lst)
|
|
|
|
scheme, netloc, _, params, query, fragment = urlparse.urlparse(self.get_url())
|
|
|
|
self.set_url(urlparse.urlunparse([scheme, netloc, path, params, query, fragment]))
|
|
|
|
|
|
|
|
def get_query(self):
|
|
|
|
"""
|
|
|
|
Gets the request query string. Returns an ODict object.
|
|
|
|
"""
|
|
|
|
_, _, _, _, query, _ = urlparse.urlparse(self.get_url())
|
|
|
|
if query:
|
|
|
|
return ODict(utils.urldecode(query))
|
|
|
|
return ODict([])
|
|
|
|
|
|
|
|
def set_query(self, odict):
|
|
|
|
"""
|
|
|
|
Takes an ODict object, and sets the request query string.
|
|
|
|
"""
|
|
|
|
scheme, netloc, path, params, _, fragment = urlparse.urlparse(self.get_url())
|
|
|
|
query = utils.urlencode(odict.lst)
|
|
|
|
self.set_url(urlparse.urlunparse([scheme, netloc, path, params, query, fragment]))
|
|
|
|
|
2014-02-05 19:26:47 +00:00
|
|
|
def get_host(self, hostheader=False):
|
2014-01-30 04:00:13 +00:00
|
|
|
"""
|
2014-02-05 19:26:47 +00:00
|
|
|
Heuristic to get the host of the request.
|
|
|
|
The host is not necessarily equal to the TCP destination of the request,
|
|
|
|
for example on a transparently proxified absolute-form request to an upstream HTTP proxy.
|
|
|
|
If hostheader is set to True, the Host: header will be used as additional (and preferred) data source.
|
2014-01-30 04:00:13 +00:00
|
|
|
"""
|
2014-01-31 03:45:39 +00:00
|
|
|
host = None
|
2014-01-30 04:00:13 +00:00
|
|
|
if hostheader:
|
2014-01-31 03:45:39 +00:00
|
|
|
host = self.headers.get_first("host")
|
|
|
|
if not host:
|
|
|
|
if self.host:
|
|
|
|
host = self.host
|
|
|
|
else:
|
|
|
|
host = self.flow.server_conn.address.host
|
2014-01-30 04:00:13 +00:00
|
|
|
host = host.encode("idna")
|
2014-02-05 19:26:47 +00:00
|
|
|
return host
|
|
|
|
|
|
|
|
def get_scheme(self):
|
|
|
|
"""
|
|
|
|
Returns the request port, either from the request itself or from the flow's server connection
|
|
|
|
"""
|
|
|
|
if self.scheme:
|
|
|
|
return self.scheme
|
|
|
|
return "https" if self.flow.server_conn.ssl_established else "http"
|
|
|
|
|
|
|
|
def get_port(self):
|
|
|
|
"""
|
|
|
|
Returns the request port, either from the request itself or from the flow's server connection
|
|
|
|
"""
|
|
|
|
if self.port:
|
|
|
|
return self.port
|
|
|
|
return self.flow.server_conn.address.port
|
|
|
|
|
|
|
|
def get_url(self, hostheader=False):
|
|
|
|
"""
|
|
|
|
Returns a URL string, constructed from the Request's URL components.
|
|
|
|
|
|
|
|
If hostheader is True, we use the value specified in the request
|
|
|
|
Host header to construct the URL.
|
|
|
|
"""
|
2014-03-18 18:38:02 +00:00
|
|
|
if self.form_out == "authority": # upstream proxy mode
|
|
|
|
return "%s:%s" % (self.get_host(hostheader), self.get_port())
|
2014-02-05 19:26:47 +00:00
|
|
|
return utils.unparse_url(self.get_scheme(),
|
|
|
|
self.get_host(hostheader),
|
|
|
|
self.get_port(),
|
|
|
|
self.path).encode('ascii')
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
def set_url(self, url):
|
|
|
|
"""
|
|
|
|
Parses a URL specification, and updates the Request's information
|
|
|
|
accordingly.
|
|
|
|
|
|
|
|
Returns False if the URL was invalid, True if the request succeeded.
|
|
|
|
"""
|
|
|
|
parts = http.parse_url(url)
|
|
|
|
if not parts:
|
|
|
|
return False
|
2014-02-05 19:26:47 +00:00
|
|
|
scheme, host, port, path = parts
|
|
|
|
is_ssl = (True if scheme == "https" else False)
|
|
|
|
|
|
|
|
self.path = path
|
|
|
|
|
|
|
|
if host != self.get_host() or port != self.get_port():
|
|
|
|
if self.flow.change_server:
|
|
|
|
self.flow.change_server((host, port), ssl=is_ssl)
|
|
|
|
else:
|
|
|
|
# There's not live server connection, we're just changing the attributes here.
|
2014-05-15 16:16:42 +00:00
|
|
|
self.flow.server_conn = ServerConnection((host, port),
|
|
|
|
proxy.AddressPriority.MANUALLY_CHANGED)
|
2014-02-05 19:26:47 +00:00
|
|
|
self.flow.server_conn.ssl_established = is_ssl
|
|
|
|
|
|
|
|
# If this is an absolute request, replace the attributes on the request object as well.
|
|
|
|
if self.host:
|
|
|
|
self.host = host
|
|
|
|
if self.port:
|
|
|
|
self.port = port
|
|
|
|
if self.scheme:
|
|
|
|
self.scheme = scheme
|
|
|
|
|
2014-01-30 04:00:13 +00:00
|
|
|
return True
|
|
|
|
|
|
|
|
def get_cookies(self):
|
|
|
|
cookie_headers = self.headers.get("cookie")
|
|
|
|
if not cookie_headers:
|
|
|
|
return None
|
|
|
|
|
|
|
|
cookies = []
|
|
|
|
for header in cookie_headers:
|
|
|
|
pairs = [pair.partition("=") for pair in header.split(';')]
|
|
|
|
cookies.extend((pair[0], (pair[2], {})) for pair in pairs)
|
|
|
|
return dict(cookies)
|
|
|
|
|
|
|
|
def replace(self, pattern, repl, *args, **kwargs):
|
|
|
|
"""
|
|
|
|
Replaces a regular expression pattern with repl in the headers, the request path
|
|
|
|
and the body of the request. Encoded content will be decoded before
|
|
|
|
replacement, and re-encoded afterwards.
|
|
|
|
|
|
|
|
Returns the number of replacements made.
|
|
|
|
"""
|
|
|
|
c = HTTPMessage.replace(self, pattern, repl, *args, **kwargs)
|
|
|
|
self.path, pc = utils.safe_subn(pattern, repl, self.path, *args, **kwargs)
|
|
|
|
c += pc
|
|
|
|
return c
|
|
|
|
|
|
|
|
|
|
|
|
class HTTPResponse(HTTPMessage):
|
|
|
|
"""
|
|
|
|
An HTTP response.
|
|
|
|
|
|
|
|
Exposes the following attributes:
|
|
|
|
|
|
|
|
flow: Flow object the request belongs to
|
|
|
|
|
|
|
|
code: HTTP response code
|
|
|
|
|
|
|
|
msg: HTTP response message
|
|
|
|
|
|
|
|
headers: ODict object
|
|
|
|
|
|
|
|
content: Content of the request, None, or CONTENT_MISSING if there
|
|
|
|
is content associated, but not present. CONTENT_MISSING evaluates
|
|
|
|
to False to make checking for the presence of content natural.
|
|
|
|
|
|
|
|
httpversion: HTTP version tuple
|
|
|
|
|
|
|
|
timestamp_start: Timestamp indicating when request transmission started
|
|
|
|
|
|
|
|
timestamp_end: Timestamp indicating when request transmission ended
|
|
|
|
"""
|
2014-05-15 16:16:42 +00:00
|
|
|
|
|
|
|
def __init__(self, httpversion, code, msg, headers, content, timestamp_start=None,
|
|
|
|
timestamp_end=None):
|
2014-01-30 19:11:01 +00:00
|
|
|
assert isinstance(headers, ODictCaseless) or headers is None
|
2014-05-15 16:16:42 +00:00
|
|
|
HTTPMessage.__init__(self, httpversion, headers, content, timestamp_start,
|
|
|
|
timestamp_end)
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
self.code = code
|
|
|
|
self.msg = msg
|
|
|
|
|
|
|
|
# Is this request replayed?
|
|
|
|
self.is_replay = False
|
|
|
|
|
2014-02-05 19:26:47 +00:00
|
|
|
_stateobject_attributes = HTTPMessage._stateobject_attributes.copy()
|
|
|
|
_stateobject_attributes.update(
|
2014-01-30 04:00:13 +00:00
|
|
|
code=int,
|
2014-02-05 19:26:47 +00:00
|
|
|
msg=str
|
2014-01-30 04:00:13 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def _from_state(cls, state):
|
2014-02-05 19:26:47 +00:00
|
|
|
f = cls(None, None, None, None, None)
|
2014-01-30 04:00:13 +00:00
|
|
|
f._load_state(state)
|
|
|
|
return f
|
|
|
|
|
|
|
|
@classmethod
|
2014-07-14 15:26:22 +00:00
|
|
|
def from_stream(cls, rfile, request_method, include_body=True, body_size_limit=None):
|
2014-01-30 04:00:13 +00:00
|
|
|
"""
|
|
|
|
Parse an HTTP response from a file stream
|
|
|
|
"""
|
2014-02-07 02:56:57 +00:00
|
|
|
|
|
|
|
if hasattr(rfile, "reset_timestamps"):
|
|
|
|
rfile.reset_timestamps()
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
httpversion, code, msg, headers, content = http.read_response(
|
|
|
|
rfile,
|
|
|
|
request_method,
|
2014-07-13 05:40:51 +00:00
|
|
|
body_size_limit,
|
2014-07-14 15:26:22 +00:00
|
|
|
include_body=include_body)
|
2014-02-07 02:56:57 +00:00
|
|
|
|
|
|
|
if hasattr(rfile, "first_byte_timestamp"):
|
|
|
|
timestamp_start = rfile.first_byte_timestamp
|
|
|
|
else:
|
|
|
|
timestamp_start = utils.timestamp()
|
|
|
|
|
2014-01-30 04:00:13 +00:00
|
|
|
timestamp_end = utils.timestamp()
|
2014-05-15 16:16:42 +00:00
|
|
|
return HTTPResponse(httpversion, code, msg, headers, content, timestamp_start,
|
|
|
|
timestamp_end)
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
def _assemble_first_line(self):
|
2014-05-15 16:16:42 +00:00
|
|
|
return 'HTTP/%s.%s %s %s' % \
|
|
|
|
(self.httpversion[0], self.httpversion[1], self.code, self.msg)
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
def _assemble_headers(self):
|
|
|
|
headers = self.headers.copy()
|
|
|
|
utils.del_all(
|
|
|
|
headers,
|
|
|
|
[
|
|
|
|
'Proxy-Connection',
|
|
|
|
'Transfer-Encoding'
|
|
|
|
]
|
|
|
|
)
|
|
|
|
if self.content:
|
|
|
|
headers["Content-Length"] = [str(len(self.content))]
|
2014-01-31 03:45:39 +00:00
|
|
|
elif 'Transfer-Encoding' in self.headers: # add content-length for chuncked transfer-encoding with no content
|
2014-01-30 04:00:13 +00:00
|
|
|
headers["Content-Length"] = ["0"]
|
|
|
|
|
|
|
|
return str(headers)
|
|
|
|
|
|
|
|
def _assemble_head(self):
|
|
|
|
return '%s\r\n%s\r\n' % (self._assemble_first_line(), self._assemble_headers())
|
|
|
|
|
|
|
|
def _assemble(self):
|
|
|
|
"""
|
|
|
|
Assembles the response for transmission to the client. We make some
|
|
|
|
modifications to make sure interception works properly.
|
|
|
|
|
|
|
|
Raises an Exception if the request cannot be assembled.
|
|
|
|
"""
|
|
|
|
if self.content == CONTENT_MISSING:
|
2014-03-10 20:57:50 +00:00
|
|
|
raise proxy.ProxyError(502, "Cannot assemble flow with CONTENT_MISSING")
|
2014-01-30 04:00:13 +00:00
|
|
|
head = self._assemble_head()
|
|
|
|
if self.content:
|
|
|
|
return head + self.content
|
|
|
|
else:
|
|
|
|
return head
|
|
|
|
|
|
|
|
def _refresh_cookie(self, c, delta):
|
|
|
|
"""
|
|
|
|
Takes a cookie string c and a time delta in seconds, and returns
|
|
|
|
a refreshed cookie string.
|
|
|
|
"""
|
|
|
|
c = Cookie.SimpleCookie(str(c))
|
|
|
|
for i in c.values():
|
|
|
|
if "expires" in i:
|
|
|
|
d = parsedate_tz(i["expires"])
|
|
|
|
if d:
|
|
|
|
d = mktime_tz(d) + delta
|
|
|
|
i["expires"] = formatdate(d)
|
|
|
|
else:
|
|
|
|
# This can happen when the expires tag is invalid.
|
|
|
|
# reddit.com sends a an expires tag like this: "Thu, 31 Dec
|
|
|
|
# 2037 23:59:59 GMT", which is valid RFC 1123, but not
|
2014-02-07 02:56:57 +00:00
|
|
|
# strictly correct according to the cookie spec. Browsers
|
2014-01-30 04:00:13 +00:00
|
|
|
# appear to parse this tolerantly - maybe we should too.
|
|
|
|
# For now, we just ignore this.
|
|
|
|
del i["expires"]
|
|
|
|
return c.output(header="").strip()
|
|
|
|
|
|
|
|
def refresh(self, now=None):
|
|
|
|
"""
|
|
|
|
This fairly complex and heuristic function refreshes a server
|
|
|
|
response for replay.
|
|
|
|
|
|
|
|
- It adjusts date, expires and last-modified headers.
|
|
|
|
- It adjusts cookie expiration.
|
|
|
|
"""
|
|
|
|
if not now:
|
|
|
|
now = time.time()
|
|
|
|
delta = now - self.timestamp_start
|
|
|
|
refresh_headers = [
|
|
|
|
"date",
|
|
|
|
"expires",
|
|
|
|
"last-modified",
|
|
|
|
]
|
|
|
|
for i in refresh_headers:
|
|
|
|
if i in self.headers:
|
|
|
|
d = parsedate_tz(self.headers[i][0])
|
|
|
|
if d:
|
|
|
|
new = mktime_tz(d) + delta
|
|
|
|
self.headers[i] = [formatdate(new)]
|
|
|
|
c = []
|
|
|
|
for i in self.headers["set-cookie"]:
|
|
|
|
c.append(self._refresh_cookie(i, delta))
|
|
|
|
if c:
|
|
|
|
self.headers["set-cookie"] = c
|
|
|
|
|
|
|
|
def get_cookies(self):
|
|
|
|
cookie_headers = self.headers.get("set-cookie")
|
|
|
|
if not cookie_headers:
|
|
|
|
return None
|
|
|
|
|
|
|
|
cookies = []
|
|
|
|
for header in cookie_headers:
|
|
|
|
pairs = [pair.partition("=") for pair in header.split(';')]
|
2014-01-31 03:45:39 +00:00
|
|
|
cookie_name = pairs[0][0] # the key of the first key/value pairs
|
|
|
|
cookie_value = pairs[0][2] # the value of the first key/value pairs
|
2014-05-15 16:16:42 +00:00
|
|
|
cookie_parameters = {key.strip().lower(): value.strip() for key, sep, value in
|
|
|
|
pairs[1:]}
|
2014-01-30 04:00:13 +00:00
|
|
|
cookies.append((cookie_name, (cookie_value, cookie_parameters)))
|
|
|
|
return dict(cookies)
|
|
|
|
|
|
|
|
|
|
|
|
class HTTPFlow(Flow):
|
|
|
|
"""
|
|
|
|
A Flow is a collection of objects representing a single HTTP
|
|
|
|
transaction. The main attributes are:
|
|
|
|
|
|
|
|
request: HTTPRequest object
|
|
|
|
response: HTTPResponse object
|
|
|
|
error: Error object
|
|
|
|
|
|
|
|
Note that it's possible for a Flow to have both a response and an error
|
|
|
|
object. This might happen, for instance, when a response was received
|
|
|
|
from the server, but there was an error sending it back to the client.
|
|
|
|
|
|
|
|
The following additional attributes are exposed:
|
|
|
|
|
|
|
|
intercepting: Is this flow currently being intercepted?
|
|
|
|
"""
|
2014-05-15 16:16:42 +00:00
|
|
|
|
2014-02-05 19:26:47 +00:00
|
|
|
def __init__(self, client_conn, server_conn, change_server=None):
|
2014-01-31 03:45:39 +00:00
|
|
|
Flow.__init__(self, "http", client_conn, server_conn)
|
|
|
|
self.request = None
|
2014-02-05 19:26:47 +00:00
|
|
|
"""@type: HTTPRequest"""
|
2014-01-31 03:45:39 +00:00
|
|
|
self.response = None
|
2014-02-05 19:26:47 +00:00
|
|
|
"""@type: HTTPResponse"""
|
2014-02-05 23:53:39 +00:00
|
|
|
self.change_server = change_server # Used by flow.request.set_url to change the server address
|
2014-01-30 04:21:53 +00:00
|
|
|
|
2014-01-31 03:45:39 +00:00
|
|
|
self.intercepting = False # FIXME: Should that rather be an attribute of Flow?
|
2014-01-30 04:21:53 +00:00
|
|
|
|
2014-01-30 04:00:13 +00:00
|
|
|
_backrefattr = Flow._backrefattr + ("request", "response")
|
2014-01-30 04:21:53 +00:00
|
|
|
|
2014-01-30 04:00:13 +00:00
|
|
|
_stateobject_attributes = Flow._stateobject_attributes.copy()
|
|
|
|
_stateobject_attributes.update(
|
|
|
|
request=HTTPRequest,
|
|
|
|
response=HTTPResponse
|
|
|
|
)
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def _from_state(cls, state):
|
2014-01-31 03:45:39 +00:00
|
|
|
f = cls(None, None)
|
2014-01-30 04:00:13 +00:00
|
|
|
f._load_state(state)
|
|
|
|
return f
|
|
|
|
|
|
|
|
def copy(self):
|
|
|
|
f = super(HTTPFlow, self).copy()
|
|
|
|
if self.request:
|
|
|
|
f.request = self.request.copy()
|
|
|
|
if self.response:
|
2014-02-04 04:02:17 +00:00
|
|
|
f.response = self.response.copy()
|
2014-01-30 04:00:13 +00:00
|
|
|
return f
|
|
|
|
|
2014-01-30 04:21:53 +00:00
|
|
|
def match(self, f):
|
|
|
|
"""
|
|
|
|
Match this flow against a compiled filter expression. Returns True
|
|
|
|
if matched, False if not.
|
|
|
|
|
|
|
|
If f is a string, it will be compiled as a filter expression. If
|
|
|
|
the expression is invalid, ValueError is raised.
|
|
|
|
"""
|
|
|
|
if isinstance(f, basestring):
|
|
|
|
f = filt.parse(f)
|
|
|
|
if not f:
|
|
|
|
raise ValueError("Invalid filter expression.")
|
|
|
|
if f:
|
|
|
|
return f(self)
|
|
|
|
return True
|
|
|
|
|
|
|
|
def kill(self, master):
|
|
|
|
"""
|
|
|
|
Kill this request.
|
|
|
|
"""
|
|
|
|
self.error = Error("Connection killed")
|
|
|
|
self.error.reply = controller.DummyReply()
|
|
|
|
if self.request and not self.request.reply.acked:
|
|
|
|
self.request.reply(KILL)
|
|
|
|
elif self.response and not self.response.reply.acked:
|
|
|
|
self.response.reply(KILL)
|
2014-01-31 03:45:39 +00:00
|
|
|
master.handle_error(self.error)
|
2014-01-30 04:21:53 +00:00
|
|
|
self.intercepting = False
|
|
|
|
|
|
|
|
def intercept(self):
|
|
|
|
"""
|
|
|
|
Intercept this Flow. Processing will stop until accept_intercept is
|
|
|
|
called.
|
|
|
|
"""
|
|
|
|
self.intercepting = True
|
|
|
|
|
|
|
|
def accept_intercept(self):
|
|
|
|
"""
|
|
|
|
Continue with the flow - called after an intercept().
|
|
|
|
"""
|
|
|
|
if self.request:
|
|
|
|
if not self.request.reply.acked:
|
|
|
|
self.request.reply()
|
|
|
|
elif self.response and not self.response.reply.acked:
|
|
|
|
self.response.reply()
|
|
|
|
self.intercepting = False
|
|
|
|
|
|
|
|
def replace(self, pattern, repl, *args, **kwargs):
|
|
|
|
"""
|
|
|
|
Replaces a regular expression pattern with repl in both request and response of the
|
|
|
|
flow. Encoded content will be decoded before replacement, and
|
|
|
|
re-encoded afterwards.
|
|
|
|
|
|
|
|
Returns the number of replacements made.
|
|
|
|
"""
|
|
|
|
c = self.request.replace(pattern, repl, *args, **kwargs)
|
|
|
|
if self.response:
|
|
|
|
c += self.response.replace(pattern, repl, *args, **kwargs)
|
|
|
|
return c
|
2014-01-30 04:00:13 +00:00
|
|
|
|
2014-01-31 02:01:51 +00:00
|
|
|
|
2014-01-30 04:00:13 +00:00
|
|
|
class HttpAuthenticationError(Exception):
|
|
|
|
def __init__(self, auth_headers=None):
|
2014-05-15 16:16:42 +00:00
|
|
|
super(HttpAuthenticationError, self).__init__("Proxy Authentication Required")
|
|
|
|
self.headers = auth_headers
|
|
|
|
self.code = 407
|
2014-01-30 04:00:13 +00:00
|
|
|
|
2014-05-15 16:16:42 +00:00
|
|
|
def __repr__(self):
|
|
|
|
return "Proxy Authentication Required"
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
|
2014-02-05 19:26:47 +00:00
|
|
|
class HTTPHandler(ProtocolHandler, TemporaryServerChangeMixin):
|
2014-03-10 01:32:27 +00:00
|
|
|
def __init__(self, c):
|
|
|
|
super(HTTPHandler, self).__init__(c)
|
|
|
|
self.expected_form_in = c.config.http_form_in
|
|
|
|
self.expected_form_out = c.config.http_form_out
|
|
|
|
self.skip_authentication = False
|
|
|
|
|
2014-01-30 04:00:13 +00:00
|
|
|
def handle_messages(self):
|
|
|
|
while self.handle_flow():
|
|
|
|
pass
|
|
|
|
self.c.close = True
|
|
|
|
|
|
|
|
def get_response_from_server(self, request):
|
2014-02-05 19:26:47 +00:00
|
|
|
self.c.establish_server_connection()
|
2014-01-30 04:00:13 +00:00
|
|
|
request_raw = request._assemble()
|
|
|
|
|
|
|
|
for i in range(2):
|
|
|
|
try:
|
2014-02-04 04:02:17 +00:00
|
|
|
self.c.server_conn.send(request_raw)
|
2014-01-30 04:00:13 +00:00
|
|
|
return HTTPResponse.from_stream(self.c.server_conn.rfile, request.method,
|
|
|
|
body_size_limit=self.c.config.body_size_limit)
|
|
|
|
except (tcp.NetLibDisconnect, http.HttpErrorConnClosed), v:
|
2014-03-13 00:04:45 +00:00
|
|
|
self.c.log("error in server communication: %s" % str(v), level="debug")
|
2014-01-30 04:00:13 +00:00
|
|
|
if i < 1:
|
|
|
|
# In any case, we try to reconnect at least once.
|
|
|
|
# This is necessary because it might be possible that we already initiated an upstream connection
|
|
|
|
# after clientconnect that has already been expired, e.g consider the following event log:
|
|
|
|
# > clientconnect (transparent mode destination known)
|
|
|
|
# > serverconnect
|
|
|
|
# > read n% of large request
|
|
|
|
# > server detects timeout, disconnects
|
|
|
|
# > read (100-n)% of large request
|
|
|
|
# > send large request upstream
|
|
|
|
self.c.server_reconnect()
|
|
|
|
else:
|
|
|
|
raise v
|
|
|
|
|
|
|
|
def handle_flow(self):
|
2014-02-05 19:26:47 +00:00
|
|
|
flow = HTTPFlow(self.c.client_conn, self.c.server_conn, self.change_server)
|
2014-01-30 04:00:13 +00:00
|
|
|
try:
|
2014-02-08 04:17:35 +00:00
|
|
|
req = HTTPRequest.from_stream(self.c.client_conn.rfile,
|
2014-03-10 01:32:27 +00:00
|
|
|
body_size_limit=self.c.config.body_size_limit)
|
2014-03-13 00:04:45 +00:00
|
|
|
self.c.log("request", "debug", [req._assemble_first_line(req.form_in)])
|
2014-06-25 21:03:15 +00:00
|
|
|
send_request_upstream = self.process_request(flow, req)
|
|
|
|
if not send_request_upstream:
|
2014-03-10 01:32:27 +00:00
|
|
|
return True
|
2014-02-08 04:17:35 +00:00
|
|
|
|
|
|
|
# Be careful NOT to assign the request to the flow before
|
|
|
|
# process_request completes. This is because the call can raise an
|
2014-03-10 01:32:27 +00:00
|
|
|
# exception. If the request object is already attached, this results
|
2014-02-08 04:17:35 +00:00
|
|
|
# in an Error object that has an attached request that has not been
|
|
|
|
# sent through to the Master.
|
|
|
|
flow.request = req
|
2014-02-06 00:19:51 +00:00
|
|
|
request_reply = self.c.channel.ask("request", flow.request)
|
2014-02-07 17:14:15 +00:00
|
|
|
flow.server_conn = self.c.server_conn
|
2014-02-06 00:19:51 +00:00
|
|
|
|
2014-01-30 04:00:13 +00:00
|
|
|
if request_reply is None or request_reply == KILL:
|
|
|
|
return False
|
|
|
|
|
2014-01-30 04:21:53 +00:00
|
|
|
if isinstance(request_reply, HTTPResponse):
|
2014-01-30 04:00:13 +00:00
|
|
|
flow.response = request_reply
|
|
|
|
else:
|
|
|
|
flow.response = self.get_response_from_server(flow.request)
|
|
|
|
|
2014-02-07 17:14:15 +00:00
|
|
|
flow.server_conn = self.c.server_conn # no further manipulation of self.c.server_conn beyond this point
|
|
|
|
# we can safely set it as the final attribute value here.
|
|
|
|
|
2014-03-13 00:04:45 +00:00
|
|
|
self.c.log("response", "debug", [flow.response._assemble_first_line()])
|
2014-02-06 00:19:51 +00:00
|
|
|
response_reply = self.c.channel.ask("response", flow.response)
|
2014-01-30 04:00:13 +00:00
|
|
|
if response_reply is None or response_reply == KILL:
|
|
|
|
return False
|
|
|
|
|
2014-02-07 17:14:15 +00:00
|
|
|
self.c.client_conn.send(flow.response._assemble())
|
2014-01-30 04:00:13 +00:00
|
|
|
flow.timestamp_end = utils.timestamp()
|
|
|
|
|
|
|
|
if (http.connection_close(flow.request.httpversion, flow.request.headers) or
|
|
|
|
http.connection_close(flow.response.httpversion, flow.response.headers)):
|
|
|
|
return False
|
|
|
|
|
|
|
|
if flow.request.form_in == "authority":
|
2014-02-07 17:14:15 +00:00
|
|
|
self.ssl_upgrade()
|
2014-01-30 19:11:01 +00:00
|
|
|
|
2014-05-15 16:16:42 +00:00
|
|
|
# If the user has changed the target server on this connection,
|
|
|
|
# restore the original target server
|
|
|
|
self.restore_server()
|
2014-01-30 04:00:13 +00:00
|
|
|
return True
|
2014-03-10 20:57:50 +00:00
|
|
|
except (HttpAuthenticationError, http.HttpError, proxy.ProxyError, tcp.NetLibError), e:
|
2014-01-30 04:00:13 +00:00
|
|
|
self.handle_error(e, flow)
|
|
|
|
return False
|
|
|
|
|
|
|
|
def handle_error(self, error, flow=None):
|
|
|
|
|
2014-05-15 16:16:42 +00:00
|
|
|
message = repr(error)
|
|
|
|
code = getattr(error, "code", 502)
|
|
|
|
headers = getattr(error, "headers", None)
|
|
|
|
|
|
|
|
if "tlsv1 alert unknown ca" in message:
|
|
|
|
message = message + " \nThe client does not trust the proxy's certificate."
|
|
|
|
|
|
|
|
self.c.log("error: %s" % message, level="info")
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
if flow:
|
2014-05-15 16:16:42 +00:00
|
|
|
flow.error = Error(message)
|
|
|
|
# FIXME: no flows without request or with both request and response at the moement.
|
2014-02-06 00:19:51 +00:00
|
|
|
if flow.request and not flow.response:
|
|
|
|
self.c.channel.ask("error", flow.error)
|
2014-01-30 04:00:13 +00:00
|
|
|
else:
|
2014-05-15 16:16:42 +00:00
|
|
|
pass # FIXME: Do we want to persist errors without flows?
|
2014-01-30 04:00:13 +00:00
|
|
|
|
2014-05-15 16:16:42 +00:00
|
|
|
try:
|
|
|
|
self.send_error(code, message, headers)
|
|
|
|
except:
|
|
|
|
pass
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
def send_error(self, code, message, headers):
|
|
|
|
response = http_status.RESPONSES.get(code, "Unknown")
|
|
|
|
html_content = '<html><head>\n<title>%d %s</title>\n</head>\n<body>\n%s\n</body>\n</html>' % \
|
|
|
|
(code, response, message)
|
|
|
|
self.c.client_conn.wfile.write("HTTP/1.1 %s %s\r\n" % (code, response))
|
|
|
|
self.c.client_conn.wfile.write("Server: %s\r\n" % self.c.server_version)
|
|
|
|
self.c.client_conn.wfile.write("Content-type: text/html\r\n")
|
|
|
|
self.c.client_conn.wfile.write("Content-Length: %d\r\n" % len(html_content))
|
|
|
|
if headers:
|
|
|
|
for key, value in headers.items():
|
|
|
|
self.c.client_conn.wfile.write("%s: %s\r\n" % (key, value))
|
|
|
|
self.c.client_conn.wfile.write("Connection: close\r\n")
|
|
|
|
self.c.client_conn.wfile.write("\r\n")
|
|
|
|
self.c.client_conn.wfile.write(html_content)
|
|
|
|
self.c.client_conn.wfile.flush()
|
|
|
|
|
2014-02-07 17:14:15 +00:00
|
|
|
def hook_reconnect(self, upstream_request):
|
2014-02-09 01:54:07 +00:00
|
|
|
"""
|
|
|
|
If the authority request has been forwarded upstream (because we have another proxy server there),
|
|
|
|
money-patch the ConnectionHandler.server_reconnect function to resend the CONNECT request on reconnect.
|
|
|
|
Hooking code isn't particulary beautiful, but it isolates this edge-case from
|
|
|
|
the protocol-agnostic ConnectionHandler
|
|
|
|
"""
|
2014-03-13 00:04:45 +00:00
|
|
|
self.c.log("Hook reconnect function", level="debug")
|
2014-02-07 17:14:15 +00:00
|
|
|
original_reconnect_func = self.c.server_reconnect
|
|
|
|
|
|
|
|
def reconnect_http_proxy():
|
2014-03-13 00:04:45 +00:00
|
|
|
self.c.log("Hooked reconnect function", "debug")
|
|
|
|
self.c.log("Hook: Run original reconnect", "debug")
|
2014-02-07 17:14:15 +00:00
|
|
|
original_reconnect_func(no_ssl=True)
|
2014-03-13 00:04:45 +00:00
|
|
|
self.c.log("Hook: Write CONNECT request to upstream proxy", "debug",
|
|
|
|
[upstream_request._assemble_first_line()])
|
2014-02-07 17:14:15 +00:00
|
|
|
self.c.server_conn.send(upstream_request._assemble())
|
2014-03-13 00:04:45 +00:00
|
|
|
self.c.log("Hook: Read answer to CONNECT request from proxy", "debug")
|
2014-02-07 17:14:15 +00:00
|
|
|
resp = HTTPResponse.from_stream(self.c.server_conn.rfile, upstream_request.method)
|
|
|
|
if resp.code != 200:
|
2014-03-10 20:57:50 +00:00
|
|
|
raise proxy.ProxyError(resp.code,
|
2014-05-15 16:16:42 +00:00
|
|
|
"Cannot reestablish SSL " +
|
|
|
|
"connection with upstream proxy: \r\n" +
|
|
|
|
str(resp.headers))
|
2014-03-13 00:04:45 +00:00
|
|
|
self.c.log("Hook: Establish SSL with upstream proxy", "debug")
|
2014-02-07 17:14:15 +00:00
|
|
|
self.c.establish_ssl(server=True)
|
|
|
|
|
|
|
|
self.c.server_reconnect = reconnect_http_proxy
|
|
|
|
|
|
|
|
def ssl_upgrade(self):
|
2014-01-30 04:00:13 +00:00
|
|
|
"""
|
|
|
|
Upgrade the connection to SSL after an authority (CONNECT) request has been made.
|
|
|
|
"""
|
2014-03-13 00:04:45 +00:00
|
|
|
self.c.log("Received CONNECT request. Upgrading to SSL...", "debug")
|
2014-03-10 01:32:27 +00:00
|
|
|
self.expected_form_in = "relative"
|
|
|
|
self.expected_form_out = "relative"
|
2014-01-30 04:00:13 +00:00
|
|
|
self.c.establish_ssl(server=True, client=True)
|
2014-03-13 00:04:45 +00:00
|
|
|
self.c.log("Upgrade to SSL completed.", "debug")
|
2014-01-30 04:00:13 +00:00
|
|
|
|
2014-02-08 04:17:35 +00:00
|
|
|
def process_request(self, flow, request):
|
2014-03-10 01:32:27 +00:00
|
|
|
|
|
|
|
if not self.skip_authentication:
|
2014-01-30 04:00:13 +00:00
|
|
|
self.authenticate(request)
|
|
|
|
|
|
|
|
if request.form_in == "authority":
|
2014-03-10 01:32:27 +00:00
|
|
|
if self.c.client_conn.ssl_established:
|
|
|
|
raise http.HttpError(400, "Must not CONNECT on already encrypted connection")
|
|
|
|
|
|
|
|
if self.expected_form_in == "absolute":
|
2014-03-10 04:11:51 +00:00
|
|
|
if not self.c.config.get_upstream_server:
|
2014-05-15 16:16:42 +00:00
|
|
|
self.c.set_server_address((request.host, request.port),
|
|
|
|
proxy.AddressPriority.FROM_PROTOCOL)
|
2014-02-08 04:17:35 +00:00
|
|
|
flow.server_conn = self.c.server_conn # Update server_conn attribute on the flow
|
2014-03-10 01:32:27 +00:00
|
|
|
self.c.client_conn.send(
|
|
|
|
'HTTP/1.1 200 Connection established\r\n' +
|
|
|
|
('Proxy-agent: %s\r\n' % self.c.server_version) +
|
|
|
|
'\r\n'
|
|
|
|
)
|
|
|
|
self.ssl_upgrade()
|
|
|
|
self.skip_authentication = True
|
|
|
|
return False
|
|
|
|
else:
|
|
|
|
self.hook_reconnect(request)
|
|
|
|
return True
|
|
|
|
elif request.form_in == self.expected_form_in:
|
|
|
|
if request.form_in == "absolute":
|
|
|
|
if request.scheme != "http":
|
|
|
|
raise http.HttpError(400, "Invalid request scheme: %s" % request.scheme)
|
|
|
|
|
2014-05-15 16:16:42 +00:00
|
|
|
self.c.set_server_address((request.host, request.port),
|
|
|
|
proxy.AddressPriority.FROM_PROTOCOL)
|
2014-03-10 01:32:27 +00:00
|
|
|
flow.server_conn = self.c.server_conn # Update server_conn attribute on the flow
|
|
|
|
|
|
|
|
request.form_out = self.expected_form_out
|
|
|
|
return True
|
|
|
|
|
2014-05-15 16:16:42 +00:00
|
|
|
raise http.HttpError(400, "Invalid HTTP request form (expected: %s, got: %s)" %
|
|
|
|
(self.expected_form_in, request.form_in))
|
2014-01-30 04:00:13 +00:00
|
|
|
|
|
|
|
def authenticate(self, request):
|
|
|
|
if self.c.config.authenticator:
|
|
|
|
if self.c.config.authenticator.authenticate(request.headers):
|
|
|
|
self.c.config.authenticator.clean(request.headers)
|
|
|
|
else:
|
2014-05-15 16:16:42 +00:00
|
|
|
raise HttpAuthenticationError(
|
|
|
|
self.c.config.authenticator.auth_challenge_headers())
|
2014-02-08 01:15:19 +00:00
|
|
|
return request.headers
|