2015-08-30 13:27:29 +00:00
|
|
|
from __future__ import (absolute_import, print_function, division)
|
2015-09-29 14:23:55 +00:00
|
|
|
|
2015-09-10 14:24:22 +00:00
|
|
|
import sys
|
2015-09-12 15:10:38 +00:00
|
|
|
import traceback
|
2015-09-10 23:18:17 +00:00
|
|
|
import six
|
2015-09-29 14:23:55 +00:00
|
|
|
import struct
|
|
|
|
import threading
|
2016-01-14 18:47:36 +00:00
|
|
|
import time
|
2015-09-29 14:23:55 +00:00
|
|
|
import Queue
|
2015-09-10 23:18:17 +00:00
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
from netlib import tcp
|
2015-11-01 17:37:20 +00:00
|
|
|
from netlib.exceptions import HttpException, HttpReadDisconnect, NetlibException
|
2015-09-29 14:23:55 +00:00
|
|
|
from netlib.http import http1, Headers, CONTENT_MISSING
|
|
|
|
from netlib.tcp import Address, ssl_read_select
|
|
|
|
|
|
|
|
import h2
|
|
|
|
from h2.connection import H2Connection
|
|
|
|
from h2.events import *
|
|
|
|
|
|
|
|
from .base import Layer, Kill
|
2015-08-30 13:27:29 +00:00
|
|
|
from .. import utils
|
2015-09-16 16:45:22 +00:00
|
|
|
from ..exceptions import HttpProtocolException, ProtocolException
|
2015-08-30 13:27:29 +00:00
|
|
|
from ..models import (
|
2015-09-29 14:23:55 +00:00
|
|
|
HTTPFlow,
|
|
|
|
HTTPRequest,
|
|
|
|
HTTPResponse,
|
|
|
|
make_error_response,
|
|
|
|
make_connect_response,
|
|
|
|
Error,
|
|
|
|
expect_continue_response
|
2015-08-30 13:27:29 +00:00
|
|
|
)
|
2015-09-10 23:18:17 +00:00
|
|
|
|
2014-01-30 04:00:13 +00:00
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
class _HttpLayer(Layer):
|
|
|
|
def read_request(self):
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
2015-09-28 08:59:10 +00:00
|
|
|
def read_request_body(self, request):
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def send_request(self, request):
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
2015-09-16 16:45:22 +00:00
|
|
|
def read_response(self, request):
|
2015-09-10 08:20:11 +00:00
|
|
|
response = self.read_response_headers()
|
2016-01-02 14:12:36 +00:00
|
|
|
response.data.content = b"".join(
|
2015-09-16 16:45:22 +00:00
|
|
|
self.read_response_body(request, response)
|
2015-09-10 08:20:11 +00:00
|
|
|
)
|
|
|
|
return response
|
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
def read_response_headers(self):
|
|
|
|
raise NotImplementedError()
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
def read_response_body(self, request, response):
|
2015-08-30 13:27:29 +00:00
|
|
|
raise NotImplementedError()
|
2015-09-29 14:23:55 +00:00
|
|
|
yield "this is a generator" # pragma: no cover
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2015-09-10 08:20:11 +00:00
|
|
|
def send_response(self, response):
|
2015-09-26 15:41:14 +00:00
|
|
|
if response.content == CONTENT_MISSING:
|
2015-09-16 16:45:22 +00:00
|
|
|
raise HttpException("Cannot assemble flow with CONTENT_MISSING")
|
2015-09-10 08:20:11 +00:00
|
|
|
self.send_response_headers(response)
|
2015-09-26 15:41:14 +00:00
|
|
|
self.send_response_body(response, [response.content])
|
2015-09-10 08:20:11 +00:00
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
def send_response_headers(self, response):
|
|
|
|
raise NotImplementedError()
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
def send_response_body(self, response, chunks):
|
|
|
|
raise NotImplementedError()
|
2016-01-27 09:12:18 +00:00
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
def check_close_connection(self, flow):
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
|
|
|
|
class Http1Layer(_HttpLayer):
|
2015-08-30 13:27:29 +00:00
|
|
|
def __init__(self, ctx, mode):
|
|
|
|
super(Http1Layer, self).__init__(ctx)
|
|
|
|
self.mode = mode
|
|
|
|
|
|
|
|
def read_request(self):
|
2015-09-16 16:45:22 +00:00
|
|
|
req = http1.read_request(self.client_conn.rfile, body_size_limit=self.config.body_size_limit)
|
|
|
|
return HTTPRequest.wrap(req)
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2015-09-28 08:59:10 +00:00
|
|
|
def read_request_body(self, request):
|
|
|
|
expected_size = http1.expected_http_body_size(request)
|
|
|
|
return http1.read_body(self.client_conn.rfile, expected_size, self.config.body_size_limit)
|
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def send_request(self, request):
|
2015-09-16 16:45:22 +00:00
|
|
|
self.server_conn.wfile.write(http1.assemble_request(request))
|
|
|
|
self.server_conn.wfile.flush()
|
2015-08-30 13:27:29 +00:00
|
|
|
|
|
|
|
def read_response_headers(self):
|
2015-09-16 16:45:22 +00:00
|
|
|
resp = http1.read_response_head(self.server_conn.rfile)
|
|
|
|
return HTTPResponse.wrap(resp)
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2015-09-16 16:45:22 +00:00
|
|
|
def read_response_body(self, request, response):
|
|
|
|
expected_size = http1.expected_http_body_size(request, response)
|
|
|
|
return http1.read_body(self.server_conn.rfile, expected_size, self.config.body_size_limit)
|
2015-08-30 13:27:29 +00:00
|
|
|
|
|
|
|
def send_response_headers(self, response):
|
2015-09-17 00:40:08 +00:00
|
|
|
raw = http1.assemble_response_head(response)
|
2015-09-16 16:45:22 +00:00
|
|
|
self.client_conn.wfile.write(raw)
|
2015-09-10 23:18:17 +00:00
|
|
|
self.client_conn.wfile.flush()
|
2015-08-30 13:27:29 +00:00
|
|
|
|
|
|
|
def send_response_body(self, response, chunks):
|
2015-09-17 00:40:08 +00:00
|
|
|
for chunk in http1.assemble_body(response.headers, chunks):
|
2015-09-10 23:18:17 +00:00
|
|
|
self.client_conn.wfile.write(chunk)
|
|
|
|
self.client_conn.wfile.flush()
|
2014-01-30 04:00:13 +00:00
|
|
|
|
2015-09-03 11:49:27 +00:00
|
|
|
def check_close_connection(self, flow):
|
2015-09-16 16:45:22 +00:00
|
|
|
request_close = http1.connection_close(
|
2015-09-17 13:16:33 +00:00
|
|
|
flow.request.http_version,
|
2015-09-16 16:45:22 +00:00
|
|
|
flow.request.headers
|
2015-09-03 11:49:27 +00:00
|
|
|
)
|
2015-09-16 16:45:22 +00:00
|
|
|
response_close = http1.connection_close(
|
2015-09-17 13:16:33 +00:00
|
|
|
flow.response.http_version,
|
2015-09-16 16:45:22 +00:00
|
|
|
flow.response.headers
|
|
|
|
)
|
|
|
|
read_until_eof = http1.expected_http_body_size(flow.request, flow.response) == -1
|
|
|
|
close_connection = request_close or response_close or read_until_eof
|
2015-09-17 13:16:33 +00:00
|
|
|
if flow.request.form_in == "authority" and flow.response.status_code == 200:
|
2015-09-16 16:45:22 +00:00
|
|
|
# Workaround for https://github.com/mitmproxy/mitmproxy/issues/313:
|
|
|
|
# Charles Proxy sends a CONNECT response with HTTP/1.0
|
2015-09-03 11:49:27 +00:00
|
|
|
# and no Content-Length header
|
|
|
|
|
|
|
|
return False
|
|
|
|
return close_connection
|
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def __call__(self):
|
|
|
|
layer = HttpLayer(self, self.mode)
|
|
|
|
layer()
|
|
|
|
|
2016-01-16 10:48:27 +00:00
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
class SafeH2Connection(H2Connection):
|
|
|
|
def __init__(self, conn, *args, **kwargs):
|
|
|
|
super(SafeH2Connection, self).__init__(*args, **kwargs)
|
|
|
|
self.conn = conn
|
|
|
|
self.lock = threading.RLock()
|
|
|
|
|
2016-01-14 18:05:15 +00:00
|
|
|
def safe_close_connection(self, error_code):
|
|
|
|
with self.lock:
|
|
|
|
self.close_connection(error_code)
|
|
|
|
self.conn.send(self.data_to_send())
|
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
def safe_increment_flow_control(self, stream_id, length):
|
2016-01-14 18:05:37 +00:00
|
|
|
if length == 0:
|
|
|
|
return
|
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
with self.lock:
|
|
|
|
self.increment_flow_control_window(length)
|
|
|
|
self.conn.send(self.data_to_send())
|
|
|
|
with self.lock:
|
|
|
|
if stream_id in self.streams and not self.streams[stream_id].closed:
|
|
|
|
self.increment_flow_control_window(length, stream_id=stream_id)
|
|
|
|
self.conn.send(self.data_to_send())
|
|
|
|
|
|
|
|
def safe_reset_stream(self, stream_id, error_code):
|
|
|
|
with self.lock:
|
2016-01-14 18:47:36 +00:00
|
|
|
try:
|
|
|
|
self.reset_stream(stream_id, error_code)
|
2016-01-18 20:44:27 +00:00
|
|
|
except h2.exceptions.StreamClosedError:
|
2016-01-14 18:47:36 +00:00
|
|
|
# stream is already closed - good
|
|
|
|
pass
|
2016-01-14 18:05:37 +00:00
|
|
|
self.conn.send(self.data_to_send())
|
2015-09-29 14:23:55 +00:00
|
|
|
|
|
|
|
def safe_update_settings(self, new_settings):
|
2016-01-20 18:31:37 +00:00
|
|
|
with self.lock:
|
2015-09-29 14:23:55 +00:00
|
|
|
self.update_settings(new_settings)
|
|
|
|
self.conn.send(self.data_to_send())
|
|
|
|
|
2016-01-16 10:31:43 +00:00
|
|
|
def safe_send_headers(self, is_zombie, stream_id, headers):
|
2015-09-29 14:23:55 +00:00
|
|
|
with self.lock:
|
2016-01-16 10:31:43 +00:00
|
|
|
if is_zombie(self, stream_id):
|
|
|
|
return
|
2015-09-29 14:23:55 +00:00
|
|
|
self.send_headers(stream_id, headers)
|
|
|
|
self.conn.send(self.data_to_send())
|
|
|
|
|
2016-01-16 10:31:43 +00:00
|
|
|
def safe_send_body(self, is_zombie, stream_id, chunks):
|
2015-09-29 14:23:55 +00:00
|
|
|
for chunk in chunks:
|
2016-01-16 10:31:43 +00:00
|
|
|
position = 0
|
|
|
|
while position < len(chunk):
|
2016-01-14 18:05:37 +00:00
|
|
|
self.lock.acquire()
|
2016-01-19 21:22:36 +00:00
|
|
|
if is_zombie(self, stream_id):
|
|
|
|
return
|
2016-01-16 10:31:43 +00:00
|
|
|
max_outbound_frame_size = self.max_outbound_frame_size
|
|
|
|
frame_chunk = chunk[position:position+max_outbound_frame_size]
|
|
|
|
if self.local_flow_control_window(stream_id) < len(frame_chunk):
|
|
|
|
self.lock.release()
|
|
|
|
time.sleep(0)
|
|
|
|
continue
|
2016-01-14 18:05:37 +00:00
|
|
|
self.send_data(stream_id, frame_chunk)
|
|
|
|
self.conn.send(self.data_to_send())
|
|
|
|
self.lock.release()
|
2016-01-16 10:31:43 +00:00
|
|
|
position += max_outbound_frame_size
|
2015-09-29 14:23:55 +00:00
|
|
|
with self.lock:
|
2016-01-16 10:31:43 +00:00
|
|
|
if is_zombie(self, stream_id):
|
|
|
|
return
|
2015-09-29 14:23:55 +00:00
|
|
|
self.end_stream(stream_id)
|
|
|
|
self.conn.send(self.data_to_send())
|
|
|
|
|
2016-01-16 10:48:27 +00:00
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
class Http2Layer(Layer):
|
2015-08-30 13:27:29 +00:00
|
|
|
def __init__(self, ctx, mode):
|
|
|
|
super(Http2Layer, self).__init__(ctx)
|
|
|
|
self.mode = mode
|
2015-09-29 14:23:55 +00:00
|
|
|
self.streams = dict()
|
|
|
|
self.server_to_client_stream_ids = dict([(0, 0)])
|
|
|
|
self.client_conn.h2 = SafeH2Connection(self.client_conn, client_side=False)
|
|
|
|
|
|
|
|
# make sure that we only pass actual SSL.Connection objects in here,
|
|
|
|
# because otherwise ssl_read_select fails!
|
|
|
|
self.active_conns = [self.client_conn.connection]
|
|
|
|
|
|
|
|
def _initiate_server_conn(self):
|
|
|
|
self.server_conn.h2 = SafeH2Connection(self.server_conn, client_side=True)
|
|
|
|
self.server_conn.h2.initiate_connection()
|
|
|
|
self.server_conn.send(self.server_conn.h2.data_to_send())
|
|
|
|
self.active_conns.append(self.server_conn.connection)
|
|
|
|
|
|
|
|
def connect(self):
|
|
|
|
self.ctx.connect()
|
|
|
|
self.server_conn.connect()
|
|
|
|
self._initiate_server_conn()
|
|
|
|
|
|
|
|
def set_server(self):
|
|
|
|
raise NotImplementedError("Cannot change server for HTTP2 connections.")
|
|
|
|
|
|
|
|
def disconnect(self):
|
|
|
|
raise NotImplementedError("Cannot dis- or reconnect in HTTP2 connections.")
|
|
|
|
|
2016-01-16 10:48:27 +00:00
|
|
|
def next_layer(self):
|
|
|
|
# WebSockets over HTTP/2?
|
|
|
|
# CONNECT for proxying?
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
2016-01-17 19:18:53 +00:00
|
|
|
def _handle_event(self, event, source_conn, other_conn, is_server):
|
|
|
|
if hasattr(event, 'stream_id'):
|
2016-01-20 09:28:04 +00:00
|
|
|
if is_server and event.stream_id % 2 == 1:
|
2016-01-17 19:18:53 +00:00
|
|
|
eid = self.server_to_client_stream_ids[event.stream_id]
|
|
|
|
else:
|
|
|
|
eid = event.stream_id
|
|
|
|
|
|
|
|
if isinstance(event, RequestReceived):
|
|
|
|
headers = Headers([[str(k), str(v)] for k, v in event.headers])
|
|
|
|
self.streams[eid] = Http2SingleStreamLayer(self, eid, headers)
|
|
|
|
self.streams[eid].timestamp_start = time.time()
|
|
|
|
self.streams[eid].start()
|
|
|
|
elif isinstance(event, ResponseReceived):
|
|
|
|
headers = Headers([[str(k), str(v)] for k, v in event.headers])
|
2016-01-19 21:10:19 +00:00
|
|
|
self.streams[eid].queued_data_length = 0
|
2016-01-17 19:18:53 +00:00
|
|
|
self.streams[eid].timestamp_start = time.time()
|
|
|
|
self.streams[eid].response_headers = headers
|
|
|
|
self.streams[eid].response_arrived.set()
|
|
|
|
elif isinstance(event, DataReceived):
|
2016-01-19 21:10:19 +00:00
|
|
|
if self.config.body_size_limit and self.streams[eid].queued_data_length > self.config.body_size_limit:
|
|
|
|
raise HttpException("HTTP body too large. Limit is {}.".format(self.config.body_size_limit))
|
2016-01-17 19:18:53 +00:00
|
|
|
self.streams[eid].data_queue.put(event.data)
|
2016-01-19 21:10:19 +00:00
|
|
|
self.streams[eid].queued_data_length += len(event.data)
|
2016-01-17 19:18:53 +00:00
|
|
|
source_conn.h2.safe_increment_flow_control(event.stream_id, len(event.data))
|
|
|
|
elif isinstance(event, StreamEnded):
|
|
|
|
self.streams[eid].timestamp_end = time.time()
|
|
|
|
self.streams[eid].data_finished.set()
|
|
|
|
elif isinstance(event, StreamReset):
|
|
|
|
self.streams[eid].zombie = time.time()
|
|
|
|
if eid in self.streams and event.error_code == 0x8:
|
|
|
|
if is_server:
|
|
|
|
other_stream_id = self.streams[eid].client_stream_id
|
|
|
|
else:
|
|
|
|
other_stream_id = self.streams[eid].server_stream_id
|
|
|
|
other_conn.h2.safe_reset_stream(other_stream_id, event.error_code)
|
|
|
|
elif isinstance(event, RemoteSettingsChanged):
|
|
|
|
new_settings = dict([(id, cs.new_value) for (id, cs) in event.changed_settings.iteritems()])
|
|
|
|
other_conn.h2.safe_update_settings(new_settings)
|
|
|
|
elif isinstance(event, ConnectionTerminated):
|
|
|
|
other_conn.h2.safe_close_connection(event.error_code)
|
2016-01-18 21:20:47 +00:00
|
|
|
return False
|
2016-01-17 19:18:53 +00:00
|
|
|
elif isinstance(event, PushedStreamReceived):
|
2016-01-20 09:28:04 +00:00
|
|
|
# pushed stream ids should be uniq and not dependent on race conditions
|
|
|
|
# only the parent stream id must be looked up first
|
|
|
|
parent_eid = self.server_to_client_stream_ids[event.parent_stream_id]
|
|
|
|
with self.client_conn.h2.lock:
|
|
|
|
self.client_conn.h2.push_stream(parent_eid, event.pushed_stream_id, event.headers)
|
|
|
|
|
|
|
|
headers = Headers([[str(k), str(v)] for k, v in event.headers])
|
|
|
|
headers['x-mitmproxy-pushed'] = 'true'
|
|
|
|
self.streams[event.pushed_stream_id] = Http2SingleStreamLayer(self, event.pushed_stream_id, headers)
|
|
|
|
self.streams[event.pushed_stream_id].timestamp_start = time.time()
|
|
|
|
self.streams[event.pushed_stream_id].pushed = True
|
|
|
|
self.streams[event.pushed_stream_id].parent_stream_id = parent_eid
|
|
|
|
self.streams[event.pushed_stream_id].timestamp_end = time.time()
|
|
|
|
self.streams[event.pushed_stream_id].data_finished.set()
|
|
|
|
self.streams[event.pushed_stream_id].start()
|
|
|
|
elif isinstance(event, TrailersReceived):
|
2016-01-17 19:18:53 +00:00
|
|
|
raise NotImplementedError()
|
|
|
|
|
2016-01-18 21:20:47 +00:00
|
|
|
return True
|
|
|
|
|
2016-01-17 19:18:53 +00:00
|
|
|
def _cleanup_streams(self):
|
|
|
|
death_time = time.time() - 10
|
|
|
|
for stream_id in self.streams.keys():
|
|
|
|
zombie = self.streams[stream_id].zombie
|
|
|
|
if zombie and zombie <= death_time:
|
|
|
|
self.streams.pop(stream_id, None)
|
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
def __call__(self):
|
2016-01-16 10:48:27 +00:00
|
|
|
if self.server_conn:
|
|
|
|
self._initiate_server_conn()
|
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
preamble = self.client_conn.rfile.read(24)
|
|
|
|
self.client_conn.h2.initiate_connection()
|
|
|
|
self.client_conn.h2.receive_data(preamble)
|
|
|
|
self.client_conn.send(self.client_conn.h2.data_to_send())
|
|
|
|
|
|
|
|
while True:
|
|
|
|
r = ssl_read_select(self.active_conns, 1)
|
|
|
|
for conn in r:
|
|
|
|
source_conn = self.client_conn if conn == self.client_conn.connection else self.server_conn
|
|
|
|
other_conn = self.server_conn if conn == self.client_conn.connection else self.client_conn
|
2016-01-18 21:20:47 +00:00
|
|
|
is_server = (conn == self.server_conn.connection)
|
2015-09-29 14:23:55 +00:00
|
|
|
|
2016-01-17 19:18:53 +00:00
|
|
|
field = source_conn.rfile.peek(3)
|
2016-01-18 21:20:47 +00:00
|
|
|
length = int(field.encode('hex'), 16)
|
2015-09-29 14:23:55 +00:00
|
|
|
raw_frame = source_conn.rfile.safe_read(9 + length)
|
|
|
|
|
|
|
|
with source_conn.h2.lock:
|
|
|
|
events = source_conn.h2.receive_data(raw_frame)
|
|
|
|
source_conn.send(source_conn.h2.data_to_send())
|
|
|
|
|
2016-01-16 10:31:43 +00:00
|
|
|
for event in events:
|
2016-01-18 21:20:47 +00:00
|
|
|
if not self._handle_event(event, source_conn, other_conn, is_server):
|
|
|
|
return
|
2016-01-17 19:18:53 +00:00
|
|
|
|
2016-01-18 21:20:47 +00:00
|
|
|
self._cleanup_streams()
|
2015-09-29 14:23:55 +00:00
|
|
|
|
|
|
|
|
|
|
|
class Http2SingleStreamLayer(_HttpLayer, threading.Thread):
|
|
|
|
def __init__(self, ctx, stream_id, request_headers):
|
|
|
|
super(Http2SingleStreamLayer, self).__init__(ctx)
|
2016-01-14 18:47:36 +00:00
|
|
|
self.zombie = None
|
2015-09-29 14:23:55 +00:00
|
|
|
self.client_stream_id = stream_id
|
|
|
|
self.server_stream_id = None
|
|
|
|
self.request_headers = request_headers
|
|
|
|
self.response_headers = None
|
|
|
|
self.data_queue = Queue.Queue()
|
2016-01-19 21:10:19 +00:00
|
|
|
self.queued_data_length = 0
|
2015-09-29 14:23:55 +00:00
|
|
|
|
|
|
|
self.response_arrived = threading.Event()
|
|
|
|
self.data_finished = threading.Event()
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2016-01-16 10:31:43 +00:00
|
|
|
def is_zombie(self, h2_conn, stream_id):
|
|
|
|
if self.zombie:
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def read_request(self):
|
2015-09-29 14:23:55 +00:00
|
|
|
self.data_finished.wait()
|
|
|
|
self.data_finished.clear()
|
|
|
|
|
|
|
|
authority = self.request_headers.get(':authority', '')
|
|
|
|
method = self.request_headers.get(':method', 'GET')
|
|
|
|
scheme = self.request_headers.get(':scheme', 'https')
|
|
|
|
path = self.request_headers.get(':path', '/')
|
|
|
|
host = None
|
|
|
|
port = None
|
|
|
|
|
|
|
|
if path == '*' or path.startswith("/"):
|
|
|
|
form_in = "relative"
|
|
|
|
elif method == 'CONNECT':
|
|
|
|
form_in = "authority"
|
|
|
|
if ":" in authority:
|
|
|
|
host, port = authority.split(":", 1)
|
|
|
|
else:
|
|
|
|
host = authority
|
|
|
|
else:
|
|
|
|
form_in = "absolute"
|
|
|
|
# FIXME: verify if path or :host contains what we need
|
|
|
|
scheme, host, port, _ = utils.parse_url(path)
|
|
|
|
|
|
|
|
if host is None:
|
|
|
|
host = 'localhost'
|
|
|
|
if port is None:
|
|
|
|
port = 80 if scheme == 'http' else 443
|
|
|
|
port = int(port)
|
|
|
|
|
|
|
|
data = []
|
|
|
|
while self.data_queue.qsize() > 0:
|
|
|
|
data.append(self.data_queue.get())
|
2016-01-19 21:10:19 +00:00
|
|
|
data = b"".join(data)
|
2015-09-29 14:23:55 +00:00
|
|
|
|
|
|
|
return HTTPRequest(
|
|
|
|
form_in,
|
|
|
|
method,
|
|
|
|
scheme,
|
|
|
|
host,
|
|
|
|
port,
|
|
|
|
path,
|
|
|
|
(2, 0),
|
|
|
|
self.request_headers,
|
|
|
|
data,
|
2016-01-16 10:48:27 +00:00
|
|
|
timestamp_start=self.timestamp_start,
|
|
|
|
timestamp_end=self.timestamp_end,
|
2015-08-30 13:27:29 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
def send_request(self, message):
|
2015-09-29 14:23:55 +00:00
|
|
|
with self.server_conn.h2.lock:
|
|
|
|
self.server_stream_id = self.server_conn.h2.get_next_available_stream_id()
|
|
|
|
self.server_to_client_stream_ids[self.server_stream_id] = self.client_stream_id
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
self.server_conn.h2.safe_send_headers(
|
2016-01-16 10:31:43 +00:00
|
|
|
self.is_zombie,
|
2015-09-29 14:23:55 +00:00
|
|
|
self.server_stream_id,
|
|
|
|
message.headers
|
|
|
|
)
|
2016-01-14 18:47:36 +00:00
|
|
|
self.server_conn.h2.safe_send_body(
|
2016-01-16 10:31:43 +00:00
|
|
|
self.is_zombie,
|
2016-01-14 18:47:36 +00:00
|
|
|
self.server_stream_id,
|
|
|
|
message.body
|
|
|
|
)
|
2015-09-29 14:23:55 +00:00
|
|
|
|
|
|
|
def read_response_headers(self):
|
|
|
|
self.response_arrived.wait()
|
|
|
|
|
|
|
|
status_code = int(self.response_headers.get(':status', 502))
|
|
|
|
|
|
|
|
return HTTPResponse(
|
|
|
|
http_version=(2, 0),
|
|
|
|
status_code=status_code,
|
|
|
|
reason='',
|
|
|
|
headers=self.response_headers,
|
|
|
|
content=None,
|
2016-01-16 10:48:27 +00:00
|
|
|
timestamp_start=self.timestamp_start,
|
|
|
|
timestamp_end=self.timestamp_end,
|
2015-09-29 14:23:55 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
def read_response_body(self, request, response):
|
|
|
|
while True:
|
|
|
|
try:
|
|
|
|
yield self.data_queue.get(timeout=1)
|
|
|
|
except Queue.Empty:
|
|
|
|
pass
|
|
|
|
if self.data_finished.is_set():
|
|
|
|
while self.data_queue.qsize() > 0:
|
|
|
|
yield self.data_queue.get()
|
|
|
|
return
|
2016-01-14 18:47:36 +00:00
|
|
|
if self.zombie:
|
|
|
|
return
|
2015-09-29 14:23:55 +00:00
|
|
|
|
|
|
|
def send_response_headers(self, response):
|
|
|
|
self.client_conn.h2.safe_send_headers(
|
2016-01-16 10:31:43 +00:00
|
|
|
self.is_zombie,
|
2015-09-29 14:23:55 +00:00
|
|
|
self.client_stream_id,
|
|
|
|
response.headers
|
2015-08-30 13:27:29 +00:00
|
|
|
)
|
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
def send_response_body(self, _response, chunks):
|
|
|
|
self.client_conn.h2.safe_send_body(
|
2016-01-16 10:31:43 +00:00
|
|
|
self.is_zombie,
|
2015-09-29 14:23:55 +00:00
|
|
|
self.client_stream_id,
|
|
|
|
chunks
|
|
|
|
)
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2015-09-03 11:49:27 +00:00
|
|
|
def check_close_connection(self, flow):
|
2015-09-29 14:23:55 +00:00
|
|
|
# This layer only handles a single stream.
|
|
|
|
# RFC 7540 8.1: An HTTP request/response exchange fully consumes a single stream.
|
|
|
|
return True
|
2015-09-03 11:49:27 +00:00
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def connect(self):
|
2015-09-29 14:23:55 +00:00
|
|
|
raise ValueError("CONNECT inside an HTTP2 stream is not supported.")
|
2015-08-30 13:27:29 +00:00
|
|
|
|
|
|
|
def set_server(self, *args, **kwargs):
|
2015-09-29 14:23:55 +00:00
|
|
|
# do not mess with the server connection - all streams share it.
|
|
|
|
pass
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
def run(self):
|
2015-08-30 13:27:29 +00:00
|
|
|
layer = HttpLayer(self, self.mode)
|
|
|
|
layer()
|
2016-01-14 18:47:36 +00:00
|
|
|
self.zombie = time.time()
|
2015-08-30 13:27:29 +00:00
|
|
|
|
|
|
|
|
|
|
|
class ConnectServerConnection(object):
|
2016-01-27 09:12:18 +00:00
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
"""
|
|
|
|
"Fake" ServerConnection to represent state after a CONNECT request to an upstream proxy.
|
2014-01-30 04:00:13 +00:00
|
|
|
"""
|
2014-05-15 16:16:42 +00:00
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def __init__(self, address, ctx):
|
|
|
|
self.address = tcp.Address.wrap(address)
|
|
|
|
self._ctx = ctx
|
2014-01-30 04:21:53 +00:00
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
@property
|
|
|
|
def via(self):
|
|
|
|
return self._ctx.server_conn
|
2014-08-08 00:45:24 +00:00
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def __getattr__(self, item):
|
|
|
|
return getattr(self.via, item)
|
2014-01-30 04:21:53 +00:00
|
|
|
|
2015-08-30 22:14:42 +00:00
|
|
|
def __nonzero__(self):
|
|
|
|
return bool(self.via)
|
|
|
|
|
2014-01-30 04:21:53 +00:00
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
class UpstreamConnectLayer(Layer):
|
2016-01-27 09:12:18 +00:00
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def __init__(self, ctx, connect_request):
|
|
|
|
super(UpstreamConnectLayer, self).__init__(ctx)
|
|
|
|
self.connect_request = connect_request
|
|
|
|
self.server_conn = ConnectServerConnection(
|
|
|
|
(connect_request.host, connect_request.port),
|
|
|
|
self.ctx
|
|
|
|
)
|
|
|
|
|
|
|
|
def __call__(self):
|
|
|
|
layer = self.ctx.next_layer(self)
|
|
|
|
layer()
|
|
|
|
|
2015-08-30 22:14:42 +00:00
|
|
|
def _send_connect_request(self):
|
|
|
|
self.send_request(self.connect_request)
|
2015-09-16 16:45:22 +00:00
|
|
|
resp = self.read_response(self.connect_request)
|
2015-09-17 13:16:33 +00:00
|
|
|
if resp.status_code != 200:
|
2015-08-30 22:14:42 +00:00
|
|
|
raise ProtocolException("Reconnect: Upstream server refuses CONNECT request")
|
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def connect(self):
|
|
|
|
if not self.server_conn:
|
|
|
|
self.ctx.connect()
|
2015-08-30 22:14:42 +00:00
|
|
|
self._send_connect_request()
|
2015-08-30 13:27:29 +00:00
|
|
|
else:
|
|
|
|
pass # swallow the message
|
|
|
|
|
2015-09-03 16:25:36 +00:00
|
|
|
def change_upstream_proxy_server(self, address):
|
|
|
|
if address != self.server_conn.via.address:
|
|
|
|
self.ctx.set_server(address)
|
|
|
|
|
|
|
|
def set_server(self, address, server_tls=None, sni=None):
|
|
|
|
if self.ctx.server_conn:
|
|
|
|
self.ctx.disconnect()
|
|
|
|
address = Address.wrap(address)
|
|
|
|
self.connect_request.host = address.host
|
|
|
|
self.connect_request.port = address.port
|
|
|
|
self.server_conn.address = address
|
|
|
|
|
|
|
|
if server_tls:
|
|
|
|
raise ProtocolException(
|
|
|
|
"Cannot upgrade to TLS, no TLS layer on the protocol stack."
|
|
|
|
)
|
2015-08-30 13:27:29 +00:00
|
|
|
|
|
|
|
|
|
|
|
class HttpLayer(Layer):
|
2016-01-27 09:12:18 +00:00
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def __init__(self, ctx, mode):
|
|
|
|
super(HttpLayer, self).__init__(ctx)
|
|
|
|
self.mode = mode
|
|
|
|
self.__original_server_conn = None
|
|
|
|
"Contains the original destination in transparent mode, which needs to be restored"
|
|
|
|
"if an inline script modified the target server for a single http request"
|
|
|
|
|
|
|
|
def __call__(self):
|
|
|
|
if self.mode == "transparent":
|
|
|
|
self.__original_server_conn = self.server_conn
|
|
|
|
while True:
|
|
|
|
try:
|
2015-09-28 08:59:10 +00:00
|
|
|
request = self.get_request_from_client()
|
2015-09-16 16:45:22 +00:00
|
|
|
self.log("request", "debug", [repr(request)])
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2015-09-16 16:45:22 +00:00
|
|
|
# Handle Proxy Authentication
|
2015-11-14 20:55:02 +00:00
|
|
|
# Proxy Authentication conceptually does not work in transparent mode.
|
|
|
|
# We catch this misconfiguration on startup. Here, we sort out requests
|
|
|
|
# after a successful CONNECT request (which do not need to be validated anymore)
|
|
|
|
if self.mode != "transparent" and not self.authenticate(request):
|
2015-08-30 13:27:29 +00:00
|
|
|
return
|
|
|
|
|
2015-09-16 16:45:22 +00:00
|
|
|
# Make sure that the incoming request matches our expectations
|
|
|
|
self.validate_request(request)
|
2015-08-30 13:27:29 +00:00
|
|
|
|
2015-11-26 22:19:43 +00:00
|
|
|
# Regular Proxy Mode: Handle CONNECT
|
|
|
|
if self.mode == "regular" and request.form_in == "authority":
|
|
|
|
self.handle_regular_mode_connect(request)
|
|
|
|
return
|
|
|
|
|
2015-09-16 16:45:22 +00:00
|
|
|
except HttpReadDisconnect:
|
|
|
|
# don't throw an error for disconnects that happen before/between requests.
|
|
|
|
return
|
2015-11-01 17:37:20 +00:00
|
|
|
except NetlibException as e:
|
2015-09-16 16:45:22 +00:00
|
|
|
self.send_error_response(400, repr(e))
|
2016-01-27 09:12:18 +00:00
|
|
|
six.reraise(ProtocolException, ProtocolException(
|
|
|
|
"Error in HTTP connection: %s" % repr(e)), sys.exc_info()[2])
|
2015-09-16 16:45:22 +00:00
|
|
|
|
|
|
|
try:
|
|
|
|
flow = HTTPFlow(self.client_conn, self.server_conn, live=self)
|
2015-08-30 13:27:29 +00:00
|
|
|
flow.request = request
|
|
|
|
self.process_request_hook(flow)
|
|
|
|
|
|
|
|
if not flow.response:
|
|
|
|
self.establish_server_connection(flow)
|
|
|
|
self.get_response_from_server(flow)
|
2015-09-18 11:51:05 +00:00
|
|
|
else:
|
|
|
|
# response was set by an inline script.
|
|
|
|
# we now need to emulate the responseheaders hook.
|
|
|
|
flow = self.channel.ask("responseheaders", flow)
|
|
|
|
if flow == Kill:
|
|
|
|
raise Kill()
|
|
|
|
|
|
|
|
self.log("response", "debug", [repr(flow.response)])
|
|
|
|
flow = self.channel.ask("response", flow)
|
|
|
|
if flow == Kill:
|
|
|
|
raise Kill()
|
2015-08-30 13:27:29 +00:00
|
|
|
self.send_response_to_client(flow)
|
|
|
|
|
|
|
|
if self.check_close_connection(flow):
|
|
|
|
return
|
|
|
|
|
2015-09-10 22:00:00 +00:00
|
|
|
# Handle 101 Switching Protocols
|
|
|
|
# It may be useful to pass additional args (such as the upgrade header)
|
|
|
|
# to next_layer in the future
|
|
|
|
if flow.response.status_code == 101:
|
|
|
|
layer = self.ctx.next_layer(self)
|
|
|
|
layer()
|
|
|
|
return
|
2015-08-30 13:27:29 +00:00
|
|
|
|
|
|
|
# Upstream Proxy Mode: Handle CONNECT
|
2015-09-17 13:16:33 +00:00
|
|
|
if flow.request.form_in == "authority" and flow.response.status_code == 200:
|
2015-08-30 13:27:29 +00:00
|
|
|
self.handle_upstream_mode_connect(flow.request.copy())
|
|
|
|
return
|
|
|
|
|
2015-11-26 22:19:43 +00:00
|
|
|
except (ProtocolException, NetlibException) as e:
|
2015-09-16 16:45:22 +00:00
|
|
|
self.send_error_response(502, repr(e))
|
|
|
|
|
|
|
|
if not flow.response:
|
2015-09-12 15:10:38 +00:00
|
|
|
flow.error = Error(str(e))
|
2015-08-30 13:27:29 +00:00
|
|
|
self.channel.ask("error", flow)
|
2015-09-12 15:10:38 +00:00
|
|
|
self.log(traceback.format_exc(), "debug")
|
2015-09-16 16:45:22 +00:00
|
|
|
return
|
|
|
|
else:
|
2016-01-27 09:12:18 +00:00
|
|
|
six.reraise(ProtocolException, ProtocolException(
|
|
|
|
"Error in HTTP connection: %s" % repr(e)), sys.exc_info()[2])
|
2015-08-30 13:27:29 +00:00
|
|
|
finally:
|
|
|
|
flow.live = False
|
|
|
|
|
2015-09-28 08:59:10 +00:00
|
|
|
def get_request_from_client(self):
|
|
|
|
request = self.read_request()
|
|
|
|
if request.headers.get("expect", "").lower() == "100-continue":
|
2015-10-03 12:48:42 +00:00
|
|
|
# TODO: We may have to use send_response_headers for HTTP2 here.
|
2015-09-28 08:59:10 +00:00
|
|
|
self.send_response(expect_continue_response)
|
|
|
|
request.headers.pop("expect")
|
|
|
|
request.body = b"".join(self.read_request_body(request))
|
|
|
|
return request
|
|
|
|
|
2015-09-16 16:45:22 +00:00
|
|
|
def send_error_response(self, code, message):
|
|
|
|
try:
|
|
|
|
response = make_error_response(code, message)
|
|
|
|
self.send_response(response)
|
2016-01-17 16:41:39 +00:00
|
|
|
except NetlibException, h2.exceptions.H2Error:
|
2015-09-16 16:45:22 +00:00
|
|
|
pass
|
|
|
|
|
2015-09-03 16:25:36 +00:00
|
|
|
def change_upstream_proxy_server(self, address):
|
|
|
|
# Make set_upstream_proxy_server always available,
|
|
|
|
# even if there's no UpstreamConnectLayer
|
|
|
|
if address != self.server_conn.address:
|
|
|
|
return self.set_server(address)
|
|
|
|
|
2015-08-30 13:27:29 +00:00
|
|
|
def handle_regular_mode_connect(self, request):
|
|
|
|
self.set_server((request.host, request.port))
|
2015-09-17 13:16:33 +00:00
|
|
|
self.send_response(make_connect_response(request.http_version))
|
2015-08-30 13:27:29 +00:00
|
|
|
layer = self.ctx.next_layer(self)
|
|
|
|
layer()
|
|
|
|
|
|
|
|
def handle_upstream_mode_connect(self, connect_request):
|
|
|
|
layer = UpstreamConnectLayer(self, connect_request)
|
|
|
|
layer()
|
|
|
|
|
|
|
|
def send_response_to_client(self, flow):
|
2015-09-29 14:23:55 +00:00
|
|
|
if not flow.response.stream:
|
2015-08-30 13:27:29 +00:00
|
|
|
# no streaming:
|
|
|
|
# we already received the full response from the server and can
|
|
|
|
# send it to the client straight away.
|
|
|
|
self.send_response(flow.response)
|
|
|
|
else:
|
|
|
|
# streaming:
|
|
|
|
# First send the headers and then transfer the response incrementally
|
|
|
|
self.send_response_headers(flow.response)
|
|
|
|
chunks = self.read_response_body(
|
2015-09-16 16:45:22 +00:00
|
|
|
flow.request,
|
|
|
|
flow.response
|
2015-08-30 13:27:29 +00:00
|
|
|
)
|
|
|
|
if callable(flow.response.stream):
|
|
|
|
chunks = flow.response.stream(chunks)
|
|
|
|
self.send_response_body(flow.response, chunks)
|
|
|
|
flow.response.timestamp_end = utils.timestamp()
|
|
|
|
|
|
|
|
def get_response_from_server(self, flow):
|
|
|
|
def get_response():
|
|
|
|
self.send_request(flow.request)
|
2015-09-29 14:23:55 +00:00
|
|
|
flow.response = self.read_response_headers()
|
2015-08-30 13:27:29 +00:00
|
|
|
|
|
|
|
try:
|
|
|
|
get_response()
|
2015-11-01 17:37:20 +00:00
|
|
|
except NetlibException as v:
|
2015-08-30 13:27:29 +00:00
|
|
|
self.log(
|
|
|
|
"server communication error: %s" % repr(v),
|
|
|
|
level="debug"
|
|
|
|
)
|
|
|
|
# In any case, we try to reconnect at least once. This is
|
|
|
|
# necessary because it might be possible that we already
|
|
|
|
# initiated an upstream connection after clientconnect that
|
|
|
|
# has already been expired, e.g consider the following event
|
|
|
|
# log:
|
|
|
|
# > clientconnect (transparent mode destination known)
|
|
|
|
# > serverconnect (required for client tls handshake)
|
|
|
|
# > read n% of large request
|
|
|
|
# > server detects timeout, disconnects
|
|
|
|
# > read (100-n)% of large request
|
|
|
|
# > send large request upstream
|
2015-09-03 15:01:25 +00:00
|
|
|
self.disconnect()
|
|
|
|
self.connect()
|
2015-08-30 13:27:29 +00:00
|
|
|
get_response()
|
|
|
|
|
|
|
|
# call the appropriate script hook - this is an opportunity for an
|
|
|
|
# inline script to set flow.stream = True
|
|
|
|
flow = self.channel.ask("responseheaders", flow)
|
2015-08-31 15:05:52 +00:00
|
|
|
if flow == Kill:
|
2015-08-30 13:27:29 +00:00
|
|
|
raise Kill()
|
|
|
|
|
2015-09-29 14:23:55 +00:00
|
|
|
if flow.response.stream:
|
|
|
|
flow.response.data.content = CONTENT_MISSING
|
|
|
|
else:
|
|
|
|
flow.response.data.content = b"".join(self.read_response_body(
|
|
|
|
flow.request,
|
|
|
|
flow.response
|
|
|
|
))
|
|
|
|
flow.response.timestamp_end = utils.timestamp()
|
2015-08-30 13:27:29 +00:00
|
|
|
|
|
|
|
# no further manipulation of self.server_conn beyond this point
|
|
|
|
# we can safely set it as the final attribute value here.
|
|
|
|
flow.server_conn = self.server_conn
|
|
|
|
|
|
|
|
def process_request_hook(self, flow):
|
|
|
|
# Determine .scheme, .host and .port attributes for inline scripts.
|
|
|
|
# For absolute-form requests, they are directly given in the request.
|
|
|
|
# For authority-form requests, we only need to determine the request scheme.
|
|
|
|
# For relative-form requests, we need to determine host and port as
|
|
|
|
# well.
|
|
|
|
if self.mode == "regular":
|
|
|
|
pass # only absolute-form at this point, nothing to do here.
|
|
|
|
elif self.mode == "upstream":
|
|
|
|
if flow.request.form_in == "authority":
|
|
|
|
flow.request.scheme = "http" # pseudo value
|
|
|
|
else:
|
2015-10-03 12:48:42 +00:00
|
|
|
# Setting request.host also updates the host header, which we want to preserve
|
|
|
|
host_header = flow.request.headers.get("host", None)
|
2015-08-30 13:27:29 +00:00
|
|
|
flow.request.host = self.__original_server_conn.address.host
|
|
|
|
flow.request.port = self.__original_server_conn.address.port
|
2015-10-03 12:48:42 +00:00
|
|
|
if host_header:
|
|
|
|
flow.request.headers["host"] = host_header
|
2015-09-28 08:59:10 +00:00
|
|
|
# TODO: This does not really work if we change the first request and --no-upstream-cert is enabled
|
2015-08-30 13:27:29 +00:00
|
|
|
flow.request.scheme = "https" if self.__original_server_conn.tls_established else "http"
|
|
|
|
|
|
|
|
request_reply = self.channel.ask("request", flow)
|
2015-08-31 15:05:52 +00:00
|
|
|
if request_reply == Kill:
|
2015-08-30 13:27:29 +00:00
|
|
|
raise Kill()
|
|
|
|
if isinstance(request_reply, HTTPResponse):
|
|
|
|
flow.response = request_reply
|
|
|
|
return
|
|
|
|
|
|
|
|
def establish_server_connection(self, flow):
|
|
|
|
address = tcp.Address((flow.request.host, flow.request.port))
|
|
|
|
tls = (flow.request.scheme == "https")
|
|
|
|
|
|
|
|
if self.mode == "regular" or self.mode == "transparent":
|
|
|
|
# If there's an existing connection that doesn't match our expectations, kill it.
|
2015-09-09 16:49:32 +00:00
|
|
|
if address != self.server_conn.address or tls != self.server_conn.tls_established:
|
2015-08-30 13:27:29 +00:00
|
|
|
self.set_server(address, tls, address.host)
|
|
|
|
# Establish connection is neccessary.
|
|
|
|
if not self.server_conn:
|
|
|
|
self.connect()
|
|
|
|
else:
|
|
|
|
if not self.server_conn:
|
|
|
|
self.connect()
|
|
|
|
if tls:
|
2015-09-16 16:45:22 +00:00
|
|
|
raise HttpProtocolException("Cannot change scheme in upstream proxy mode.")
|
2015-08-30 13:27:29 +00:00
|
|
|
"""
|
|
|
|
# This is a very ugly (untested) workaround to solve a very ugly problem.
|
|
|
|
if self.server_conn and self.server_conn.tls_established and not ssl:
|
2015-09-03 15:01:25 +00:00
|
|
|
self.disconnect()
|
|
|
|
self.connect()
|
2015-08-30 13:27:29 +00:00
|
|
|
elif ssl and not hasattr(self, "connected_to") or self.connected_to != address:
|
|
|
|
if self.server_conn.tls_established:
|
2015-09-03 15:01:25 +00:00
|
|
|
self.disconnect()
|
|
|
|
self.connect()
|
2015-08-30 13:27:29 +00:00
|
|
|
|
|
|
|
self.send_request(make_connect_request(address))
|
|
|
|
tls_layer = TlsLayer(self, False, True)
|
|
|
|
tls_layer._establish_tls_with_server()
|
|
|
|
"""
|
|
|
|
|
|
|
|
def validate_request(self, request):
|
|
|
|
if request.form_in == "absolute" and request.scheme != "http":
|
|
|
|
raise HttpException("Invalid request scheme: %s" % request.scheme)
|
|
|
|
|
|
|
|
expected_request_forms = {
|
2015-09-16 16:45:22 +00:00
|
|
|
"regular": ("authority", "absolute",),
|
2015-08-30 13:27:29 +00:00
|
|
|
"upstream": ("authority", "absolute"),
|
|
|
|
"transparent": ("relative",)
|
|
|
|
}
|
|
|
|
|
|
|
|
allowed_request_forms = expected_request_forms[self.mode]
|
|
|
|
if request.form_in not in allowed_request_forms:
|
|
|
|
err_message = "Invalid HTTP request form (expected: %s, got: %s)" % (
|
|
|
|
" or ".join(allowed_request_forms), request.form_in
|
|
|
|
)
|
|
|
|
raise HttpException(err_message)
|
|
|
|
|
2015-09-16 16:45:22 +00:00
|
|
|
if self.mode == "regular" and request.form_in == "absolute":
|
2015-08-30 13:27:29 +00:00
|
|
|
request.form_out = "relative"
|
|
|
|
|
|
|
|
def authenticate(self, request):
|
|
|
|
if self.config.authenticator:
|
|
|
|
if self.config.authenticator.authenticate(request.headers):
|
|
|
|
self.config.authenticator.clean(request.headers)
|
|
|
|
else:
|
|
|
|
self.send_response(make_error_response(
|
|
|
|
407,
|
|
|
|
"Proxy Authentication Required",
|
2015-09-05 18:45:58 +00:00
|
|
|
Headers(**self.config.authenticator.auth_challenge_headers())
|
2015-08-30 13:27:29 +00:00
|
|
|
))
|
2015-09-16 16:45:22 +00:00
|
|
|
return False
|
|
|
|
return True
|