mirror of
https://github.com/Grasscutters/mitmproxy.git
synced 2024-11-23 00:01:36 +00:00
http2: respect MAX_CONCURRENT_STREAMS by hold-off
This commit is contained in:
parent
e880f532ad
commit
7f4ac6f27b
@ -73,7 +73,7 @@ class SafeH2Connection(connection.H2Connection):
|
||||
frame_chunk = chunk[position:position + max_outbound_frame_size]
|
||||
if self.local_flow_control_window(stream_id) < len(frame_chunk):
|
||||
self.lock.release()
|
||||
time.sleep(0)
|
||||
time.sleep(0.1)
|
||||
continue
|
||||
self.send_data(stream_id, frame_chunk)
|
||||
self.conn.send(self.data_to_send())
|
||||
@ -352,8 +352,22 @@ class Http2SingleStreamLayer(http._HttpTransmissionLayer, threading.Thread):
|
||||
raise NotImplementedError()
|
||||
|
||||
def send_request(self, message):
|
||||
if not hasattr(self.server_conn, 'h2'):
|
||||
raise exceptions.Http2ProtocolException("Zombie Stream")
|
||||
|
||||
while True:
|
||||
self.server_conn.h2.lock.acquire()
|
||||
max_streams = self.server_conn.h2.remote_settings.max_concurrent_streams
|
||||
if self.server_conn.h2.open_outbound_streams + 1 >= max_streams:
|
||||
# wait until we get a free slot for a new outgoing stream
|
||||
self.server_conn.h2.lock.release()
|
||||
time.sleep(0.1)
|
||||
else:
|
||||
break
|
||||
|
||||
if self.pushed:
|
||||
# nothing to do here
|
||||
self.server_conn.h2.lock.release()
|
||||
return
|
||||
|
||||
with self.server_conn.h2.lock:
|
||||
@ -364,16 +378,19 @@ class Http2SingleStreamLayer(http._HttpTransmissionLayer, threading.Thread):
|
||||
self.server_stream_id = self.server_conn.h2.get_next_available_stream_id()
|
||||
self.server_to_client_stream_ids[self.server_stream_id] = self.client_stream_id
|
||||
|
||||
headers = message.headers.copy()
|
||||
headers.insert(0, ":path", message.path)
|
||||
headers.insert(0, ":method", message.method)
|
||||
headers.insert(0, ":scheme", message.scheme)
|
||||
headers = message.headers.copy()
|
||||
headers.insert(0, ":path", message.path)
|
||||
headers.insert(0, ":method", message.method)
|
||||
headers.insert(0, ":scheme", message.scheme)
|
||||
self.server_stream_id = self.server_conn.h2.get_next_available_stream_id()
|
||||
self.server_to_client_stream_ids[self.server_stream_id] = self.client_stream_id
|
||||
self.server_conn.h2.safe_send_headers(
|
||||
self.is_zombie,
|
||||
self.server_stream_id,
|
||||
headers,
|
||||
)
|
||||
self.server_conn.h2.lock.release()
|
||||
|
||||
self.server_conn.h2.safe_send_headers(
|
||||
self.is_zombie,
|
||||
self.server_stream_id,
|
||||
headers
|
||||
)
|
||||
self.server_conn.h2.safe_send_body(
|
||||
self.is_zombie,
|
||||
self.server_stream_id,
|
||||
@ -408,7 +425,7 @@ class Http2SingleStreamLayer(http._HttpTransmissionLayer, threading.Thread):
|
||||
if self.response_data_finished.is_set():
|
||||
while self.response_data_queue.qsize() > 0:
|
||||
yield self.response_data_queue.get()
|
||||
return
|
||||
break
|
||||
if self.zombie: # pragma: no cover
|
||||
raise exceptions.Http2ProtocolException("Zombie Stream")
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user