mirror of
https://github.com/Grasscutters/mitmproxy.git
synced 2024-11-22 07:08:10 +00:00
pyupgrade --keep-runtime-typing --py38-plus
This commit is contained in:
parent
beb49ab121
commit
9d1e3107e8
@ -21,7 +21,7 @@ class MyAddon:
|
|||||||
totals[f.request.host] = totals.setdefault(f.request.host, 0) + 1
|
totals[f.request.host] = totals.setdefault(f.request.host, 0) + 1
|
||||||
|
|
||||||
with open(path, "w+") as fp:
|
with open(path, "w+") as fp:
|
||||||
for cnt, dom in sorted([(v, k) for (k, v) in totals.items()]):
|
for cnt, dom in sorted((v, k) for (k, v) in totals.items()):
|
||||||
fp.write(f"{cnt}: {dom}\n")
|
fp.write(f"{cnt}: {dom}\n")
|
||||||
|
|
||||||
ctx.log.alert("done")
|
ctx.log.alert("done")
|
||||||
|
@ -15,7 +15,7 @@ def websocket_message(flow: http.HTTPFlow):
|
|||||||
last_message = flow.websocket.messages[-1]
|
last_message = flow.websocket.messages[-1]
|
||||||
if last_message.is_text and "secret" in last_message.text:
|
if last_message.is_text and "secret" in last_message.text:
|
||||||
last_message.drop()
|
last_message.drop()
|
||||||
ctx.master.commands.call("inject.websocket", flow, last_message.from_client, "ssssssh".encode())
|
ctx.master.commands.call("inject.websocket", flow, last_message.from_client, b"ssssssh")
|
||||||
|
|
||||||
|
|
||||||
# Complex example: Schedule a periodic timer
|
# Complex example: Schedule a periodic timer
|
||||||
|
@ -3,14 +3,15 @@ import binascii
|
|||||||
import socket
|
import socket
|
||||||
import typing
|
import typing
|
||||||
|
|
||||||
|
from ntlm_auth import gss_channel_bindings, ntlm
|
||||||
|
|
||||||
|
from mitmproxy import addonmanager, http
|
||||||
from mitmproxy import ctx
|
from mitmproxy import ctx
|
||||||
from mitmproxy import http, addonmanager
|
|
||||||
from mitmproxy.net.http import http1
|
from mitmproxy.net.http import http1
|
||||||
from mitmproxy.proxy import layer, commands
|
from mitmproxy.proxy import commands, layer
|
||||||
from mitmproxy.proxy.context import Context
|
from mitmproxy.proxy.context import Context
|
||||||
from mitmproxy.proxy.layers.http import HttpConnectUpstreamHook, HttpLayer, HttpStream
|
from mitmproxy.proxy.layers.http import HttpConnectUpstreamHook, HttpLayer, HttpStream
|
||||||
from mitmproxy.proxy.layers.http._upstream_proxy import HttpUpstreamProxy
|
from mitmproxy.proxy.layers.http._upstream_proxy import HttpUpstreamProxy
|
||||||
from ntlm_auth import gss_channel_bindings, ntlm
|
|
||||||
|
|
||||||
|
|
||||||
class NTLMUpstreamAuth:
|
class NTLMUpstreamAuth:
|
||||||
@ -148,7 +149,7 @@ class CustomNTLMContext:
|
|||||||
ntlm_compatibility: int = ctx.options.upstream_ntlm_compatibility
|
ntlm_compatibility: int = ctx.options.upstream_ntlm_compatibility
|
||||||
username, password = tuple(auth.split(":"))
|
username, password = tuple(auth.split(":"))
|
||||||
workstation = socket.gethostname().upper()
|
workstation = socket.gethostname().upper()
|
||||||
ctx.log.debug('\nntlm context with the details: "{}\\{}", *****'.format(domain, username))
|
ctx.log.debug(f'\nntlm context with the details: "{domain}\\{username}", *****')
|
||||||
self.ctx_log = ctx.log
|
self.ctx_log = ctx.log
|
||||||
self.preferred_type = preferred_type
|
self.preferred_type = preferred_type
|
||||||
self.ntlm_context = ntlm.NtlmContext(
|
self.ntlm_context = ntlm.NtlmContext(
|
||||||
@ -163,7 +164,7 @@ class CustomNTLMContext:
|
|||||||
negotiate_message = self.ntlm_context.step()
|
negotiate_message = self.ntlm_context.step()
|
||||||
negotiate_message_base_64_in_bytes = base64.b64encode(negotiate_message)
|
negotiate_message_base_64_in_bytes = base64.b64encode(negotiate_message)
|
||||||
negotiate_message_base_64_ascii = negotiate_message_base_64_in_bytes.decode("ascii")
|
negotiate_message_base_64_ascii = negotiate_message_base_64_in_bytes.decode("ascii")
|
||||||
negotiate_message_base_64_final = u'%s %s' % (self.preferred_type, negotiate_message_base_64_ascii)
|
negotiate_message_base_64_final = f'{self.preferred_type} {negotiate_message_base_64_ascii}'
|
||||||
self.ctx_log.debug(
|
self.ctx_log.debug(
|
||||||
f'{self.preferred_type} Authentication, negotiate message: {negotiate_message_base_64_final}'
|
f'{self.preferred_type} Authentication, negotiate message: {negotiate_message_base_64_final}'
|
||||||
)
|
)
|
||||||
@ -174,10 +175,10 @@ class CustomNTLMContext:
|
|||||||
try:
|
try:
|
||||||
challenge_message_ascii_bytes = base64.b64decode(challenge_message, validate=True)
|
challenge_message_ascii_bytes = base64.b64decode(challenge_message, validate=True)
|
||||||
except binascii.Error as err:
|
except binascii.Error as err:
|
||||||
self.ctx_log.debug('{} Authentication fail with error {}'.format(self.preferred_type, err.__str__()))
|
self.ctx_log.debug(f'{self.preferred_type} Authentication fail with error {err.__str__()}')
|
||||||
return False
|
return False
|
||||||
authenticate_message = self.ntlm_context.step(challenge_message_ascii_bytes)
|
authenticate_message = self.ntlm_context.step(challenge_message_ascii_bytes)
|
||||||
negotiate_message_base_64 = u'%s %s' % (self.preferred_type,
|
negotiate_message_base_64 = '{} {}'.format(self.preferred_type,
|
||||||
base64.b64encode(authenticate_message).decode('ascii'))
|
base64.b64encode(authenticate_message).decode('ascii'))
|
||||||
self.ctx_log.debug(
|
self.ctx_log.debug(
|
||||||
f'{self.preferred_type} Authentication, response to challenge message: {negotiate_message_base_64}'
|
f'{self.preferred_type} Authentication, response to challenge message: {negotiate_message_base_64}'
|
||||||
|
@ -178,7 +178,7 @@ class Core:
|
|||||||
req.url = val
|
req.url = val
|
||||||
except ValueError as e:
|
except ValueError as e:
|
||||||
raise exceptions.CommandError(
|
raise exceptions.CommandError(
|
||||||
"URL {} is invalid: {}".format(repr(val), e)
|
f"URL {repr(val)} is invalid: {e}"
|
||||||
) from e
|
) from e
|
||||||
else:
|
else:
|
||||||
self.rupdate = False
|
self.rupdate = False
|
||||||
@ -199,7 +199,7 @@ class Core:
|
|||||||
updated.append(f)
|
updated.append(f)
|
||||||
|
|
||||||
ctx.master.addons.trigger(hooks.UpdateHook(updated))
|
ctx.master.addons.trigger(hooks.UpdateHook(updated))
|
||||||
ctx.log.alert("Set {} on {} flows.".format(attr, len(updated)))
|
ctx.log.alert(f"Set {attr} on {len(updated)} flows.")
|
||||||
|
|
||||||
@command.command("flow.decode")
|
@command.command("flow.decode")
|
||||||
def decode(self, flows: typing.Sequence[flow.Flow], part: str) -> None:
|
def decode(self, flows: typing.Sequence[flow.Flow], part: str) -> None:
|
||||||
|
@ -63,7 +63,7 @@ def curl_command(f: flow.Flow) -> str:
|
|||||||
server_addr = f.server_conn.peername[0] if f.server_conn.peername else None
|
server_addr = f.server_conn.peername[0] if f.server_conn.peername else None
|
||||||
|
|
||||||
if ctx.options.export_preserve_original_ip and server_addr and request.pretty_host != server_addr:
|
if ctx.options.export_preserve_original_ip and server_addr and request.pretty_host != server_addr:
|
||||||
resolve = "{}:{}:[{}]".format(request.pretty_host, request.port, server_addr)
|
resolve = f"{request.pretty_host}:{request.port}:[{server_addr}]"
|
||||||
args.append("--resolve")
|
args.append("--resolve")
|
||||||
args.append(resolve)
|
args.append(resolve)
|
||||||
|
|
||||||
|
@ -112,7 +112,7 @@ class ServerPlayback:
|
|||||||
|
|
||||||
@command.command("replay.server.count")
|
@command.command("replay.server.count")
|
||||||
def count(self) -> int:
|
def count(self) -> int:
|
||||||
return sum([len(i) for i in self.flowmap.values()])
|
return sum(len(i) for i in self.flowmap.values())
|
||||||
|
|
||||||
def _hash(self, flow: http.HTTPFlow) -> typing.Hashable:
|
def _hash(self, flow: http.HTTPFlow) -> typing.Hashable:
|
||||||
"""
|
"""
|
||||||
|
@ -265,11 +265,11 @@ class CommandManager:
|
|||||||
parts, _ = self.parse_partial(cmdstr)
|
parts, _ = self.parse_partial(cmdstr)
|
||||||
if not parts:
|
if not parts:
|
||||||
raise exceptions.CommandError(f"Invalid command: {cmdstr!r}")
|
raise exceptions.CommandError(f"Invalid command: {cmdstr!r}")
|
||||||
command_name, *args = [
|
command_name, *args = (
|
||||||
unquote(part.value)
|
unquote(part.value)
|
||||||
for part in parts
|
for part in parts
|
||||||
if part.type != mitmproxy.types.Space
|
if part.type != mitmproxy.types.Space
|
||||||
]
|
)
|
||||||
return self.call_strings(command_name, args)
|
return self.call_strings(command_name, args)
|
||||||
|
|
||||||
def dump(self, out=sys.stdout) -> None:
|
def dump(self, out=sys.stdout) -> None:
|
||||||
|
@ -67,4 +67,4 @@ if __name__ == "__main__": # pragma: no cover
|
|||||||
|
|
||||||
t = time.time()
|
t = time.time()
|
||||||
x = beautify(data)
|
x = beautify(data)
|
||||||
print("Beautifying vendor.css took {:.2}s".format(time.time() - t))
|
print(f"Beautifying vendor.css took {time.time() - t:.2}s")
|
||||||
|
@ -20,7 +20,7 @@ def format_query_list(data: typing.List[typing.Any]):
|
|||||||
num_queries = len(data) - 1
|
num_queries = len(data) - 1
|
||||||
result = ""
|
result = ""
|
||||||
for i, op in enumerate(data):
|
for i, op in enumerate(data):
|
||||||
result += "--- {i}/{num_queries}\n".format(i=i, num_queries=num_queries)
|
result += f"--- {i}/{num_queries}\n"
|
||||||
result += format_graphql(op)
|
result += format_graphql(op)
|
||||||
return result
|
return result
|
||||||
|
|
||||||
|
@ -200,7 +200,7 @@ class ProtoParser:
|
|||||||
# read field key (tag and wire_type)
|
# read field key (tag and wire_type)
|
||||||
offset, key = ProtoParser._read_base128le(wire_data[pos:])
|
offset, key = ProtoParser._read_base128le(wire_data[pos:])
|
||||||
# casting raises exception for invalid WireTypes
|
# casting raises exception for invalid WireTypes
|
||||||
wt = ProtoParser.WireTypes((key & 7))
|
wt = ProtoParser.WireTypes(key & 7)
|
||||||
tag = (key >> 3)
|
tag = (key >> 3)
|
||||||
pos += offset
|
pos += offset
|
||||||
|
|
||||||
@ -232,7 +232,7 @@ class ProtoParser:
|
|||||||
wt == ProtoParser.WireTypes.group_start or
|
wt == ProtoParser.WireTypes.group_start or
|
||||||
wt == ProtoParser.WireTypes.group_end
|
wt == ProtoParser.WireTypes.group_end
|
||||||
):
|
):
|
||||||
raise ValueError("deprecated field: {}".format(wt))
|
raise ValueError(f"deprecated field: {wt}")
|
||||||
elif wt == ProtoParser.WireTypes.bit_32:
|
elif wt == ProtoParser.WireTypes.bit_32:
|
||||||
offset, val = ProtoParser._read_u32(wire_data[pos:])
|
offset, val = ProtoParser._read_u32(wire_data[pos:])
|
||||||
pos += offset
|
pos += offset
|
||||||
@ -735,8 +735,7 @@ class ProtoParser:
|
|||||||
yield field_desc_dict
|
yield field_desc_dict
|
||||||
# add sub-fields of messages or packed fields
|
# add sub-fields of messages or packed fields
|
||||||
for f in decoded_val:
|
for f in decoded_val:
|
||||||
for field_dict in f.gen_flat_decoded_field_dicts():
|
yield from f.gen_flat_decoded_field_dicts()
|
||||||
yield field_dict
|
|
||||||
else:
|
else:
|
||||||
field_desc_dict["val"] = decoded_val
|
field_desc_dict["val"] = decoded_val
|
||||||
yield field_desc_dict
|
yield field_desc_dict
|
||||||
@ -767,8 +766,7 @@ class ProtoParser:
|
|||||||
|
|
||||||
def gen_flat_decoded_field_dicts(self) -> Generator[Dict, None, None]:
|
def gen_flat_decoded_field_dicts(self) -> Generator[Dict, None, None]:
|
||||||
for f in self.root_fields:
|
for f in self.root_fields:
|
||||||
for field_dict in f.gen_flat_decoded_field_dicts():
|
yield from f.gen_flat_decoded_field_dicts()
|
||||||
yield field_dict
|
|
||||||
|
|
||||||
def gen_str_rows(self) -> Generator[Tuple[str, ...], None, None]:
|
def gen_str_rows(self) -> Generator[Tuple[str, ...], None, None]:
|
||||||
for field_dict in self.gen_flat_decoded_field_dicts():
|
for field_dict in self.gen_flat_decoded_field_dicts():
|
||||||
@ -879,8 +877,7 @@ def hack_generator_to_list(generator_func):
|
|||||||
|
|
||||||
|
|
||||||
def format_pbuf(message: bytes, parser_options: ProtoParser.ParserOptions, rules: List[ProtoParser.ParserRule]):
|
def format_pbuf(message: bytes, parser_options: ProtoParser.ParserOptions, rules: List[ProtoParser.ParserRule]):
|
||||||
for l in format_table(ProtoParser(data=message, parser_options=parser_options, rules=rules).gen_str_rows()):
|
yield from format_table(ProtoParser(data=message, parser_options=parser_options, rules=rules).gen_str_rows())
|
||||||
yield l
|
|
||||||
|
|
||||||
|
|
||||||
def format_grpc(
|
def format_grpc(
|
||||||
@ -895,12 +892,11 @@ def format_grpc(
|
|||||||
compression_scheme if compressed else compressed) + ')'
|
compression_scheme if compressed else compressed) + ')'
|
||||||
|
|
||||||
yield [("text", headline)]
|
yield [("text", headline)]
|
||||||
for l in format_pbuf(
|
yield from format_pbuf(
|
||||||
message=pb_message,
|
message=pb_message,
|
||||||
parser_options=parser_options,
|
parser_options=parser_options,
|
||||||
rules=rules
|
rules=rules
|
||||||
):
|
)
|
||||||
yield l
|
|
||||||
|
|
||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
|
@ -3,10 +3,10 @@ import typing
|
|||||||
|
|
||||||
from kaitaistruct import KaitaiStream
|
from kaitaistruct import KaitaiStream
|
||||||
|
|
||||||
from mitmproxy.contrib.kaitaistruct import png
|
|
||||||
from mitmproxy.contrib.kaitaistruct import gif
|
from mitmproxy.contrib.kaitaistruct import gif
|
||||||
from mitmproxy.contrib.kaitaistruct import jpeg
|
|
||||||
from mitmproxy.contrib.kaitaistruct import ico
|
from mitmproxy.contrib.kaitaistruct import ico
|
||||||
|
from mitmproxy.contrib.kaitaistruct import jpeg
|
||||||
|
from mitmproxy.contrib.kaitaistruct import png
|
||||||
|
|
||||||
Metadata = typing.List[typing.Tuple[str, str]]
|
Metadata = typing.List[typing.Tuple[str, str]]
|
||||||
|
|
||||||
@ -91,7 +91,8 @@ def parse_ico(data: bytes) -> Metadata:
|
|||||||
for i, image in enumerate(img.images):
|
for i, image in enumerate(img.images):
|
||||||
parts.append(
|
parts.append(
|
||||||
(
|
(
|
||||||
'Image {}'.format(i + 1), "Size: {} x {}\n"
|
f'Image {i + 1}',
|
||||||
|
"Size: {} x {}\n"
|
||||||
"{: >18}Bits per pixel: {}\n"
|
"{: >18}Bits per pixel: {}\n"
|
||||||
"{: >18}PNG: {}".format(256 if not image.width else image.width,
|
"{: >18}PNG: {}".format(256 if not image.width else image.width,
|
||||||
256 if not image.height else image.height,
|
256 if not image.height else image.height,
|
||||||
|
@ -376,7 +376,7 @@ class FSrc(_Rex):
|
|||||||
def __call__(self, f):
|
def __call__(self, f):
|
||||||
if not f.client_conn or not f.client_conn.peername:
|
if not f.client_conn or not f.client_conn.peername:
|
||||||
return False
|
return False
|
||||||
r = "{}:{}".format(f.client_conn.peername[0], f.client_conn.peername[1])
|
r = f"{f.client_conn.peername[0]}:{f.client_conn.peername[1]}"
|
||||||
return f.client_conn.peername and self.re.search(r)
|
return f.client_conn.peername and self.re.search(r)
|
||||||
|
|
||||||
|
|
||||||
@ -388,7 +388,7 @@ class FDst(_Rex):
|
|||||||
def __call__(self, f):
|
def __call__(self, f):
|
||||||
if not f.server_conn or not f.server_conn.address:
|
if not f.server_conn or not f.server_conn.address:
|
||||||
return False
|
return False
|
||||||
r = "{}:{}".format(f.server_conn.address[0], f.server_conn.address[1])
|
r = f"{f.server_conn.address[0]}:{f.server_conn.address[1]}"
|
||||||
return f.server_conn.address and self.re.search(r)
|
return f.server_conn.address and self.re.search(r)
|
||||||
|
|
||||||
|
|
||||||
|
@ -510,7 +510,7 @@ class Message(serializable.Serializable):
|
|||||||
self.headers["content-encoding"] = encoding
|
self.headers["content-encoding"] = encoding
|
||||||
self.content = self.raw_content
|
self.content = self.raw_content
|
||||||
if "content-encoding" not in self.headers:
|
if "content-encoding" not in self.headers:
|
||||||
raise ValueError("Invalid content encoding {}".format(repr(encoding)))
|
raise ValueError(f"Invalid content encoding {repr(encoding)}")
|
||||||
|
|
||||||
def json(self, **kwargs: Any) -> Any:
|
def json(self, **kwargs: Any) -> Any:
|
||||||
"""
|
"""
|
||||||
@ -1033,7 +1033,7 @@ class Response(Message):
|
|||||||
reason = reason.encode("ascii", "strict")
|
reason = reason.encode("ascii", "strict")
|
||||||
|
|
||||||
if isinstance(content, str):
|
if isinstance(content, str):
|
||||||
raise ValueError("Content must be bytes, not {}".format(type(content).__name__))
|
raise ValueError(f"Content must be bytes, not {type(content).__name__}")
|
||||||
if not isinstance(headers, Headers):
|
if not isinstance(headers, Headers):
|
||||||
headers = Headers(headers)
|
headers = Headers(headers)
|
||||||
if trailers is not None and not isinstance(trailers, Headers):
|
if trailers is not None and not isinstance(trailers, Headers):
|
||||||
|
@ -146,7 +146,7 @@ def _rdumpq(q: collections.deque, size: int, value: TSerializable) -> int:
|
|||||||
write(span)
|
write(span)
|
||||||
return size + 1 + len(span)
|
return size + 1 + len(span)
|
||||||
else:
|
else:
|
||||||
raise ValueError("unserializable object: {} ({})".format(value, type(value)))
|
raise ValueError(f"unserializable object: {value} ({type(value)})")
|
||||||
|
|
||||||
|
|
||||||
def loads(string: bytes) -> TSerializable:
|
def loads(string: bytes) -> TSerializable:
|
||||||
|
@ -63,7 +63,7 @@ class WebsocketConnection(wsproto.Connection):
|
|||||||
frame_buf: List[bytes]
|
frame_buf: List[bytes]
|
||||||
|
|
||||||
def __init__(self, *args, conn: connection.Connection, **kwargs):
|
def __init__(self, *args, conn: connection.Connection, **kwargs):
|
||||||
super(WebsocketConnection, self).__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
self.conn = conn
|
self.conn = conn
|
||||||
self.frame_buf = [b""]
|
self.frame_buf = [b""]
|
||||||
|
|
||||||
|
@ -153,7 +153,7 @@ class ConnectionHandler(metaclass=abc.ABCMeta):
|
|||||||
try:
|
try:
|
||||||
command.connection.timestamp_start = time.time()
|
command.connection.timestamp_start = time.time()
|
||||||
reader, writer = await asyncio.open_connection(*command.connection.address)
|
reader, writer = await asyncio.open_connection(*command.connection.address)
|
||||||
except (IOError, asyncio.CancelledError) as e:
|
except (OSError, asyncio.CancelledError) as e:
|
||||||
err = str(e)
|
err = str(e)
|
||||||
if not err: # str(CancelledError()) returns empty string.
|
if not err: # str(CancelledError()) returns empty string.
|
||||||
err = "connection cancelled"
|
err = "connection cancelled"
|
||||||
|
@ -55,7 +55,7 @@ class TCPFlow(flow.Flow):
|
|||||||
_stateobject_attributes["messages"] = List[TCPMessage]
|
_stateobject_attributes["messages"] = List[TCPMessage]
|
||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return "<TCPFlow ({} messages)>".format(len(self.messages))
|
return f"<TCPFlow ({len(self.messages)} messages)>"
|
||||||
|
|
||||||
|
|
||||||
__all__ = [
|
__all__ = [
|
||||||
|
@ -46,7 +46,7 @@ class EventLog(urwid.ListBox, layoutwidget.LayoutWidget):
|
|||||||
def add_event(self, event_store, entry: log.LogEntry):
|
def add_event(self, event_store, entry: log.LogEntry):
|
||||||
if log.log_tier(self.master.options.console_eventlog_verbosity) < log.log_tier(entry.level):
|
if log.log_tier(self.master.options.console_eventlog_verbosity) < log.log_tier(entry.level):
|
||||||
return
|
return
|
||||||
txt = "{}: {}".format(entry.level, str(entry.msg))
|
txt = f"{entry.level}: {str(entry.msg)}"
|
||||||
if entry.level in ("error", "warn", "alert"):
|
if entry.level in ("error", "warn", "alert"):
|
||||||
e = urwid.Text((entry.level, txt))
|
e = urwid.Text((entry.level, txt))
|
||||||
else:
|
else:
|
||||||
|
@ -89,7 +89,7 @@ class ConsoleMaster(master.Master):
|
|||||||
signals.status_message.send(
|
signals.status_message.send(
|
||||||
message = (
|
message = (
|
||||||
entry.level,
|
entry.level,
|
||||||
"{}: {}".format(entry.level.title(), str(entry.msg).lstrip())
|
f"{entry.level.title()}: {str(entry.msg).lstrip()}"
|
||||||
),
|
),
|
||||||
expire=5
|
expire=5
|
||||||
)
|
)
|
||||||
|
@ -108,7 +108,7 @@ class Chooser(urwid.WidgetWrap, layoutwidget.LayoutWidget):
|
|||||||
self.master = master
|
self.master = master
|
||||||
self.choices = choices
|
self.choices = choices
|
||||||
self.callback = callback
|
self.callback = callback
|
||||||
choicewidth = max([len(i) for i in choices])
|
choicewidth = max(len(i) for i in choices)
|
||||||
self.width = max(choicewidth, len(title)) + 7
|
self.width = max(choicewidth, len(title)) + 7
|
||||||
|
|
||||||
self.walker = ChooserListWalker(choices, current)
|
self.walker = ChooserListWalker(choices, current)
|
||||||
|
@ -100,7 +100,7 @@ def run(
|
|||||||
opts.update(**extra(args))
|
opts.update(**extra(args))
|
||||||
|
|
||||||
except exceptions.OptionsError as e:
|
except exceptions.OptionsError as e:
|
||||||
print("{}: {}".format(sys.argv[0], e), file=sys.stderr)
|
print(f"{sys.argv[0]}: {e}", file=sys.stderr)
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
loop = asyncio.get_running_loop()
|
loop = asyncio.get_running_loop()
|
||||||
|
@ -206,7 +206,7 @@ class RequestHandler(tornado.web.RequestHandler):
|
|||||||
try:
|
try:
|
||||||
return json.loads(self.request.body.decode())
|
return json.loads(self.request.body.decode())
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise APIError(400, "Malformed JSON: {}".format(str(e)))
|
raise APIError(400, f"Malformed JSON: {str(e)}")
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def filecontents(self):
|
def filecontents(self):
|
||||||
|
@ -20,7 +20,7 @@ def dump_system_info():
|
|||||||
data = [
|
data = [
|
||||||
f"Mitmproxy: {mitmproxy_version}",
|
f"Mitmproxy: {mitmproxy_version}",
|
||||||
f"Python: {platform.python_version()}",
|
f"Python: {platform.python_version()}",
|
||||||
"OpenSSL: {}".format(SSL.SSLeay_version(SSL.SSLEAY_VERSION).decode()),
|
f"OpenSSL: {SSL.SSLeay_version(SSL.SSLEAY_VERSION).decode()}",
|
||||||
f"Platform: {platform.platform()}",
|
f"Platform: {platform.platform()}",
|
||||||
]
|
]
|
||||||
return "\n".join(data)
|
return "\n".join(data)
|
||||||
|
@ -31,7 +31,7 @@ def pretty_size(size: int) -> str:
|
|||||||
raise AssertionError
|
raise AssertionError
|
||||||
|
|
||||||
|
|
||||||
@functools.lru_cache()
|
@functools.lru_cache
|
||||||
def parse_size(s: typing.Optional[str]) -> typing.Optional[int]:
|
def parse_size(s: typing.Optional[str]) -> typing.Optional[int]:
|
||||||
"""
|
"""
|
||||||
Parse a size with an optional k/m/... suffix.
|
Parse a size with an optional k/m/... suffix.
|
||||||
@ -65,7 +65,7 @@ def pretty_duration(secs: typing.Optional[float]) -> str:
|
|||||||
if secs >= limit:
|
if secs >= limit:
|
||||||
return formatter.format(secs)
|
return formatter.format(secs)
|
||||||
# less than 1 sec
|
# less than 1 sec
|
||||||
return "{:.0f}ms".format(secs * 1000)
|
return f"{secs * 1000:.0f}ms"
|
||||||
|
|
||||||
|
|
||||||
def format_timestamp(s):
|
def format_timestamp(s):
|
||||||
@ -79,7 +79,7 @@ def format_timestamp_with_milli(s):
|
|||||||
return d.strftime("%Y-%m-%d %H:%M:%S.%f")[:-3]
|
return d.strftime("%Y-%m-%d %H:%M:%S.%f")[:-3]
|
||||||
|
|
||||||
|
|
||||||
@functools.lru_cache()
|
@functools.lru_cache
|
||||||
def format_address(address: typing.Optional[tuple]) -> str:
|
def format_address(address: typing.Optional[tuple]) -> str:
|
||||||
"""
|
"""
|
||||||
This function accepts IPv4/IPv6 tuples and
|
This function accepts IPv4/IPv6 tuples and
|
||||||
@ -90,12 +90,12 @@ def format_address(address: typing.Optional[tuple]) -> str:
|
|||||||
try:
|
try:
|
||||||
host = ipaddress.ip_address(address[0])
|
host = ipaddress.ip_address(address[0])
|
||||||
if host.is_unspecified:
|
if host.is_unspecified:
|
||||||
return "*:{}".format(address[1])
|
return f"*:{address[1]}"
|
||||||
if isinstance(host, ipaddress.IPv4Address):
|
if isinstance(host, ipaddress.IPv4Address):
|
||||||
return "{}:{}".format(str(host), address[1])
|
return f"{str(host)}:{address[1]}"
|
||||||
# If IPv6 is mapped to IPv4
|
# If IPv6 is mapped to IPv4
|
||||||
elif host.ipv4_mapped:
|
elif host.ipv4_mapped:
|
||||||
return "{}:{}".format(str(host.ipv4_mapped), address[1])
|
return f"{str(host.ipv4_mapped)}:{address[1]}"
|
||||||
return "[{}]:{}".format(str(host), address[1])
|
return f"[{str(host)}]:{address[1]}"
|
||||||
except ValueError:
|
except ValueError:
|
||||||
return "{}:{}".format(address[0], address[1])
|
return f"{address[0]}:{address[1]}"
|
||||||
|
@ -22,7 +22,7 @@ def always_bytes(str_or_bytes: Union[None, str, bytes], *encode_args) -> Union[N
|
|||||||
elif isinstance(str_or_bytes, str):
|
elif isinstance(str_or_bytes, str):
|
||||||
return str_or_bytes.encode(*encode_args)
|
return str_or_bytes.encode(*encode_args)
|
||||||
else:
|
else:
|
||||||
raise TypeError("Expected str or bytes, but got {}.".format(type(str_or_bytes).__name__))
|
raise TypeError(f"Expected str or bytes, but got {type(str_or_bytes).__name__}.")
|
||||||
|
|
||||||
|
|
||||||
@overload
|
@overload
|
||||||
@ -45,7 +45,7 @@ def always_str(str_or_bytes: Union[None, str, bytes], *decode_args) -> Union[Non
|
|||||||
elif isinstance(str_or_bytes, bytes):
|
elif isinstance(str_or_bytes, bytes):
|
||||||
return str_or_bytes.decode(*decode_args)
|
return str_or_bytes.decode(*decode_args)
|
||||||
else:
|
else:
|
||||||
raise TypeError("Expected str or bytes, but got {}.".format(type(str_or_bytes).__name__))
|
raise TypeError(f"Expected str or bytes, but got {type(str_or_bytes).__name__}.")
|
||||||
|
|
||||||
|
|
||||||
# Translate control characters to "safe" characters. This implementation
|
# Translate control characters to "safe" characters. This implementation
|
||||||
@ -73,7 +73,7 @@ def escape_control_characters(text: str, keep_spacing=True) -> str:
|
|||||||
keep_spacing: If True, tabs and newlines will not be replaced.
|
keep_spacing: If True, tabs and newlines will not be replaced.
|
||||||
"""
|
"""
|
||||||
if not isinstance(text, str):
|
if not isinstance(text, str):
|
||||||
raise ValueError("text type must be unicode but is {}".format(type(text).__name__))
|
raise ValueError(f"text type must be unicode but is {type(text).__name__}")
|
||||||
|
|
||||||
trans = _control_char_trans_newline if keep_spacing else _control_char_trans
|
trans = _control_char_trans_newline if keep_spacing else _control_char_trans
|
||||||
return text.translate(trans)
|
return text.translate(trans)
|
||||||
|
@ -388,7 +388,7 @@ def build_pyinstaller(be: BuildEnviron) -> None: # pragma: no cover
|
|||||||
click.echo(subprocess.check_output([executable, "--version"]).decode())
|
click.echo(subprocess.check_output([executable, "--version"]).decode())
|
||||||
|
|
||||||
archive.add(str(executable), str(executable.name))
|
archive.add(str(executable), str(executable.name))
|
||||||
click.echo("Packed {}.".format(be.archive_path.name))
|
click.echo(f"Packed {be.archive_path.name}.")
|
||||||
|
|
||||||
|
|
||||||
def build_wininstaller(be: BuildEnviron) -> None: # pragma: no cover
|
def build_wininstaller(be: BuildEnviron) -> None: # pragma: no cover
|
||||||
|
@ -108,7 +108,7 @@ def pytest_terminal_summary(terminalreporter, exitstatus, config):
|
|||||||
terminalreporter.write(msg, **markup)
|
terminalreporter.write(msg, **markup)
|
||||||
|
|
||||||
for name in sorted(coverage_values.keys()):
|
for name in sorted(coverage_values.keys()):
|
||||||
msg = 'Coverage for {}: {:.2f}%\n'.format(name, coverage_values[name][0])
|
msg = f'Coverage for {name}: {coverage_values[name][0]:.2f}%\n'
|
||||||
if coverage_values[name][0] < 100:
|
if coverage_values[name][0] < 100:
|
||||||
markup = {'red': True, 'bold': True}
|
markup = {'red': True, 'bold': True}
|
||||||
for s, v in sorted(coverage_values[name][1]):
|
for s, v in sorted(coverage_values[name][1]):
|
||||||
|
@ -62,7 +62,7 @@ def get_random_object(random=random, depth=0):
|
|||||||
else:
|
else:
|
||||||
return -1 * random.randint(0, MAXINT)
|
return -1 * random.randint(0, MAXINT)
|
||||||
n = random.randint(0, 100)
|
n = random.randint(0, 100)
|
||||||
return bytes([random.randint(32, 126) for _ in range(n)])
|
return bytes(random.randint(32, 126) for _ in range(n))
|
||||||
|
|
||||||
|
|
||||||
class Test_Format(unittest.TestCase):
|
class Test_Format(unittest.TestCase):
|
||||||
|
@ -22,7 +22,7 @@ SAMPLE_SETTINGS = {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
class FrameFactory(object):
|
class FrameFactory:
|
||||||
"""
|
"""
|
||||||
A class containing lots of helper methods and state to build frames. This
|
A class containing lots of helper methods and state to build frames. This
|
||||||
allows test cases to easily build correct HTTP/2 frames to feed to
|
allows test cases to easily build correct HTTP/2 frames to feed to
|
||||||
|
@ -1312,9 +1312,9 @@ def test_invalid_content_length(tctx):
|
|||||||
flow = Placeholder(HTTPFlow)
|
flow = Placeholder(HTTPFlow)
|
||||||
assert (
|
assert (
|
||||||
Playbook(http.HttpLayer(tctx, HTTPMode.regular))
|
Playbook(http.HttpLayer(tctx, HTTPMode.regular))
|
||||||
>> DataReceived(tctx.client, ("GET http://example.com/ HTTP/1.1\r\n"
|
>> DataReceived(tctx.client, (b"GET http://example.com/ HTTP/1.1\r\n"
|
||||||
"Host: example.com\r\n"
|
b"Host: example.com\r\n"
|
||||||
"Content-Length: NaN\r\n\r\n").encode())
|
b"Content-Length: NaN\r\n\r\n"))
|
||||||
<< SendData(tctx.client, err)
|
<< SendData(tctx.client, err)
|
||||||
<< CloseConnection(tctx.client)
|
<< CloseConnection(tctx.client)
|
||||||
<< http.HttpRequestHeadersHook(flow)
|
<< http.HttpRequestHeadersHook(flow)
|
||||||
|
@ -27,7 +27,7 @@ class _Masked:
|
|||||||
other[1] &= 0b0111_1111 # remove mask bit
|
other[1] &= 0b0111_1111 # remove mask bit
|
||||||
assert other[1] < 126 # (we don't support extended payload length here)
|
assert other[1] < 126 # (we don't support extended payload length here)
|
||||||
mask = other[2:6]
|
mask = other[2:6]
|
||||||
payload = bytes([x ^ mask[i % 4] for i, x in enumerate(other[6:])])
|
payload = bytes(x ^ mask[i % 4] for i, x in enumerate(other[6:]))
|
||||||
return self.unmasked == other[:2] + payload
|
return self.unmasked == other[:2] + payload
|
||||||
|
|
||||||
|
|
||||||
@ -41,7 +41,7 @@ def masked_bytes(unmasked: bytes) -> bytes:
|
|||||||
assert header[1] < 126 # assert that this is neither masked nor extended payload
|
assert header[1] < 126 # assert that this is neither masked nor extended payload
|
||||||
header[1] |= 0b1000_0000
|
header[1] |= 0b1000_0000
|
||||||
mask = secrets.token_bytes(4)
|
mask = secrets.token_bytes(4)
|
||||||
masked = bytes([x ^ mask[i % 4] for i, x in enumerate(unmasked[2:])])
|
masked = bytes(x ^ mask[i % 4] for i, x in enumerate(unmasked[2:]))
|
||||||
return bytes(header + mask + masked)
|
return bytes(header + mask + masked)
|
||||||
|
|
||||||
|
|
||||||
|
@ -264,13 +264,13 @@ class TestCert:
|
|||||||
def test_multi_valued_rdns(self, tdata):
|
def test_multi_valued_rdns(self, tdata):
|
||||||
subject = x509.Name([
|
subject = x509.Name([
|
||||||
x509.RelativeDistinguishedName([
|
x509.RelativeDistinguishedName([
|
||||||
x509.NameAttribute(NameOID.TITLE, u'Test'),
|
x509.NameAttribute(NameOID.TITLE, 'Test'),
|
||||||
x509.NameAttribute(NameOID.COMMON_NAME, u'Multivalue'),
|
x509.NameAttribute(NameOID.COMMON_NAME, 'Multivalue'),
|
||||||
x509.NameAttribute(NameOID.SURNAME, u'RDNs'),
|
x509.NameAttribute(NameOID.SURNAME, 'RDNs'),
|
||||||
x509.NameAttribute(NameOID.ORGANIZATION_NAME, u'TSLA'),
|
x509.NameAttribute(NameOID.ORGANIZATION_NAME, 'TSLA'),
|
||||||
]),
|
]),
|
||||||
x509.RelativeDistinguishedName([
|
x509.RelativeDistinguishedName([
|
||||||
x509.NameAttribute(NameOID.ORGANIZATION_NAME, u'PyCA')
|
x509.NameAttribute(NameOID.ORGANIZATION_NAME, 'PyCA')
|
||||||
]),
|
]),
|
||||||
])
|
])
|
||||||
expected = [('2.5.4.12', 'Test'), ('CN', 'Multivalue'), ('2.5.4.4', 'RDNs'), ('O', 'TSLA'), ('O', 'PyCA')]
|
expected = [('2.5.4.12', 'Test'), ('CN', 'Multivalue'), ('2.5.4.4', 'RDNs'), ('O', 'TSLA'), ('O', 'PyCA')]
|
||||||
|
Loading…
Reference in New Issue
Block a user