mirror of
https://github.com/Grasscutters/mitmproxy.git
synced 2024-11-21 22:58:24 +00:00
Grpc contentview (#4851)
* Partial gRPC contentview prototype, not linted, no tests, not as add-on * Linted (flake8) * Save dev state * Rewrote of protobuf parser, use decoding strategy, reduced rendered data. Parser uses generators * minor cleanup * fix: preferred encoding was provided as function instead of value * flake8: line length * Backlinked message tree objects, temporary debug out * Partial implementation of gRPC definitions. Save state to fix a cras (data invalidate in edit mode) * hack: deal with missing exception handling for generator based content views * gRPC/Protoparser descriptions (with test code) * replaced manual gzip decoding with mitmproxy.net.encoding.decode * Refactored typing imports * Reafctoring * distinguish request vs response definitions, separate view config from parser config * Code cleaning, moved customized protobuf definitions to example addon * final cleanup * changelog * Stubs for tests * Fixed render_riority of addon example * Started adding tests * Work on tests * mypy * Added pseudo encoder to tests, to cover special decodings * Example addon test added * finalized tests, no 100 percent coverage possible, see comments un uncovered code * minor adjustments * fixup tests * Typos Co-authored-by: Maximilian Hils <git@maximilianhils.com>
This commit is contained in:
parent
aa2ec224c7
commit
480052f58b
@ -9,6 +9,8 @@
|
||||
* Windows: Switch to Python's default asyncio event loop, which increases the number of sockets
|
||||
that can be processed simultaneously (@mhils)
|
||||
* Add `client_replay_concurrency` option, which allows more than one client replay request to be in-flight at a time. (@rbdixon)
|
||||
* New content view which handles gRPC/protobuf. Allows to apply custom definitions to visualize different field decodings.
|
||||
Includes example addon which applies custom definitions for selected gRPC traffic (@mame82)
|
||||
|
||||
## 28 September 2021: mitmproxy 7.0.4
|
||||
|
||||
|
83
examples/addons/contentview-custom-grpc.py
Normal file
83
examples/addons/contentview-custom-grpc.py
Normal file
@ -0,0 +1,83 @@
|
||||
"""
|
||||
Add a custom version of the gRPC/protobuf content view, which parses
|
||||
protobuf messages based on a user defined rule set.
|
||||
|
||||
"""
|
||||
from mitmproxy import contentviews
|
||||
from mitmproxy.contentviews.grpc import ViewGrpcProtobuf, ViewConfig, ProtoParser
|
||||
|
||||
config: ViewConfig = ViewConfig()
|
||||
config.parser_rules = [
|
||||
# Note:
|
||||
#
|
||||
# The first two ParserRules use the same flow filter, although one should reply to request messages and the other to responses.
|
||||
# Even with '~s' and '~q' filter expressions, the whole flow would be matched (for '~s') or not matched at all (for '~q'), if
|
||||
# the contentview displays a http.Message belonging to a flow with existing request and response.
|
||||
# The rules would have to be applied on per-message-basis, instead of per-flow-basis to distinguish request and response (the
|
||||
# contentview deals with a single message, either request or response, the flow filter with a flow contiaing both).
|
||||
#
|
||||
# Thus different ParserRule classes are used to restrict rules to requests or responses were needed:
|
||||
#
|
||||
# - ParserRule: applied to requests and responses
|
||||
# - ParserRuleRequest: applies to requests only
|
||||
# - ParserRuleResponse: applies to responses only
|
||||
#
|
||||
# The actual 'filter' definition in the rule, would still match the whole flow. This means '~u' expressions could
|
||||
# be used, to match the URL from the request of a flow, while the ParserRuleResponse is only applied to the response.
|
||||
|
||||
ProtoParser.ParserRuleRequest(
|
||||
name = "Geo coordinate lookup request",
|
||||
# note on flowfilter: for tflow the port gets appended to the URL's host part
|
||||
filter = "example\\.com.*/ReverseGeocode",
|
||||
field_definitions=[
|
||||
ProtoParser.ParserFieldDefinition(tag="1", name="position"),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.1", name="latitude", intended_decoding=ProtoParser.DecodedTypes.double),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.2", name="longitude", intended_decoding=ProtoParser.DecodedTypes.double),
|
||||
ProtoParser.ParserFieldDefinition(tag="3", name="country"),
|
||||
ProtoParser.ParserFieldDefinition(tag="7", name="app"),
|
||||
]
|
||||
),
|
||||
ProtoParser.ParserRuleResponse(
|
||||
name = "Geo coordinate lookup response",
|
||||
# note on flowfilter: for tflow the port gets appended to the URL's host part
|
||||
filter = "example\\.com.*/ReverseGeocode",
|
||||
field_definitions=[
|
||||
ProtoParser.ParserFieldDefinition(tag="1.2", name="address"),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.3", name="address array element"),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.3.1", name="unknown bytes", intended_decoding=ProtoParser.DecodedTypes.bytes),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.3.2", name="element value long"),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.3.3", name="element value short"),
|
||||
ProtoParser.ParserFieldDefinition(tag="", tag_prefixes=["1.5.1", "1.5.3", "1.5.4", "1.5.5", "1.5.6"], name="position"),
|
||||
ProtoParser.ParserFieldDefinition(tag=".1", tag_prefixes=["1.5.1", "1.5.3", "1.5.4", "1.5.5", "1.5.6"], name="latitude", intended_decoding=ProtoParser.DecodedTypes.double), # noqa: E501
|
||||
ProtoParser.ParserFieldDefinition(tag=".2", tag_prefixes=["1.5.1", "1.5.3", "1.5.4", "1.5.5", "1.5.6"], name="longitude", intended_decoding=ProtoParser.DecodedTypes.double), # noqa: E501
|
||||
ProtoParser.ParserFieldDefinition(tag="7", name="app"),
|
||||
]
|
||||
),
|
||||
]
|
||||
|
||||
|
||||
class ViewGrpcWithRules(ViewGrpcProtobuf):
|
||||
name = "customized gRPC/protobuf"
|
||||
|
||||
def __init__(self) -> None:
|
||||
super().__init__(config=config)
|
||||
|
||||
def __call__(self, *args, **kwargs) -> contentviews.TViewResult:
|
||||
heading, lines = super().__call__(*args, **kwargs)
|
||||
return heading + " (addon with custom rules)", lines
|
||||
|
||||
def render_priority(self, *args, **kwargs) -> float:
|
||||
# increase priority above default gRPC view
|
||||
s_prio = super().render_priority(*args, **kwargs)
|
||||
return s_prio + 1 if s_prio > 0 else s_prio
|
||||
|
||||
|
||||
view = ViewGrpcWithRules()
|
||||
|
||||
|
||||
def load(l):
|
||||
contentviews.add(view)
|
||||
|
||||
|
||||
def done():
|
||||
contentviews.remove(view)
|
@ -20,7 +20,7 @@ from mitmproxy import http
|
||||
from mitmproxy.utils import strutils
|
||||
from . import (
|
||||
auto, raw, hex, json, xml_html, wbxml, javascript, css,
|
||||
urlencoded, multipart, image, query, protobuf, msgpack, graphql
|
||||
urlencoded, multipart, image, query, protobuf, msgpack, graphql, grpc
|
||||
)
|
||||
from .base import View, KEY_MAX, format_text, format_dict, TViewResult
|
||||
from ..http import HTTPFlow
|
||||
@ -187,6 +187,7 @@ add(image.ViewImage())
|
||||
add(query.ViewQuery())
|
||||
add(protobuf.ViewProtobuf())
|
||||
add(msgpack.ViewMsgPack())
|
||||
add(grpc.ViewGrpcProtobuf())
|
||||
|
||||
__all__ = [
|
||||
"View", "KEY_MAX", "format_text", "format_dict", "TViewResult",
|
||||
|
881
mitmproxy/contentviews/grpc.py
Normal file
881
mitmproxy/contentviews/grpc.py
Normal file
@ -0,0 +1,881 @@
|
||||
from __future__ import annotations
|
||||
|
||||
import struct
|
||||
from dataclasses import dataclass, field
|
||||
from enum import Enum
|
||||
from typing import Dict, Generator, Iterable, Iterator, List, Optional, Tuple, Union
|
||||
|
||||
from mitmproxy import contentviews, ctx, flow, flowfilter, http
|
||||
from mitmproxy.contentviews import base
|
||||
from mitmproxy.contrib.kaitaistruct.google_protobuf import GoogleProtobuf
|
||||
from mitmproxy.contrib.kaitaistruct.vlq_base128_le import VlqBase128Le
|
||||
from mitmproxy.net.encoding import decode
|
||||
|
||||
|
||||
class ProtoParser:
|
||||
@dataclass
|
||||
class ParserRule:
|
||||
"""
|
||||
A parser rule lists Field definitions which are applied if the filter rule matches the flow.
|
||||
|
||||
Matching on flow-level also means, a match applies to request AND response messages.
|
||||
To restrict a rule to a requests only use 'ParserRuleRequest', instead.
|
||||
To restrict a rule to a responses only use 'ParserRuleResponse', instead.
|
||||
"""
|
||||
|
||||
field_definitions: List[ProtoParser.ParserFieldDefinition]
|
||||
"""List of field definitions for this rule """
|
||||
|
||||
name: str = ""
|
||||
"""Name of this rule, only used for debugging"""
|
||||
|
||||
filter: str = ""
|
||||
"""
|
||||
Flowfilter to select which flows to apply to ('~q' and '~s' can not be used to distinguish
|
||||
if the rule should apply to the request or response of a flow. To do so, use ParserRuleRequest
|
||||
or ParserRuleResponse. ParserRule always applies to request and response.)
|
||||
"""
|
||||
|
||||
@dataclass
|
||||
class ParserRuleResponse(ParserRule):
|
||||
"""
|
||||
A parser rule lists Field definitions which are applied if the filter rule matches the flow.
|
||||
|
||||
The rule only applies if the processed message is a server response.
|
||||
"""
|
||||
pass
|
||||
|
||||
@dataclass
|
||||
class ParserRuleRequest(ParserRule):
|
||||
"""
|
||||
A parser rule lists Field definitions which are applied if the filter rule matches the flow.
|
||||
|
||||
The rule only applies if the processed message is a client request.
|
||||
"""
|
||||
pass
|
||||
|
||||
@dataclass
|
||||
class ParserFieldDefinition:
|
||||
"""
|
||||
Defines how to parse a field (or multiple fields with the same tag) in a protobuf messages.
|
||||
|
||||
This allows to apply an intended decoding (f.e. decode uint64 as double instead) and to assign
|
||||
a descriptive name to a field. Field definitions are aggregated into rules, which also holds
|
||||
a filter to match selected HTTP messages.
|
||||
|
||||
The most natural way to use this, is to describe known parts of a single protobuf message
|
||||
in a set of field descriptors, pack them into a rule and set the filter of the rule in a way,
|
||||
that it only applies to proper protobuf messages (f.e. to request traffic against an API endpoint
|
||||
matched by an URL flowfilter)
|
||||
"""
|
||||
|
||||
# A 'tag' could be considered as "absolute path" to match a unique field, yet
|
||||
# protobuf allows to uses the same nested message in different positions of the parent message
|
||||
# The 'tag_prefixes' parameter allows to apply the field definition to different "leafs nodes"
|
||||
# of a message.
|
||||
#
|
||||
# Example 1: match a single, absolute tag
|
||||
# ----------
|
||||
# tag = '1.2'
|
||||
# tag_prefixes = [] (default)
|
||||
#
|
||||
# applies to: tag '1.2'
|
||||
#
|
||||
# Example 2: match multiple tags with same ending
|
||||
# ----------
|
||||
# tag = '1.3'
|
||||
# tag_prefixes = ['1.2.', '2.5.']
|
||||
#
|
||||
# applies to: tag '1.2.1.3' and tag '2.5.1.3'
|
||||
# does not apply to: '1.3', unless tag_prefixes is extended to tag_prefixes = ['1.2', '2.5', '']
|
||||
#
|
||||
# Example 3: match multiple tags
|
||||
# ----------
|
||||
# tag = ''
|
||||
# tag_prefixes = ['1.2', '2.5']
|
||||
#
|
||||
# applies to: tag '1.2' and tag '1.5'
|
||||
|
||||
tag: str
|
||||
"""Field tag for which this description applies (including flattened tag path, f.e. '1.2.2.4')"""
|
||||
|
||||
tag_prefixes: List[str] = field(default_factory=list)
|
||||
"""List of prefixes for tag matching (f.e. tag_prefixes=['1.2.', '2.2.'] with tag='1' matches '1.2.1' and '2.2.1')"""
|
||||
|
||||
intended_decoding: Optional[ProtoParser.DecodedTypes] = None
|
||||
"""optional: intended decoding for visualization (parser fails over to alternate decoding if not possible)"""
|
||||
|
||||
name: Optional[str] = None
|
||||
"""optional: intended field for visualization (parser fails over to alternate decoding if not possible)"""
|
||||
|
||||
@dataclass
|
||||
class ParserOptions:
|
||||
# output should contain wiretype of fields
|
||||
include_wiretype: bool = False
|
||||
|
||||
# output should contain the fields which describe nested messages
|
||||
# (the nested messages bodies are always included, but the "header fields" could
|
||||
# add unnecessary output overhead)
|
||||
exclude_message_headers: bool = False
|
||||
|
||||
# optional: rules
|
||||
# rules: List[ProtoParser.ParserRule] = field(default_factory=list)
|
||||
|
||||
class DecodedTypes(Enum):
|
||||
# varint
|
||||
int32 = 0
|
||||
int64 = 1
|
||||
uint32 = 2
|
||||
uint64 = 3
|
||||
sint32 = 4 # ZigZag encoding
|
||||
sint64 = 5 # ZigZag encoding
|
||||
bool = 6
|
||||
enum = 7
|
||||
# bit_32
|
||||
fixed32 = 8
|
||||
sfixed32 = 9
|
||||
float = 10
|
||||
# bit_64
|
||||
fixed64 = 11
|
||||
sfixed64 = 12
|
||||
double = 13
|
||||
# len_delimited
|
||||
string = 14
|
||||
bytes = 15
|
||||
message = 16
|
||||
packed_repeated_field = 17
|
||||
# helper
|
||||
unknown = 18
|
||||
|
||||
class Message:
|
||||
def __init__(
|
||||
self,
|
||||
data: bytes,
|
||||
options: ProtoParser.ParserOptions,
|
||||
rules: List[ProtoParser.ParserRule],
|
||||
parent_field: ProtoParser.Field = None,
|
||||
) -> None:
|
||||
self.data: bytes = data
|
||||
self.parent_field: Optional[ProtoParser.Field] = parent_field
|
||||
self.options: ProtoParser.ParserOptions = options
|
||||
self.rules: List[ProtoParser.ParserRule] = rules
|
||||
try:
|
||||
self.fields: List[ProtoParser.Field] = self.parse_message_fields(data)
|
||||
except:
|
||||
raise ValueError("not a valid protobuf message")
|
||||
|
||||
def parse_message_fields(self, message: bytes) -> List:
|
||||
res: List[ProtoParser.Field] = []
|
||||
|
||||
pb: GoogleProtobuf = GoogleProtobuf.from_bytes(message)
|
||||
for pair in pb.pairs:
|
||||
tag = pair.field_tag
|
||||
wt = pair.wire_type
|
||||
if wt == GoogleProtobuf.Pair.WireTypes.group_start or wt == GoogleProtobuf.Pair.WireTypes.group_end:
|
||||
# ignore deprecated types without values
|
||||
continue
|
||||
v: Union[GoogleProtobuf.DelimitedBytes, VlqBase128Le] = pair.value # for WireType bit-32 and bit-64
|
||||
preferred_decoding = ProtoParser.DecodedTypes.unknown
|
||||
# see: https://www.oreilly.com/library/view/grpc-up-and/9781492058328/ch04.html
|
||||
if wt == GoogleProtobuf.Pair.WireTypes.len_delimited:
|
||||
assert isinstance(v, GoogleProtobuf.DelimitedBytes)
|
||||
v = v.body
|
||||
assert isinstance(v, bytes)
|
||||
# always try to parse length delimited data as nested protobuf message
|
||||
preferred_decoding = ProtoParser.DecodedTypes.message
|
||||
if wt == GoogleProtobuf.Pair.WireTypes.varint:
|
||||
assert isinstance(v, VlqBase128Le)
|
||||
v = v.value
|
||||
assert isinstance(v, int)
|
||||
if v.bit_length() > 32:
|
||||
preferred_decoding = ProtoParser.DecodedTypes.uint64
|
||||
else:
|
||||
preferred_decoding = ProtoParser.DecodedTypes.uint32
|
||||
if wt == GoogleProtobuf.Pair.WireTypes.bit_64:
|
||||
# exists in Protobuf for efficient encoding, when decoded comes down to uint64
|
||||
assert isinstance(v, int)
|
||||
preferred_decoding = ProtoParser.DecodedTypes.fixed64
|
||||
if wt == GoogleProtobuf.Pair.WireTypes.bit_32:
|
||||
# exists in Protobuf for efficient encoding, when decoded comes down to uint32
|
||||
assert isinstance(v, int)
|
||||
preferred_decoding = ProtoParser.DecodedTypes.fixed32
|
||||
|
||||
field = ProtoParser.Field(
|
||||
preferred_decoding=preferred_decoding,
|
||||
wire_type=wt,
|
||||
tag=tag,
|
||||
wire_value=v,
|
||||
owning_message=self,
|
||||
options=self.options,
|
||||
rules=self.rules
|
||||
)
|
||||
res.append(field)
|
||||
return res
|
||||
|
||||
def gen_fields(self) -> Generator[ProtoParser.Field, None, None]:
|
||||
for f in self.fields:
|
||||
yield f
|
||||
|
||||
def gen_flat_decoded_field_dicts(self) -> Generator[Dict, None, None]:
|
||||
"""
|
||||
This generator returns a flattened version of the fields from a message (including nested fields)
|
||||
|
||||
A single entry has the form:
|
||||
{
|
||||
"tag": str # fully qualified tag (all tags starting from the root message, concatenated with '.' delimiter)
|
||||
"wireType": str # describes the wire encoding used by the field
|
||||
"decoding": str # describes the chosen decoding (interpretation of wire encoding, according to protobuf types)
|
||||
"val": Union[bool, str, bytes, int, float] # the decoded value in python representation
|
||||
}
|
||||
"""
|
||||
# iterate over fields
|
||||
for f in self.gen_fields():
|
||||
# convert field and nested fields to dicts
|
||||
for d in f.gen_flat_decoded_field_dicts():
|
||||
yield d
|
||||
|
||||
def gen_string_rows(self) -> Generator[Tuple[str, ...], None, None]:
|
||||
# Excluding fields containing message headers simplifies the view, but without
|
||||
# knowing the message tags, they can not be used in a custom definition, in order
|
||||
# to declare a different interpretation for the message (the message is a length-delimeted
|
||||
# field value, which could alternatively be parsed as 'str' or 'bytes' if the field tag
|
||||
# is known)
|
||||
for field_dict in self.gen_flat_decoded_field_dicts():
|
||||
if self.options.exclude_message_headers and field_dict["decoding"] == "message":
|
||||
continue
|
||||
|
||||
if self.options.include_wiretype:
|
||||
col1 = "[{}->{}]".format(field_dict["wireType"], field_dict["decoding"])
|
||||
else:
|
||||
col1 = "[{}]".format(field_dict["decoding"])
|
||||
col2 = field_dict["name"] # empty string if not set (consumes no space)
|
||||
col3 = field_dict["tag"]
|
||||
col4 = str(field_dict["val"])
|
||||
yield col1, col2, col3, col4
|
||||
|
||||
class Field:
|
||||
"""
|
||||
Represents a single field of a protobuf message and handles the varios encodings.
|
||||
|
||||
As mitmproxy sees the data passing by as raw protobuf message, it only knows the
|
||||
WireTypes. Each of the WireTypes could represent different Protobuf field types.
|
||||
The exact Protobuf field type can not be determined from the wire format, thus different
|
||||
options for decoding have to be supported.
|
||||
In addition the parsed WireTypes are (intermediary) stored in Python types, which adds
|
||||
some additional overhead type conversions.
|
||||
|
||||
WireType represented Protobuf Types Python type (intermediary)
|
||||
|
||||
0: varint int32, int64, uint32, uint64, enum, int (*)
|
||||
sint32, sint64 (both ZigZag encoded), int
|
||||
bool bool
|
||||
float (**)
|
||||
|
||||
1: bit_64 fixed64, sfixed64, int (*)
|
||||
double float
|
||||
|
||||
2: len_delimited string, str
|
||||
message, class 'Message'
|
||||
bytes, bytes (*)
|
||||
packed_repeated_field class 'Message' (fields with same tag)
|
||||
|
||||
3: group_start unused (deprecated) -
|
||||
4: group_end unused (deprecated) -
|
||||
|
||||
5: bit_32 fixed32, sfixed32, int (*)
|
||||
float float
|
||||
|
||||
(*) Note 1: Conversion between WireType and intermediary python representation
|
||||
is handled by Kaitai protobuf decoder and always uses the python
|
||||
representation marked with (*). Converting to alternative representations
|
||||
is handled inside this class.
|
||||
(**) Note 2: Varint is not used to represent floating point values, but some applications
|
||||
store native floats in uint32 protobuf types (or native double in uint64).
|
||||
Thus we allow conversion of varint to floating point values for convenience
|
||||
(A well known APIs "hide" GPS latitude and longitude values in varint types,
|
||||
much easier to spot such things when rendered as float)
|
||||
|
||||
Ref: - https://developers.google.com/protocol-buffers/docs/proto3
|
||||
- https://developers.google.com/protocol-buffers/docs/encoding
|
||||
"""
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
wire_type: GoogleProtobuf.Pair.WireTypes,
|
||||
preferred_decoding: ProtoParser.DecodedTypes,
|
||||
tag: int,
|
||||
wire_value: Union[int, bytes],
|
||||
owning_message: ProtoParser.Message,
|
||||
options: ProtoParser.ParserOptions,
|
||||
rules: List[ProtoParser.ParserRule]
|
||||
) -> None:
|
||||
self.wire_type: GoogleProtobuf.Pair.WireTypes = wire_type
|
||||
self.preferred_decoding: ProtoParser.DecodedTypes = preferred_decoding
|
||||
self.wire_value: Union[int, bytes] = wire_value
|
||||
self.tag: int = tag
|
||||
self.owning_message: ProtoParser.Message = owning_message
|
||||
self.options: ProtoParser.ParserOptions = options
|
||||
self.name: str = ""
|
||||
self.rules: List[ProtoParser.ParserRule] = rules
|
||||
self.parent_tags: List[int]
|
||||
if not self.owning_message.parent_field:
|
||||
self.parent_tags = []
|
||||
else:
|
||||
self.parent_tags = self.owning_message.parent_field.parent_tags[:]
|
||||
self.parent_tags.append(self.owning_message.parent_field.tag)
|
||||
|
||||
self.apply_rules()
|
||||
|
||||
# no tests for only_first_hit=False, as not user-changable
|
||||
def apply_rules(self, only_first_hit=True):
|
||||
tag_str = self._gen_tag_str()
|
||||
name = None
|
||||
decoding = None
|
||||
try:
|
||||
for rule in self.rules:
|
||||
for fd in rule.field_definitions:
|
||||
match = False
|
||||
if len(fd.tag_prefixes) == 0 and fd.tag == tag_str:
|
||||
match = True
|
||||
else:
|
||||
for rt in fd.tag_prefixes:
|
||||
if rt + fd.tag == tag_str:
|
||||
match = True
|
||||
break
|
||||
if match:
|
||||
if only_first_hit:
|
||||
# only first match
|
||||
self.name = fd.name
|
||||
self.preferred_decoding = fd.intended_decoding
|
||||
return
|
||||
else:
|
||||
# overwrite matches till last rule was inspected
|
||||
# (f.e. allows to define name in one rule and intended_decoding in another one)
|
||||
name = fd.name if fd.name else name
|
||||
decoding = fd.intended_decoding if fd.intended_decoding else decoding
|
||||
|
||||
if name:
|
||||
self.name = name
|
||||
if decoding:
|
||||
self.preferred_decoding = decoding
|
||||
except Exception as e:
|
||||
ctx.log.warn(e)
|
||||
pass
|
||||
|
||||
def _gen_tag_str(self):
|
||||
tags = self.parent_tags[:]
|
||||
tags.append(self.tag)
|
||||
return ".".join([str(tag) for tag in tags])
|
||||
|
||||
def safe_decode_as(
|
||||
self,
|
||||
intended_decoding: ProtoParser.DecodedTypes
|
||||
) -> Tuple[ProtoParser.DecodedTypes, Union[bool, float, int, bytes, str, ProtoParser.Message]]:
|
||||
"""
|
||||
Tries to decode as intended, applies failover, if not possible
|
||||
|
||||
Returns selected decoding and decoded value
|
||||
"""
|
||||
if self.wire_type == GoogleProtobuf.Pair.WireTypes.varint:
|
||||
try:
|
||||
return intended_decoding, self.decode_as(intended_decoding)
|
||||
except:
|
||||
if int(self.wire_value).bit_length() > 32:
|
||||
# ignore the fact that varint could exceed 64bit (would violate the specs)
|
||||
return ProtoParser.DecodedTypes.uint64, self.wire_value
|
||||
else:
|
||||
return ProtoParser.DecodedTypes.uint32, self.wire_value
|
||||
elif self.wire_type == GoogleProtobuf.Pair.WireTypes.bit_64:
|
||||
try:
|
||||
return intended_decoding, self.decode_as(intended_decoding)
|
||||
except:
|
||||
return ProtoParser.DecodedTypes.fixed64, self.wire_value
|
||||
elif self.wire_type == GoogleProtobuf.Pair.WireTypes.bit_32:
|
||||
try:
|
||||
return intended_decoding, self.decode_as(intended_decoding)
|
||||
except:
|
||||
return ProtoParser.DecodedTypes.fixed32, self.wire_value
|
||||
elif self.wire_type == GoogleProtobuf.Pair.WireTypes.len_delimited:
|
||||
try:
|
||||
return intended_decoding, self.decode_as(intended_decoding)
|
||||
except:
|
||||
# failover strategy: message --> string (valid UTF-8) --> bytes
|
||||
len_delimited_strategy: List[ProtoParser.DecodedTypes] = [
|
||||
ProtoParser.DecodedTypes.message,
|
||||
ProtoParser.DecodedTypes.string,
|
||||
ProtoParser.DecodedTypes.bytes # should always work
|
||||
]
|
||||
for failover_decoding in len_delimited_strategy:
|
||||
if failover_decoding == intended_decoding:
|
||||
continue # don't try it twice
|
||||
try:
|
||||
return failover_decoding, self.decode_as(failover_decoding)
|
||||
except:
|
||||
# move on with next
|
||||
pass
|
||||
|
||||
# we should never get here (could not be added to tests)
|
||||
return ProtoParser.DecodedTypes.unknown, self.wire_value
|
||||
|
||||
def decode_as(
|
||||
self,
|
||||
intended_decoding: ProtoParser.DecodedTypes
|
||||
) -> Union[bool, int, float, bytes, str, ProtoParser.Message]:
|
||||
if self.wire_type == GoogleProtobuf.Pair.WireTypes.varint:
|
||||
assert isinstance(self.wire_value, int)
|
||||
if intended_decoding == ProtoParser.DecodedTypes.bool:
|
||||
return self.wire_value != 0
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.int32:
|
||||
if self.wire_value.bit_length() > 32:
|
||||
raise TypeError("wire value too large for int32")
|
||||
return struct.unpack("!i", struct.pack("!I", self.wire_value))[0]
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.int64:
|
||||
if self.wire_value.bit_length() > 64:
|
||||
# currently avoided by kaitai decoder (can not be added to tests)
|
||||
raise TypeError("wire value too large for int64")
|
||||
return struct.unpack("!q", struct.pack("!Q", self.wire_value))[0]
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.uint32:
|
||||
if self.wire_value.bit_length() > 32:
|
||||
raise TypeError("wire value too large for uint32")
|
||||
return self.wire_value # already 'int' which was parsed as unsigned
|
||||
elif (
|
||||
intended_decoding == ProtoParser.DecodedTypes.uint64 or
|
||||
intended_decoding == ProtoParser.DecodedTypes.enum
|
||||
):
|
||||
if self.wire_value.bit_length() > 64:
|
||||
# currently avoided by kaitai decoder (can not be added to tests)
|
||||
raise TypeError("wire value too large")
|
||||
return self.wire_value # already 'int' which was parsed as unsigned
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.sint32:
|
||||
if self.wire_value.bit_length() > 32:
|
||||
raise TypeError("wire value too large for sint32")
|
||||
return (self.wire_value >> 1) ^ -(self.wire_value & 1) # zigzag_decode
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.sint64:
|
||||
if self.wire_value.bit_length() > 64:
|
||||
# currently avoided by kaitai decoder (can not be added to tests)
|
||||
raise TypeError("wire value too large for sint64")
|
||||
# ZigZag decode
|
||||
# Ref: https://gist.github.com/mfuerstenau/ba870a29e16536fdbaba
|
||||
return (self.wire_value >> 1) ^ -(self.wire_value & 1)
|
||||
elif (
|
||||
intended_decoding == ProtoParser.DecodedTypes.float or
|
||||
intended_decoding == ProtoParser.DecodedTypes.double
|
||||
):
|
||||
# special case, not complying to protobuf specs
|
||||
return self._wire_value_as_float()
|
||||
elif self.wire_type == GoogleProtobuf.Pair.WireTypes.bit_64:
|
||||
if intended_decoding == ProtoParser.DecodedTypes.fixed64:
|
||||
return self.wire_value
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.sfixed64:
|
||||
return struct.unpack("!q", struct.pack("!Q", self.wire_value))[0]
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.double:
|
||||
return self._wire_value_as_float()
|
||||
elif self.wire_type == GoogleProtobuf.Pair.WireTypes.bit_32:
|
||||
if intended_decoding == ProtoParser.DecodedTypes.fixed32:
|
||||
return self.wire_value
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.sfixed32:
|
||||
return struct.unpack("!i", struct.pack("!I", self.wire_value))[0]
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.float:
|
||||
return self._wire_value_as_float()
|
||||
elif self.wire_type == GoogleProtobuf.Pair.WireTypes.len_delimited:
|
||||
assert isinstance(self.wire_value, bytes)
|
||||
if intended_decoding == ProtoParser.DecodedTypes.string:
|
||||
# According to specs, a protobuf string HAS TO be UTF-8 parsable
|
||||
# throw exception on invalid UTF-8 chars, but escape linebreaks
|
||||
return self.wire_value_as_utf8(escape_invalid=False, escape_newline=True)
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.bytes:
|
||||
# always works, assure to hand back a copy
|
||||
return self.wire_value[:]
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.packed_repeated_field:
|
||||
raise NotImplementedError("currently not needed")
|
||||
elif intended_decoding == ProtoParser.DecodedTypes.message:
|
||||
return ProtoParser.Message(
|
||||
data=self.wire_value,
|
||||
options=self.options,
|
||||
parent_field=self,
|
||||
rules=self.rules
|
||||
)
|
||||
|
||||
# if here, there is no valid decoding
|
||||
raise TypeError("intended decoding mismatches wire type")
|
||||
|
||||
def encode_from(inputval, intended_encoding: ProtoParser.DecodedTypes):
|
||||
raise NotImplementedError(
|
||||
"Future work, needed to manipulate and re-encode protobuf message, with respect to given wire types"
|
||||
)
|
||||
|
||||
def _wire_value_as_float(self) -> float:
|
||||
"""
|
||||
Handles double (64bit) and float (32bit).
|
||||
Assumes Network Byte Order (big endian).
|
||||
|
||||
Usable for:
|
||||
|
||||
WireType --> Protobuf Type):
|
||||
----------------------------
|
||||
varint --> double/float (not intended by ProtoBuf, but used in the wild)
|
||||
bit_32 --> float
|
||||
bit_64 --> double
|
||||
len_delimited --> 4 bytes: float / 8 bytes: double / other sizes return NaN
|
||||
"""
|
||||
v = self._value_as_bytes()
|
||||
if len(v) == 4:
|
||||
return struct.unpack("!f", v)[0]
|
||||
elif len(v) == 8:
|
||||
return struct.unpack("!d", v)[0]
|
||||
# no need to raise an Exception
|
||||
raise TypeError("can not be converted to floatingpoint representation")
|
||||
|
||||
def _value_as_bytes(self) -> bytes:
|
||||
if isinstance(self.wire_value, bytes):
|
||||
return self.wire_value
|
||||
elif isinstance(self.wire_value, int):
|
||||
if self.wire_value.bit_length() > 64:
|
||||
# source for a python int are wiretypes varint/bit_32/bit64 and should never convert to int values 64bit
|
||||
# currently avoided by kaitai decoder (can not be added to tests)
|
||||
raise ValueError("Value exceeds 64bit, violating protobuf specs")
|
||||
elif self.wire_value.bit_length() > 32:
|
||||
# packing uses network byte order (to assure consistent results across architectures)
|
||||
return struct.pack("!Q", self.wire_value)
|
||||
else:
|
||||
# packing uses network byte order (to assure consistent results across architectures)
|
||||
return struct.pack("!I", self.wire_value)
|
||||
else:
|
||||
# should never happen, no tests
|
||||
raise ValueError("can not be converted to bytes")
|
||||
|
||||
def _wire_type_str(self):
|
||||
return str(self.wire_type).split(".")[-1]
|
||||
|
||||
def _decoding_str(self, decoding: ProtoParser.DecodedTypes):
|
||||
return str(decoding).split(".")[-1]
|
||||
|
||||
def wire_value_as_utf8(self, escape_invalid=True, escape_newline=True) -> str:
|
||||
if isinstance(self.wire_value, bytes):
|
||||
if escape_invalid:
|
||||
res = self.wire_value.decode("utf-8", "backslashreplace")
|
||||
else:
|
||||
res = self.wire_value.decode("utf-8")
|
||||
return res.replace("\n", "\\n") if escape_newline else res
|
||||
return str(self.wire_value)
|
||||
|
||||
def gen_flat_decoded_field_dicts(self) -> Generator[Dict, None, None]:
|
||||
"""
|
||||
Returns a generator which passes the field as a dict.
|
||||
|
||||
In order to return the field value it gets decoded (based on a failover strategy and
|
||||
provided ParserRules).
|
||||
If the field holds a nested message, the fields contained in the message are appended.
|
||||
Ultimately this flattens all fields recursively.
|
||||
"""
|
||||
selected_decoding, decoded_val = self.safe_decode_as(self.preferred_decoding)
|
||||
field_desc_dict = {
|
||||
"tag": self._gen_tag_str(),
|
||||
"wireType": self._wire_type_str(),
|
||||
"decoding": self._decoding_str(selected_decoding),
|
||||
"name": self.name,
|
||||
}
|
||||
if isinstance(decoded_val, ProtoParser.Message):
|
||||
field_desc_dict["val"] = "" # message has no value, because contained fields get appended (flattened)
|
||||
yield field_desc_dict
|
||||
# the value is an embedded message, thus add the message fields
|
||||
for f in decoded_val.gen_fields():
|
||||
for field_dict in f.gen_flat_decoded_field_dicts():
|
||||
yield field_dict
|
||||
else:
|
||||
field_desc_dict["val"] = decoded_val
|
||||
yield field_desc_dict
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
data: bytes,
|
||||
rules: List[ProtoParser.ParserRule] = None,
|
||||
parser_options: ParserOptions = None
|
||||
) -> None:
|
||||
self.data: bytes = data
|
||||
if parser_options is None:
|
||||
parser_options = ProtoParser.ParserOptions()
|
||||
self.options = parser_options
|
||||
if rules is None:
|
||||
rules = []
|
||||
self.rules = rules
|
||||
self.root_message: ProtoParser.Message = ProtoParser.Message(
|
||||
data=data,
|
||||
options=self.options,
|
||||
rules=self.rules
|
||||
)
|
||||
|
||||
def gen_str_rows(self) -> Generator[Tuple[str, ...], None, None]:
|
||||
for f in self.root_message.gen_string_rows():
|
||||
yield f
|
||||
|
||||
|
||||
# Note: all content view formating functionality is kept out of the ProtoParser class, to
|
||||
# allow it to be use independently.
|
||||
# This function is generic enough, to consider moving it to mitmproxy.contentviews.base
|
||||
def format_table(
|
||||
table_rows: Iterable[Tuple[str, ...]],
|
||||
max_col_width=100,
|
||||
) -> Iterator[base.TViewLine]:
|
||||
"""
|
||||
Helper function to render tables with variable column count (move to contentview base, if needed elsewhere)
|
||||
|
||||
Note: The function has to convert generators to a list, as all rows have to be processed twice (to determine
|
||||
the column widths first).
|
||||
"""
|
||||
rows: List[Tuple[str, ...]] = []
|
||||
col_count = 0
|
||||
cols_width: List[int] = []
|
||||
for row in table_rows:
|
||||
col_count = max(col_count, len(row))
|
||||
while len(cols_width) < col_count:
|
||||
cols_width.append(0)
|
||||
for col_num in range(len(row)):
|
||||
cols_width[col_num] = max(len(row[col_num]), cols_width[col_num])
|
||||
|
||||
# store row in list
|
||||
rows.append(row)
|
||||
|
||||
for i in range(len(cols_width)):
|
||||
cols_width[i] = min(cols_width[i], max_col_width)
|
||||
|
||||
for row in rows:
|
||||
line: base.TViewLine = []
|
||||
for col_num in range(len(row)):
|
||||
col_val = row[col_num].ljust(cols_width[col_num] + 2)
|
||||
line.append(("text", col_val))
|
||||
yield line
|
||||
|
||||
|
||||
def parse_grpc_messages(data, compression_scheme) -> Generator[Tuple[bool, bytes], None, None]:
|
||||
"""Generator iterates over body data and returns a boolean indicating if the messages
|
||||
was compressed, along with the raw message data (decompressed) for each gRPC message
|
||||
contained in the body data"""
|
||||
while data:
|
||||
try:
|
||||
msg_is_compressed, length = struct.unpack('!?i', data[:5])
|
||||
decoded_message = struct.unpack('!%is' % length, data[5:5 + length])[0]
|
||||
except Exception as e:
|
||||
raise ValueError("invalid gRPC message") from e
|
||||
|
||||
if msg_is_compressed:
|
||||
try:
|
||||
decoded_message = decode(encoded=decoded_message, encoding=compression_scheme)
|
||||
except Exception as e:
|
||||
raise ValueError("Failed to decompress gRPC message with gzip") from e
|
||||
|
||||
yield msg_is_compressed, decoded_message
|
||||
data = data[5 + length:]
|
||||
|
||||
|
||||
# hacky fix for mitmproxy issue:
|
||||
#
|
||||
# mitmproxy handles Exceptions in the contenview's __call__ function, by
|
||||
# failing over to 'Raw' view. The intention was to use this behavior to
|
||||
# pass up Exceptions thrown inside the generator function ('format_pbuf'
|
||||
# and 'format_grpc') to the __call__ function.
|
||||
# This usually works fine if the contentview is initialized on a flow
|
||||
# with invalid data.
|
||||
# When the flow data gets invalidated in the edit mode, mitmproxy re-calls
|
||||
# the generator functions outside the contentviews '__call__' method.
|
||||
#
|
||||
# This happens in the 'safe_to_print' function of 'mitmproxy/contentvies/__init__.py'
|
||||
#
|
||||
# def safe_to_print(lines, encoding="utf8"):
|
||||
# """
|
||||
# Wraps a content generator so that each text portion is a *safe to print* unicode string.
|
||||
# """
|
||||
# for line in lines: # <------ this code re-iterates lines and thus calls generators, without using the views __call__ function
|
||||
# clean_line = []
|
||||
# for (style, text) in line:
|
||||
# if isinstance(text, bytes):
|
||||
# text = text.decode(encoding, "replace")
|
||||
# text = strutils.escape_control_characters(text)
|
||||
# clean_line.append((style, text))
|
||||
# yield clean_line
|
||||
#
|
||||
# In result, mitmproxy crashes if the generator functions raise Exception to indicate
|
||||
# data parsing errors.
|
||||
# To deal with this, the generator function gets converted into a list inside the
|
||||
# __call__ function. Ultimately, exceptions are raised directly from within __call__
|
||||
# instead in cases where the generator is accessed externally without exception handling.
|
||||
def hack_generator_to_list(generator_func):
|
||||
return list(generator_func)
|
||||
|
||||
|
||||
def format_pbuf(message: bytes, parser_options: ProtoParser.ParserOptions, rules: List[ProtoParser.ParserRule]):
|
||||
for l in format_table(ProtoParser(data=message, parser_options=parser_options, rules=rules).gen_str_rows()):
|
||||
yield l
|
||||
|
||||
|
||||
def format_grpc(
|
||||
data: bytes,
|
||||
parser_options: ProtoParser.ParserOptions,
|
||||
rules: List[ProtoParser.ParserRule],
|
||||
compression_scheme="gzip"
|
||||
):
|
||||
message_count = 0
|
||||
for compressed, pb_message in parse_grpc_messages(data=data, compression_scheme=compression_scheme):
|
||||
headline = 'gRPC message ' + str(message_count) + ' (compressed ' + str(
|
||||
compression_scheme if compressed else compressed) + ')'
|
||||
|
||||
yield [("text", headline)]
|
||||
for l in format_pbuf(
|
||||
message=pb_message,
|
||||
parser_options=parser_options,
|
||||
rules=rules
|
||||
):
|
||||
yield l
|
||||
|
||||
|
||||
@dataclass
|
||||
class ViewConfig:
|
||||
parser_options: ProtoParser.ParserOptions = ProtoParser.ParserOptions()
|
||||
parser_rules: List[ProtoParser.ParserRule] = field(default_factory=list)
|
||||
|
||||
|
||||
class ViewGrpcProtobuf(base.View):
|
||||
"""Human friendly view of protocol buffers"""
|
||||
|
||||
name = "gRPC/Protocol Buffer"
|
||||
__content_types_pb = [
|
||||
"application/x-protobuf",
|
||||
"application/x-protobuffer",
|
||||
"application/grpc-proto",
|
||||
]
|
||||
__content_types_grpc = [
|
||||
"application/grpc",
|
||||
]
|
||||
|
||||
# first value serves as default algorithm for compressed messages, if 'grpc-encoding' header is missing
|
||||
__valid_grpc_encodings = [
|
||||
"gzip",
|
||||
"identity",
|
||||
"deflate",
|
||||
]
|
||||
|
||||
# allows to take external ParserOptions object. goes with defaults otherwise
|
||||
def __init__(self, config: ViewConfig = None) -> None:
|
||||
super().__init__()
|
||||
if config is None:
|
||||
config = ViewConfig()
|
||||
self.config = config
|
||||
|
||||
def _matching_rules(
|
||||
self,
|
||||
rules: List[ProtoParser.ParserRule],
|
||||
message: Optional[http.Message],
|
||||
flow: Optional[flow.Flow]
|
||||
) -> List[ProtoParser.ParserRule]:
|
||||
"""
|
||||
Checks which of the give rules applies and returns a List only containing those rules
|
||||
|
||||
Each rule defines a flow filter in rule.filter which is usually matched against a flow.
|
||||
When it comes to protobuf parsing, in most cases request messages differ from response messages.
|
||||
Thus, it has to be possible to apply a rule to a http.Request or a http.Response, only.
|
||||
|
||||
As the name flowfilter suggests, filters are working on a flow-level, not on message-level.
|
||||
This means:
|
||||
|
||||
- the filter expression '~q' matches all flows with a request, but no response
|
||||
- the filter expression '~s' matches all flows with a response
|
||||
|
||||
In result, for complete flows (with a gRPC message in the request and the response), ParserRules would
|
||||
either be applied to request and response at the same time ('~s') or neither would match request, nor
|
||||
response (~q).
|
||||
|
||||
To distinguish between rules which should be applied to response messages, request messages or both
|
||||
(while being applied to the whole flow), different classes with same behavior are used to wrap rules:
|
||||
|
||||
- ParserRule: applies to requests and responses
|
||||
- ParserRuleRequest: applies to requests only
|
||||
- ParserRuleResponse: applies to responses only
|
||||
"""
|
||||
res: List[ProtoParser.ParserRule] = []
|
||||
if not flow:
|
||||
return res
|
||||
is_request = isinstance(message, http.Request)
|
||||
for rule in rules:
|
||||
# message based rule matching
|
||||
if is_request and isinstance(rule, ProtoParser.ParserRuleResponse):
|
||||
continue
|
||||
elif not is_request and isinstance(rule, ProtoParser.ParserRuleRequest):
|
||||
continue
|
||||
# flow based rule matching
|
||||
if flowfilter.match(rule.filter, flow=flow):
|
||||
res.append(rule)
|
||||
return res
|
||||
|
||||
def __call__(
|
||||
self,
|
||||
data: bytes,
|
||||
*,
|
||||
content_type: Optional[str] = None,
|
||||
flow: Optional[flow.Flow] = None,
|
||||
http_message: Optional[http.Message] = None,
|
||||
**unknown_metadata,
|
||||
) -> contentviews.TViewResult:
|
||||
applicabble_rules = self._matching_rules(rules=self.config.parser_rules, flow=flow, message=http_message)
|
||||
if content_type in self.__content_types_grpc:
|
||||
# If gRPC messages are flagged to be compressed, the compression algorithm is expressed in the
|
||||
# 'grpc-encoding' header.
|
||||
#
|
||||
# The following code tries to determine the compression algorithm base on this header.
|
||||
# If the header is not present or contains an unsupported compression, the logic falls back to
|
||||
# 'gzip'.
|
||||
#
|
||||
# If a compressed gRPC message is found in the body data (compressed flag set), the information
|
||||
# on the compression scheme is needed (even if not set by a header), in order to process the message.
|
||||
# Thus we assure there is always an encoding selected. An encoding of 'Identity' would not make
|
||||
# sense, if a message is flagged as being compressed, that's why a default is chosen.
|
||||
try:
|
||||
assert http_message is not None
|
||||
h = http_message.headers["grpc-encoding"]
|
||||
grpc_encoding = h if h in self.__valid_grpc_encodings else self.__valid_grpc_encodings[0]
|
||||
except:
|
||||
grpc_encoding = self.__valid_grpc_encodings[0]
|
||||
|
||||
text_iter = format_grpc(
|
||||
data=data,
|
||||
parser_options=self.config.parser_options,
|
||||
compression_scheme=grpc_encoding,
|
||||
rules=applicabble_rules
|
||||
)
|
||||
title = "gRPC"
|
||||
else:
|
||||
text_iter = format_pbuf(
|
||||
message=data,
|
||||
parser_options=self.config.parser_options,
|
||||
rules=applicabble_rules
|
||||
)
|
||||
title = "Protobuf (flattened)"
|
||||
|
||||
# hacky bugfix, see description above generator functions format_pbuf/format_grpc
|
||||
try:
|
||||
text_iter = hack_generator_to_list(text_iter)
|
||||
except Exception as e:
|
||||
# hook to log exception tracebacks on iterators
|
||||
|
||||
# import traceback
|
||||
# ctx.log.warn("gRPC contentview: {}".format(traceback.format_exc()))
|
||||
raise e
|
||||
|
||||
return title, text_iter
|
||||
|
||||
def render_priority(
|
||||
self,
|
||||
data: bytes,
|
||||
*,
|
||||
content_type: Optional[str] = None,
|
||||
flow: Optional[flow.Flow] = None,
|
||||
http_message: Optional[http.Message] = None,
|
||||
**unknown_metadata,
|
||||
) -> float:
|
||||
|
||||
if bool(data) and content_type in self.__content_types_grpc:
|
||||
return 1
|
||||
if bool(data) and content_type in self.__content_types_pb:
|
||||
# replace existing protobuf renderer preference (adjust by option)
|
||||
return 1.5
|
||||
else:
|
||||
return 0
|
@ -51,6 +51,7 @@ exclude =
|
||||
mitmproxy/addons/termlog.py
|
||||
mitmproxy/connections.py
|
||||
mitmproxy/contentviews/base.py
|
||||
mitmproxy/contentviews/grpc.py
|
||||
mitmproxy/controller.py
|
||||
mitmproxy/ctx.py
|
||||
mitmproxy/exceptions.py
|
||||
|
@ -22,6 +22,40 @@ class TestScripts(tservers.MasterTest):
|
||||
_, fmt = swapcase(b"<html>Test!</html>")
|
||||
assert any(b'tEST!' in val[0][1] for val in fmt)
|
||||
|
||||
def test_custom_grpc_contentview(self, tdata):
|
||||
with taddons.context() as tctx:
|
||||
tctx.script(tdata.path("../examples/addons/contentview-custom-grpc.py"))
|
||||
v = contentviews.get("customized gRPC/protobuf")
|
||||
|
||||
p = tdata.path("mitmproxy/contentviews/test_grpc_data/msg1.bin")
|
||||
with open(p, "rb") as f:
|
||||
raw = f.read()
|
||||
|
||||
sim_msg_req = tutils.treq(
|
||||
port=443,
|
||||
host="example.com",
|
||||
path="/ReverseGeocode"
|
||||
)
|
||||
|
||||
sim_msg_resp = tutils.tresp()
|
||||
|
||||
sim_flow = tflow.tflow(
|
||||
req=sim_msg_req,
|
||||
resp=sim_msg_resp
|
||||
)
|
||||
|
||||
view_text, output = v(raw, flow=sim_flow, http_message=sim_flow.request) # simulate request message
|
||||
assert view_text == "Protobuf (flattened) (addon with custom rules)"
|
||||
output = list(output) # assure list conversion if generator
|
||||
assert output == [
|
||||
[('text', '[message] '), ('text', 'position '), ('text', '1 '), ('text', ' ')],
|
||||
[('text', '[double] '), ('text', 'latitude '), ('text', '1.1 '), ('text', '38.89816675798073 ')],
|
||||
[('text', '[double] '), ('text', 'longitude '), ('text', '1.2 '), ('text', '-77.03829828366696 ')],
|
||||
[('text', '[string] '), ('text', 'country '), ('text', '3 '), ('text', 'de_DE ')],
|
||||
[('text', '[uint32] '), ('text', ' '), ('text', '6 '), ('text', '1 ')],
|
||||
[('text', '[string] '), ('text', 'app '), ('text', '7 '), ('text', 'de.mcdonalds.mcdonaldsinfoapp ')]
|
||||
]
|
||||
|
||||
def test_modify_form(self, tdata):
|
||||
with taddons.context() as tctx:
|
||||
sc = tctx.script(tdata.path("../examples/addons/http-modify-form.py"))
|
||||
|
387
test/mitmproxy/contentviews/test_grpc.py
Normal file
387
test/mitmproxy/contentviews/test_grpc.py
Normal file
@ -0,0 +1,387 @@
|
||||
import pytest
|
||||
|
||||
from mitmproxy.contentviews import grpc
|
||||
from mitmproxy.contentviews.grpc import ViewGrpcProtobuf, ViewConfig, ProtoParser, parse_grpc_messages
|
||||
from mitmproxy.net.encoding import encode
|
||||
from mitmproxy.test import tflow, tutils
|
||||
import struct
|
||||
from . import full_eval
|
||||
|
||||
datadir = "mitmproxy/contentviews/test_grpc_data/"
|
||||
|
||||
|
||||
def helper_pack_grpc_message(data: bytes, compress=False, encoding="gzip") -> bytes:
|
||||
if compress:
|
||||
data = encode(data, encoding)
|
||||
header = struct.pack('!?i', compress, len(data))
|
||||
return header + data
|
||||
|
||||
|
||||
custom_parser_rules = [
|
||||
ProtoParser.ParserRuleRequest(
|
||||
name = "Geo coordinate lookup request",
|
||||
# note on flowfilter: for tflow the port gets appended to the URL's host part
|
||||
filter = "example\\.com.*/ReverseGeocode",
|
||||
field_definitions=[
|
||||
ProtoParser.ParserFieldDefinition(tag="1", name="position"),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.1", name="latitude", intended_decoding=ProtoParser.DecodedTypes.double),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.2", name="longitude", intended_decoding=ProtoParser.DecodedTypes.double),
|
||||
ProtoParser.ParserFieldDefinition(tag="3", name="country"),
|
||||
ProtoParser.ParserFieldDefinition(tag="7", name="app"),
|
||||
]
|
||||
),
|
||||
ProtoParser.ParserRuleResponse(
|
||||
name = "Geo coordinate lookup response",
|
||||
# note on flowfilter: for tflow the port gets appended to the URL's host part
|
||||
filter = "example\\.com.*/ReverseGeocode",
|
||||
field_definitions=[
|
||||
ProtoParser.ParserFieldDefinition(tag="1.2", name="address"),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.3", name="address array element"),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.3.1", name="unknown bytes", intended_decoding=ProtoParser.DecodedTypes.bytes),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.3.2", name="element value long"),
|
||||
ProtoParser.ParserFieldDefinition(tag="1.3.3", name="element value short"),
|
||||
ProtoParser.ParserFieldDefinition(tag="", tag_prefixes=["1.5.1", "1.5.3", "1.5.4", "1.5.5", "1.5.6"], name="position"),
|
||||
ProtoParser.ParserFieldDefinition(tag=".1", tag_prefixes=["1.5.1", "1.5.3", "1.5.4", "1.5.5", "1.5.6"], name="latitude", intended_decoding=ProtoParser.DecodedTypes.double), # noqa: E501
|
||||
ProtoParser.ParserFieldDefinition(tag=".2", tag_prefixes=["1.5.1", "1.5.3", "1.5.4", "1.5.5", "1.5.6"], name="longitude", intended_decoding=ProtoParser.DecodedTypes.double), # noqa: E501
|
||||
ProtoParser.ParserFieldDefinition(tag="7", name="app"),
|
||||
]
|
||||
),
|
||||
]
|
||||
|
||||
custom_view_config = ViewConfig(
|
||||
parser_options=ProtoParser.ParserOptions(exclude_message_headers=True, include_wiretype=True)
|
||||
)
|
||||
|
||||
custom_view_config_parser_rules = ViewConfig(
|
||||
parser_rules=custom_parser_rules
|
||||
)
|
||||
|
||||
sim_msg_req = tutils.treq(
|
||||
port=443,
|
||||
host="example.com",
|
||||
path="/ReverseGeocode"
|
||||
)
|
||||
sim_msg_req.headers["grpc-encoding"] = "gzip"
|
||||
sim_msg_resp = tutils.tresp()
|
||||
|
||||
sim_flow = tflow.tflow(
|
||||
req=sim_msg_req,
|
||||
resp=sim_msg_resp
|
||||
)
|
||||
|
||||
|
||||
def test_view_protobuf(tdata):
|
||||
v = full_eval(ViewGrpcProtobuf())
|
||||
p = tdata.path(datadir + "msg1.bin")
|
||||
|
||||
with open(p, "rb") as f:
|
||||
raw = f.read()
|
||||
view_text, output = v(raw)
|
||||
assert view_text == "Protobuf (flattened)"
|
||||
output = list(output) # assure list conversion if generator
|
||||
assert output == [
|
||||
[('text', '[message] '), ('text', ' '), ('text', '1 '), ('text', ' ')],
|
||||
[('text', '[fixed64] '), ('text', ' '), ('text', '1.1 '), ('text', '4630671247600644312 ')],
|
||||
[('text', '[fixed64] '), ('text', ' '), ('text', '1.2 '), ('text', '13858493542095451628 ')],
|
||||
[('text', '[string] '), ('text', ' '), ('text', '3 '), ('text', 'de_DE ')],
|
||||
[('text', '[uint32] '), ('text', ' '), ('text', '6 '), ('text', '1 ')],
|
||||
[('text', '[string] '), ('text', ' '), ('text', '7 '), ('text', 'de.mcdonalds.mcdonaldsinfoapp ')]
|
||||
]
|
||||
with pytest.raises(ValueError, match='not a valid protobuf message'):
|
||||
v(b'foobar')
|
||||
|
||||
|
||||
def test_view_protobuf_custom_parsing_request(tdata):
|
||||
v = full_eval(ViewGrpcProtobuf(custom_view_config_parser_rules))
|
||||
p = tdata.path(datadir + "msg1.bin")
|
||||
with open(p, "rb") as f:
|
||||
raw = f.read()
|
||||
view_text, output = v(raw, flow=sim_flow, http_message=sim_flow.request) # simulate request message
|
||||
assert view_text == "Protobuf (flattened)"
|
||||
output = list(output) # assure list conversion if generator
|
||||
assert output == [
|
||||
[('text', '[message] '), ('text', 'position '), ('text', '1 '), ('text', ' ')],
|
||||
[('text', '[double] '), ('text', 'latitude '), ('text', '1.1 '), ('text', '38.89816675798073 ')],
|
||||
[('text', '[double] '), ('text', 'longitude '), ('text', '1.2 '), ('text', '-77.03829828366696 ')],
|
||||
[('text', '[string] '), ('text', 'country '), ('text', '3 '), ('text', 'de_DE ')],
|
||||
[('text', '[uint32] '), ('text', ' '), ('text', '6 '), ('text', '1 ')],
|
||||
[('text', '[string] '), ('text', 'app '), ('text', '7 '), ('text', 'de.mcdonalds.mcdonaldsinfoapp ')]
|
||||
]
|
||||
|
||||
|
||||
def test_view_protobuf_custom_parsing_response(tdata):
|
||||
v = full_eval(ViewGrpcProtobuf(custom_view_config_parser_rules))
|
||||
p = tdata.path(datadir + "msg3.bin")
|
||||
|
||||
with open(p, "rb") as f:
|
||||
raw = f.read()
|
||||
view_text, output = v(raw, flow=sim_flow, http_message=sim_flow.response) # simulate response message
|
||||
assert view_text == "Protobuf (flattened)"
|
||||
output = list(output) # assure list conversion if generator
|
||||
assert output == [
|
||||
[('text', '[message] '), ('text', ' '), ('text', '1 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[string] '), ('text', ' '), ('text', '1.1 '), ('text', '\x15 ')], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'address '), ('text', '1.2 '), ('text', '1650 Pennsylvania Avenue NW, Washington, DC 20502, USA ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'address array element '), ('text', '1.3 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[bytes] '), ('text', 'unknown bytes '), ('text', '1.3.1 '), ('text', 'b\'"\' ')], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value long '), ('text', '1.3.2 '), ('text', '1650 ')], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value short '), ('text', '1.3.3 '), ('text', '1650 ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'address array element '), ('text', '1.3 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[bytes] '), ('text', 'unknown bytes '), ('text', '1.3.1 '), ('text', "b'\\x02' ")], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value long '), ('text', '1.3.2 '), ('text', 'Pennsylvania Avenue Northwest ')], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value short '), ('text', '1.3.3 '), ('text', 'Pennsylvania Avenue NW ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'address array element '), ('text', '1.3 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[bytes] '), ('text', 'unknown bytes '), ('text', '1.3.1 '), ('text', "b'\\x14\\x04' ")], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value long '), ('text', '1.3.2 '), ('text', 'Northwest Washington ')], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value short '), ('text', '1.3.3 '), ('text', 'Northwest Washington ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'address array element '), ('text', '1.3 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[bytes] '), ('text', 'unknown bytes '), ('text', '1.3.1 '), ('text', "b'\\x0c\\x04' ")], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value long '), ('text', '1.3.2 '), ('text', 'Washington ')], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value short '), ('text', '1.3.3 '), ('text', 'Washington ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'address array element '), ('text', '1.3 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[bytes] '), ('text', 'unknown bytes '), ('text', '1.3.1 '), ('text', "b'\\x06\\x04' ")], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value long '), ('text', '1.3.2 '), ('text', 'District of Columbia ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'element value short '), ('text', '1.3.3 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'address array element '), ('text', '1.3 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[bytes] '), ('text', 'unknown bytes '), ('text', '1.3.1 '), ('text', "b'\\x05\\x04' ")], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value long '), ('text', '1.3.2 '), ('text', 'USA ')], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value short '), ('text', '1.3.3 '), ('text', 'US ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'address array element '), ('text', '1.3 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[bytes] '), ('text', 'unknown bytes '), ('text', '1.3.1 '), ('text', "b'\\x17' ")], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value long '), ('text', '1.3.2 '), ('text', '20502 ')], # noqa: E501
|
||||
[('text', '[string] '), ('text', 'element value short '), ('text', '1.3.3 '), ('text', '20502 ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', ' '), ('text', '1.5 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'position '), ('text', '1.5.1 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[double] '), ('text', 'latitude '), ('text', '1.5.1.1 '), ('text', '38.8970309 ')], # noqa: E501
|
||||
[('text', '[double] '), ('text', 'longitude '), ('text', '1.5.1.2 '), ('text', '-77.03872559999999 ')], # noqa: E501
|
||||
[('text', '[uint32] '), ('text', ' '), ('text', '1.5.2 '), ('text', '1 ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'position '), ('text', '1.5.3 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[double] '), ('text', 'latitude '), ('text', '1.5.3.1 '), ('text', '38.8962271697085 ')], # noqa: E501
|
||||
[('text', '[double] '), ('text', 'longitude '), ('text', '1.5.3.2 '), ('text', '-77.0400511802915 ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'position '), ('text', '1.5.4 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[double] '), ('text', 'latitude '), ('text', '1.5.4.1 '), ('text', '38.8989251302915 ')], # noqa: E501
|
||||
[('text', '[double] '), ('text', 'longitude '), ('text', '1.5.4.2 '), ('text', '-77.03735321970849 ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'position '), ('text', '1.5.5 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[double] '), ('text', 'latitude '), ('text', '1.5.5.1 '), ('text', '38.896898 ')], # noqa: E501
|
||||
[('text', '[double] '), ('text', 'longitude '), ('text', '1.5.5.2 '), ('text', '-77.03917229999999 ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', 'position '), ('text', '1.5.6 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[double] '), ('text', 'latitude '), ('text', '1.5.6.1 '), ('text', '38.8982543 ')], # noqa: E501
|
||||
[('text', '[double] '), ('text', 'longitude '), ('text', '1.5.6.2 '), ('text', '-77.0382321 ')], # noqa: E501
|
||||
[('text', '[string] '), ('text', ' '), ('text', '1.7 '), ('text', 'ChIJAXiAory3t4kRpkrvas9dYmQ ')], # noqa: E501
|
||||
[('text', '[message] '), ('text', ' '), ('text', '2 '), ('text', ' ')], # noqa: E501
|
||||
[('text', '[uint32] '), ('text', ' '), ('text', '2.1 '), ('text', '21 ')], # noqa: E501
|
||||
]
|
||||
|
||||
|
||||
def test_view_protobuf_custom_config(tdata):
|
||||
v = full_eval(ViewGrpcProtobuf(custom_view_config))
|
||||
p = tdata.path(datadir + "msg1.bin")
|
||||
|
||||
with open(p, "rb") as f:
|
||||
raw = f.read()
|
||||
view_text, output = v(raw)
|
||||
assert view_text == "Protobuf (flattened)"
|
||||
output = list(output) # assure list conversion if generator
|
||||
assert output == [
|
||||
[('text', '[bit_64->fixed64] '), ('text', ' '), ('text', '1.1 '), ('text', '4630671247600644312 ')],
|
||||
[('text', '[bit_64->fixed64] '), ('text', ' '), ('text', '1.2 '), ('text', '13858493542095451628 ')],
|
||||
[('text', '[len_delimited->string] '), ('text', ' '), ('text', '3 '), ('text', 'de_DE ')],
|
||||
[('text', '[varint->uint32] '), ('text', ' '), ('text', '6 '), ('text', '1 ')],
|
||||
[('text', '[len_delimited->string] '), ('text', ' '), ('text', '7 '), ('text', 'de.mcdonalds.mcdonaldsinfoapp ')]
|
||||
]
|
||||
|
||||
|
||||
def test_view_grpc(tdata):
|
||||
v = full_eval(ViewGrpcProtobuf())
|
||||
p = tdata.path(datadir + "msg1.bin")
|
||||
|
||||
with open(p, "rb") as f:
|
||||
raw = f.read()
|
||||
# pack into protobuf message
|
||||
raw = helper_pack_grpc_message(raw)
|
||||
|
||||
view_text, output = v(raw, content_type="application/grpc", http_message=sim_msg_req)
|
||||
assert view_text == "gRPC"
|
||||
output = list(output) # assure list conversion if generator
|
||||
|
||||
assert output == [
|
||||
[('text', 'gRPC message 0 (compressed False)')],
|
||||
[('text', '[message] '), ('text', ' '), ('text', '1 '), ('text', ' ')],
|
||||
[('text', '[fixed64] '), ('text', ' '), ('text', '1.1 '), ('text', '4630671247600644312 ')],
|
||||
[('text', '[fixed64] '), ('text', ' '), ('text', '1.2 '), ('text', '13858493542095451628 ')],
|
||||
[('text', '[string] '), ('text', ' '), ('text', '3 '), ('text', 'de_DE ')],
|
||||
[('text', '[uint32] '), ('text', ' '), ('text', '6 '), ('text', '1 ')],
|
||||
[('text', '[string] '), ('text', ' '), ('text', '7 '), ('text', 'de.mcdonalds.mcdonaldsinfoapp ')]
|
||||
]
|
||||
with pytest.raises(ValueError, match='invalid gRPC message'):
|
||||
v(b'foobar', content_type="application/grpc")
|
||||
with pytest.raises(ValueError, match='Failed to decompress gRPC message with gzip'):
|
||||
list(parse_grpc_messages(data=b'\x01\x00\x00\x00\x01foobar', compression_scheme="gzip"))
|
||||
|
||||
|
||||
def test_view_grpc_compressed(tdata):
|
||||
v = full_eval(grpc.ViewGrpcProtobuf())
|
||||
p = tdata.path(datadir + "msg1.bin")
|
||||
|
||||
with open(p, "rb") as f:
|
||||
raw = f.read()
|
||||
# pack into protobuf message
|
||||
raw = helper_pack_grpc_message(raw, True, "gzip")
|
||||
|
||||
view_text, output = v(raw, content_type="application/grpc")
|
||||
assert view_text == "gRPC"
|
||||
output = list(output) # assure list conversion if generator
|
||||
|
||||
assert output == [
|
||||
[('text', 'gRPC message 0 (compressed gzip)')],
|
||||
[('text', '[message] '), ('text', ' '), ('text', '1 '), ('text', ' ')],
|
||||
[('text', '[fixed64] '), ('text', ' '), ('text', '1.1 '), ('text', '4630671247600644312 ')],
|
||||
[('text', '[fixed64] '), ('text', ' '), ('text', '1.2 '), ('text', '13858493542095451628 ')],
|
||||
[('text', '[string] '), ('text', ' '), ('text', '3 '), ('text', 'de_DE ')],
|
||||
[('text', '[uint32] '), ('text', ' '), ('text', '6 '), ('text', '1 ')],
|
||||
[('text', '[string] '), ('text', ' '), ('text', '7 '), ('text', 'de.mcdonalds.mcdonaldsinfoapp ')]
|
||||
]
|
||||
|
||||
|
||||
def helper_encode_base128le(val: int):
|
||||
# hacky base128le encoding
|
||||
if val <= 0:
|
||||
return b'\x00'
|
||||
res = []
|
||||
while val > 0:
|
||||
part = val & 0b1111111
|
||||
val = val >> 7
|
||||
if val > 0:
|
||||
res.append(part + 0x80)
|
||||
else:
|
||||
res.append(part)
|
||||
return bytes(res)
|
||||
|
||||
|
||||
def helper_gen_varint_msg_field(f_idx: int, f_val: int):
|
||||
# manual encoding of protobuf data
|
||||
f_wt = 0 # field type 0 (varint)
|
||||
tag = (f_idx << 3) | f_wt # combined tag
|
||||
msg = helper_encode_base128le(tag) # add encoded tag to message
|
||||
msg = msg + helper_encode_base128le(f_val) # add varint encoded field value
|
||||
return msg
|
||||
|
||||
|
||||
def helper_gen_bits32_msg_field(f_idx: int, f_val: int):
|
||||
# manual encoding of protobuf data
|
||||
f_wt = 5 # field type 5 (bits32)
|
||||
tag = (f_idx << 3) | f_wt # combined tag
|
||||
msg = helper_encode_base128le(tag) # add encoded tag to message
|
||||
msg = msg + struct.pack("<I", f_val) # add varint encoded field value
|
||||
return msg
|
||||
|
||||
|
||||
def helper_gen_bits64_msg_field(f_idx: int, f_val: int):
|
||||
# manual encoding of protobuf data
|
||||
f_wt = 1 # field type 1 (bits32)
|
||||
tag = (f_idx << 3) | f_wt # combined tag
|
||||
msg = helper_encode_base128le(tag) # add encoded tag to message
|
||||
msg = msg + struct.pack("<Q", f_val) # add varint encoded field value
|
||||
return msg
|
||||
|
||||
|
||||
def helper_gen_lendel_msg_field(f_idx: int, f_val: bytes):
|
||||
# manual encoding of protobuf data
|
||||
f_wt = 2 # field type 2 (length delimited messag)
|
||||
tag = (f_idx << 3) | f_wt # combined tag
|
||||
msg = helper_encode_base128le(tag) # add encoded tag to message
|
||||
msg = msg + helper_encode_base128le(len(f_val)) # add length of message
|
||||
msg = msg + f_val
|
||||
return msg
|
||||
|
||||
|
||||
def test_special_decoding():
|
||||
from mitmproxy.contrib.kaitaistruct.google_protobuf import GoogleProtobuf
|
||||
|
||||
msg = helper_gen_varint_msg_field(1, 1) # small varint
|
||||
msg += helper_gen_varint_msg_field(2, 1 << 32) # varint > 32bit
|
||||
msg += helper_gen_varint_msg_field(3, 1 << 64) # varint > 64bit (returned as 0x0 by Kaitai protobuf decoder)
|
||||
msg += helper_gen_bits32_msg_field(4, 0xbf8ccccd) # bits32
|
||||
msg += helper_gen_bits64_msg_field(5, 0xbff199999999999a) # bits64
|
||||
msg += helper_gen_varint_msg_field(6, 0xffffffff) # 32 bit varint negative
|
||||
msg += helper_gen_lendel_msg_field(7, b"hello world") # length delimted message, UTF-8 parsable
|
||||
|
||||
parser = ProtoParser.Message(
|
||||
data=msg,
|
||||
parent_field=[],
|
||||
rules=[],
|
||||
options=ProtoParser.ParserOptions()
|
||||
)
|
||||
# print(list(parser.gen_string_rows()))
|
||||
# return
|
||||
|
||||
fields = list(parser.gen_fields())
|
||||
assert fields[0].wire_value == 1
|
||||
assert fields[1].wire_value == 1 << 32
|
||||
as_bool = fields[1].decode_as(ProtoParser.DecodedTypes.bool)
|
||||
assert isinstance(as_bool, bool)
|
||||
assert as_bool
|
||||
as_bool = fields[2].decode_as(ProtoParser.DecodedTypes.bool)
|
||||
assert isinstance(as_bool, bool)
|
||||
assert not as_bool
|
||||
assert fields[1].decode_as(ProtoParser.DecodedTypes.float) == 2.121995791e-314
|
||||
assert fields[1].safe_decode_as(ProtoParser.DecodedTypes.uint32) == (ProtoParser.DecodedTypes.uint64, 1 << 32)
|
||||
assert fields[0].safe_decode_as(ProtoParser.DecodedTypes.sfixed32) == (ProtoParser.DecodedTypes.uint32, 1)
|
||||
assert fields[3].wire_type == GoogleProtobuf.Pair.WireTypes.bit_32
|
||||
assert fields[4].wire_type == GoogleProtobuf.Pair.WireTypes.bit_64
|
||||
# signed 32 bit int (standard encoding)
|
||||
assert fields[5].safe_decode_as(ProtoParser.DecodedTypes.int32) == (ProtoParser.DecodedTypes.int32, -1)
|
||||
# fixed (signed) 32bit int (ZigZag encoding)
|
||||
assert fields[5].safe_decode_as(ProtoParser.DecodedTypes.sint32) == (ProtoParser.DecodedTypes.sint32, -2147483648)
|
||||
# sint64
|
||||
assert fields[1].safe_decode_as(ProtoParser.DecodedTypes.sint64) == (ProtoParser.DecodedTypes.sint64, 2147483648)
|
||||
|
||||
# varint 64bit to enum
|
||||
assert fields[1].safe_decode_as(ProtoParser.DecodedTypes.enum) == (ProtoParser.DecodedTypes.enum, 4294967296)
|
||||
|
||||
# bits64 to sfixed64
|
||||
assert fields[4].safe_decode_as(ProtoParser.DecodedTypes.sfixed64) == (ProtoParser.DecodedTypes.sfixed64, -4615739258092021350)
|
||||
# bits64 to fixed64
|
||||
assert fields[4].safe_decode_as(ProtoParser.DecodedTypes.fixed64) == (ProtoParser.DecodedTypes.fixed64, 0xbff199999999999a)
|
||||
# bits64 to double
|
||||
assert fields[4].safe_decode_as(ProtoParser.DecodedTypes.double) == (ProtoParser.DecodedTypes.double, -1.1)
|
||||
# bits64 to float --> failover fixed64 (64bit to large for double)
|
||||
assert fields[4].safe_decode_as(ProtoParser.DecodedTypes.float) == (ProtoParser.DecodedTypes.fixed64, 0xbff199999999999a)
|
||||
|
||||
# bits32 to sfixed32
|
||||
assert fields[3].safe_decode_as(ProtoParser.DecodedTypes.sfixed32) == (ProtoParser.DecodedTypes.sfixed32, -1081291571)
|
||||
# bits32 to fixed32
|
||||
assert fields[3].safe_decode_as(ProtoParser.DecodedTypes.fixed32) == (ProtoParser.DecodedTypes.fixed32, 0xbf8ccccd)
|
||||
# bits32 to float
|
||||
assert fields[3].safe_decode_as(ProtoParser.DecodedTypes.float) == (ProtoParser.DecodedTypes.float, -1.100000023841858)
|
||||
# bits32 to string --> failover fixed32
|
||||
assert fields[3].safe_decode_as(ProtoParser.DecodedTypes.string) == (ProtoParser.DecodedTypes.fixed32, 0xbf8ccccd)
|
||||
|
||||
# length delimeted to string
|
||||
assert fields[6].safe_decode_as(ProtoParser.DecodedTypes.string) == (ProtoParser.DecodedTypes.string, "hello world")
|
||||
# length delimeted to bytes
|
||||
assert fields[6].safe_decode_as(ProtoParser.DecodedTypes.bytes) == (ProtoParser.DecodedTypes.bytes, b"hello world")
|
||||
|
||||
assert fields[0].wire_value_as_utf8() == "1"
|
||||
|
||||
with pytest.raises(TypeError, match="intended decoding mismatches wire type"):
|
||||
fields[0].decode_as(ProtoParser.DecodedTypes.sfixed32)
|
||||
with pytest.raises(TypeError, match="wire value too large for int32"):
|
||||
fields[1].decode_as(ProtoParser.DecodedTypes.int32)
|
||||
with pytest.raises(TypeError, match="wire value too large for sint32"):
|
||||
fields[1].decode_as(ProtoParser.DecodedTypes.sint32)
|
||||
with pytest.raises(TypeError, match="wire value too large for uint32"):
|
||||
fields[1].decode_as(ProtoParser.DecodedTypes.uint32)
|
||||
with pytest.raises(TypeError, match="can not be converted to floatingpoint representation"):
|
||||
fields[6]._wire_value_as_float()
|
||||
|
||||
print(fields[6])
|
||||
|
||||
|
||||
def test_render_priority():
|
||||
v = grpc.ViewGrpcProtobuf()
|
||||
assert v.render_priority(b"data", content_type="application/x-protobuf")
|
||||
assert v.render_priority(b"data", content_type="application/x-protobuffer")
|
||||
assert v.render_priority(b"data", content_type="application/grpc-proto")
|
||||
assert v.render_priority(b"data", content_type="application/grpc")
|
||||
assert not v.render_priority(b"data", content_type="text/plain")
|
2
test/mitmproxy/contentviews/test_grpc_data/msg1.bin
Normal file
2
test/mitmproxy/contentviews/test_grpc_data/msg1.bin
Normal file
@ -0,0 +1,2 @@
|
||||
|
||||
ØðÙ ÷rC@ìõ¤zsBSÀde_DE0:de.mcdonalds.mcdonaldsinfoapp
|
BIN
test/mitmproxy/contentviews/test_grpc_data/msg2.bin
Normal file
BIN
test/mitmproxy/contentviews/test_grpc_data/msg2.bin
Normal file
Binary file not shown.
13
test/mitmproxy/contentviews/test_grpc_data/msg3.bin
Normal file
13
test/mitmproxy/contentviews/test_grpc_data/msg3.bin
Normal file
@ -0,0 +1,13 @@
|
||||
|
||||
Ÿ
|
||||
61650 Pennsylvania Avenue NW, Washington, DC 20502, USA
|
||||
"16501650:
|
||||
Pennsylvania Avenue NorthwestPennsylvania Avenue NW0
|
||||
Northwest WashingtonNorthwest Washington
|
||||
|
||||
Washington
|
||||
Washington
|
||||
District of ColumbiaDC
|
||||
USAUS
|
||||
2050220502*f
|
||||
/€•èÑrC@&aðzzBSÀ £×g’·rC@*aÓ2<C393>BSÀ" ÈŠúsC@CÂþcBSÀ* …ì¼<C3AC>ÍrC@.ÚˆÌ<CB86>BSÀ2 åþ4ÿùrC@ÊerBSÀ:ChIJAXiAory3t4kRpkrvas9dYmQ
|
@ -14,6 +14,7 @@ export interface OptionsState {
|
||||
ciphers_server: string | undefined
|
||||
client_certs: string | undefined
|
||||
client_replay: string[]
|
||||
client_replay_concurrency: number
|
||||
command_history: boolean
|
||||
confdir: string
|
||||
connection_strategy: string
|
||||
@ -99,6 +100,7 @@ export const defaultState: OptionsState = {
|
||||
ciphers_server: undefined,
|
||||
client_certs: undefined,
|
||||
client_replay: [],
|
||||
client_replay_concurrency: 1,
|
||||
command_history: true,
|
||||
confdir: "~/.mitmproxy",
|
||||
connection_strategy: "eager",
|
||||
|
Loading…
Reference in New Issue
Block a user