Updated setup.py and moved requirements to examples section. Included examples section in requirements.txt. Updated har_extractor to use command line arguments.

This commit is contained in:
Justus Wingert 2014-11-15 22:37:32 +01:00
parent a7ab06d80e
commit c84ad384f6
3 changed files with 27 additions and 17 deletions

View File

@ -56,9 +56,14 @@ def start(context, argv):
of HAR generation. As it will probably be necessary to cluster logs by IPs or reset them of HAR generation. As it will probably be necessary to cluster logs by IPs or reset them
from time to time. from time to time.
""" """
context.dump_file = None
if len(argv) > 1:
context.dump_file = argv[1]
else:
raise ValueError('Usage: -s "har_extractor.py filename" '
'(- will output to stdout, filenames ending with .zhar will result in compressed har)')
context.HARLog = _HARLog(['https://github.com']) context.HARLog = _HARLog(['https://github.com'])
context.seen_server_connect = set() context.seen_server = set()
context.seen_server_ssl = set()
def response(context, flow): def response(context, flow):
@ -66,22 +71,20 @@ def response(context, flow):
Called when a server response has been received. At the time of this message both Called when a server response has been received. At the time of this message both
a request and a response are present and completely done. a request and a response are present and completely done.
""" """
# Values are converted from float seconds to int milliseconds later.
ssl_time = -.001
connect_time = -.001 connect_time = -.001
if flow.server_conn not in context.seen_server_connect: if flow.server_conn not in context.seen_server:
# Calculate the connect_time for this server_conn. Afterwards add it to seen list, in # Calculate the connect_time for this server_conn. Afterwards add it to seen list, in
# order to avoid the connect_time being present in entries that use an existing connection. # order to avoid the connect_time being present in entries that use an existing connection.
connect_time = flow.server_conn.timestamp_tcp_setup - flow.server_conn.timestamp_start connect_time = flow.server_conn.timestamp_tcp_setup - flow.server_conn.timestamp_start
context.seen_server_connect.add(flow.server_conn) context.seen_server.add(flow.server_conn)
ssl_time = -.001 if flow.server_conn.timestamp_ssl_setup is not None:
if flow.server_conn not in context.seen_server_ssl \ # Get the ssl_time for this server_conn as the difference between the start of the successful
and flow.server_conn.timestamp_ssl_setup is not None: # tcp setup and the successful ssl setup. If no ssl setup has been made it is left as -1 since
# Get the ssl_time for this server_conn as the difference between the start of the successful # it doesn't apply to this connection.
# tcp setup and the successful ssl setup. Afterwards add it to seen list, in order to avoid ssl_time = flow.server_conn.timestamp_ssl_setup - flow.server_conn.timestamp_tcp_setup
# the ssl_time being present in entries that use an existing connection. If no ssl setup has
# been made it is also left as -1 since it doesn't apply to this connection.
ssl_time = flow.server_conn.timestamp_ssl_setup - flow.server_conn.timestamp_tcp_setup
context.seen_server_ssl.add(flow.server_conn)
# Calculate the raw timings from the different timestamps present in the request and response object. # Calculate the raw timings from the different timestamps present in the request and response object.
# For lack of a way to measure it dns timings can not be calculated. The same goes for HAR blocked: # For lack of a way to measure it dns timings can not be calculated. The same goes for HAR blocked:
@ -178,7 +181,12 @@ def done(context):
compressed_json_dump = context.HARLog.compress() compressed_json_dump = context.HARLog.compress()
print "=" * 100 print "=" * 100
pprint(json.loads(json_dump)) if context.dump_file == '-':
pprint(json.loads(json_dump))
elif context.dump_file.endswith('.zhar'):
file(context.dump_file, "w").write(compressed_json_dump)
else:
file(context.dump_file, "w").write(json_dump)
print "=" * 100 print "=" * 100
print "HAR log finished with %s bytes (%s bytes compressed)" % (len(json_dump), len(compressed_json_dump)) print "HAR log finished with %s bytes (%s bytes compressed)" % (len(json_dump), len(compressed_json_dump))
print "Compression rate is %s%%" % str(100. * len(compressed_json_dump) / len(json_dump)) print "Compression rate is %s%%" % str(100. * len(compressed_json_dump) / len(json_dump))

View File

@ -1,3 +1,3 @@
-e git+https://github.com/mitmproxy/netlib.git#egg=netlib -e git+https://github.com/mitmproxy/netlib.git#egg=netlib
-e git+https://github.com/mitmproxy/pathod.git#egg=pathod -e git+https://github.com/mitmproxy/pathod.git#egg=pathod
-e .[dev] -e .[dev,examples]

View File

@ -28,8 +28,6 @@ script_deps = {
"urwid>=1.1", "urwid>=1.1",
"lxml>=3.3.6", "lxml>=3.3.6",
"Pillow>=2.3.0", "Pillow>=2.3.0",
"pytz",
"harparser",
}, },
"mitmdump": set() "mitmdump": set()
} }
@ -80,6 +78,10 @@ setup(
"pyamf>=0.6.1", "pyamf>=0.6.1",
"protobuf>=2.5.0", "protobuf>=2.5.0",
"cssutils>=1.0" "cssutils>=1.0"
],
'examples': [
"pytz",
"harparser",
] ]
} }
) )