forked from DidierStevens/DidierStevensSuite
-
Notifications
You must be signed in to change notification settings - Fork 0
/
base64dump.py
1630 lines (1450 loc) · 71.1 KB
/
base64dump.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
#!/usr/bin/env python
__description__ = 'Extract base64 strings from file'
__author__ = 'Didier Stevens'
__version__ = '0.0.25'
__date__ = '2024/04/16'
"""
Source code put in public domain by Didier Stevens, no Copyright
https://DidierStevens.com
Use at your own risk
History:
2015/06/30: start
2015/07/01: added header
2015/07/14: 0.0.2: added option -n
2015/07/28: fixed option -n
2015/09/12: 0.0.3: added option -c
2015/11/18: 0.0.4 added support for -c :-number
2016/01/22: added option ignorewhitespace
2016/01/23: updated CutData
2016/11/08: 0.0.5 added option -e; unique bytes; option -S
2016/11/18: added hex encoding
2017/02/12: 0.0.6 added encoding all and option -u
2017/02/13: updated man
2017/07/01: 0.0.7 added option -z
2017/10/21: 0.0.8 added option -t
2018/05/07: 0.0.9: added bx and ah encoding; added YARA support; added decoders
2018/05/23: 0.0.10: added zxle and zxbe encoding; added option --ignore
2018/07/23: 0.0.11: added option -I
2020/07/02: 0.0.12: added zxc encoding, verbose YARACompile, updated CutData, option -A, selection warning, DSS_DEFAULT_HASH_ALGORITHMS, option --jsonoutput, option -T, option -p
2020/12/25: 0.0.13 added dec encoding; Translate refactoring
2021/05/23: 0.0.14 added nb decoding
2021/07/16: 0.0.15 bug fix -i -I options; man page changes
2021/07/16: 0.0.16 added b85 decoding
2021/09/27: 0.0.17 added select L; added a85 decoding
2021/11/14: 0.0.18 fixed DecodeDataDecimal
2021/12/15: 0.0.19 Python 3 bugfix
2021/12/28: 0.0.20 added zxcn decoding
2022/05/11: 0.0.21 added nbl decoding
2022/06/17: 0.0.22 added statistics for encodings
2022/07/15: 0.0.23 added option --jsoninput & --postprocess
2022/07/17: continue
2022/10/13: 0.0.24 update CalculateByteStatistics
2024/04/16: 0.0.25 added selection of original; ExtractLongestString; --sort
Todo:
add base64 url
"""
import optparse
import sys
import os
import zipfile
import binascii
import textwrap
import re
import hashlib
import string
import math
import string
import codecs
import zlib
import json
import operator
import base64
try:
import yara
except:
pass
if sys.version_info[0] >= 3:
from io import StringIO
else:
from cStringIO import StringIO
if sys.version_info[0] >= 3:
from io import BytesIO as DataIO
else:
from cStringIO import StringIO as DataIO
dumplinelength = 16
MALWARE_PASSWORD = b'infected'
REGEX_STANDARD = b'[\x09\x20-\x7E]'
global dEncodings
def PrintManual():
manual = '''
Manual:
base64dump is a program that extracts and decodes base64 strings (or other encodings) found inside the provided file. base64dump looks for sequences of base64 characters (or other encodings) in the provided file and tries to decode them.
Other encodings than base64 can be used via option -e.
The result is displayed in a table like this (base64 encoding example):
ID Size Encoded Decoded MD5 decoded
-- ---- ------- ------- -----------
1: 400728 TVqQAAMAAAAEAAAA MZ.............. d611941e0d24cb6b59c7b6b2add4fd8f
2: 36 U2NyaXB0aW5nLkZp Scripting.FileSy a1c46f599699a442a5ae0454467f6d63
3: 4 exel {.. f1b1127ffb842243f9a03e67036d4bb6
The first column (ID) is the number (ID) assigned to the datastream by base64dump. This ID is used when selecting a datastream for further analysis with option -s.
The second column (Size) is the length of the base64 string.
The third column (Encoded) is the start of the base64 string.
The fourth column (Decoded) is the ASCII dump of the start of the decoded base64 string.
The fifth column (MD5 decoded) is the MD5 hash of the decoded base64 string.
By default, it's the MD5 hash, but this can be changed by setting environment variable DSS_DEFAULT_HASH_ALGORITHMS.
Like this: set DSS_DEFAULT_HASH_ALGORITHMS=sha256
By default, base64dump will search for base64 encoding strings. It's possible to specify other encodings by using option -e. This option takes the following values:
''' + '\n'.join(dEncodings.keys()) + '''
b64 stands for BASE64, it looks like this: TVqQAAMAAAAEAAAA...
bu stands for "backslash UNICODE" (\\u), it looks like this: \\u9090\\ueb77...
pu stands for "percent UNICODE" (%u), it looks like this: %u9090%ueb77...
hex stands for "hexadecimal", it looks like this: 6D6573736167652C...
bx stands for "backslash hexadecimal" (\\x), it looks like this: \\x90\\x90...
ah stands for "ampersand hexadecimal" (&H), it looks like this: &H90&H90...
zxle stands for "zero hexadecimal little-endian" (0x), it looks like this: 0x909090900xeb77...
zxbe stands for "zero hexadecimal big-endian" (0x), it looks like this: 0x909090900x77eb...
zxc stands for "zero hexadecimal comma" (0x), it looks like this: 0x90,0x90,0x90,0x90...
zxcn stands for "zero hexadecimal comma no-leading-zero" (0x), it looks like this: 0x90,0xA,0x2A,0x90...
dec stands for "decimal", it looks like this: 80;75;3;4...
nb stands for "NETBIOS", it looks like this: ENFKOIAA
nbl stands for "NETBIOS lowercase", it looks like this: enfkoiaa
b85 stands for BASE85 RFC 1924, it looks like this: X>D+Ca&#bLba`-Pb31o...
a85 stands for ASCII85, it looks like this: BOu!rD]...
zxle and zxbe encoding looks for 1 to 8 hexadecimal characters after the prefix 0x. If the number of hexadecimal characters is uneven, a 0 (digit zero) will be prefixed to have an even number of hexadecimal digits.
Select a datastream for further analysis with option -s followed by the ID number of the datastream, L for the largest one, or a for all.
Output example for -s 2:
Info:
MD5: d611941e0d24cb6b59c7b6b2add4fd8f
Filesize: 300544
Entropy: 6.900531
Unique bytes: 256 (100.00%)
Magic HEX: 4d5a9000
Magic ASCII: MZ..
Null bytes: 41139
Control bytes: 38024
Whitespace bytes: 9369
Printable bytes: 122967
High bytes: 89045
This displays information for the datastream, like the entropy of the datastream.
Append o (original) to the ID number to select the unprocessed decoded data (when using option -P --postprocess). Like this: -s32o.
The selected stream can be dumped (-d), hexdumped (-x), ASCII dumped (-a) or dump the strings (-S). Use the dump option (-d) to extract the stream and save it to disk (with file redirection >) or to pipe it (|) into the next command.
If the dump needs to be processed by a string codec, like utf16, use option -t instead of -d and provide the codec:
C:\Demo>base64dump.py -s 1 -t utf16 test.bin
You can also provide a Python string expression, like .decode('utf16').encode('utf8').
Here is an example of an ascii dump (-s 2 -a):
00000000: 53 63 72 69 70 74 69 6E 67 2E 46 69 6C 65 53 79 Scripting.FileSy
00000010: 73 74 65 6D 4F 62 6A 65 63 74 stemObject
You can also specify the minimum length of the decoded base64 datastream with option -n.
With option -w (ignorewhitespace), you can instruct base64dump to ignore all whitespace characters. So for example if the base64 text is split into lines, then you will get one base64 stream.
With option -z (ignorenullbytes), you can instruct base64dump to ignore all leading 0x00 bytes. This can help to decode UNICODE text.
With option -i (ignore), you can instruct base64dump to ignore all characters you provide as a value to option -i. For example, with -i , you can parse strings like 0x44332211,0x88776655,... with encoding zxle by ignoring the separator character ,.
Option -I is like option -i, except that you specify the characters to ignore as an hexadecimal string. For example, -I 2226 will ignore characters " (22) and & (26).
It's also possible to try all encodings: all
Example:
base64dump.py -e all -n 80 sample.vir
Enc Size Chars Encoded Decoded md5 decoded
--- ---- ----- ------- ------- -----------
b64: 176 11 bebafeca41414141 m..}...^5.^5.^5. e3bed37dcd137c9bb5da103d3c45be49
hex: 176 11 bebafeca41414141 ....AAAAAAAAAAAA 56464bd2b2c42bbf2edda87d54ab91f5
a85: 178 13 <bebafeca4141414 Vj...4V.2| .;... 9385715088c31ac9a8b4fa06166ccb97
b85: 178 13 <bebafeca4141414 ......Q..B...{.. 1b14eef572cc16916b30e32c490f2cc9
b64: 192 13 28000e1358000e13 ..4..w..4..wm..} 81c587874b1c3ddc5479a283179d29f7
hex: 192 13 28000e1358000e13 (...X.......AAAA 3a604ca304d1dbbcc1734a430cf6dc82
a85: 194 15 <28000e1358000e1 T...1*..H.(`2x9. b714fd16ad0d77d5e5391d9197ad1b9f
b85: 194 15 <28000e1358000e1 .5v..v.E..%..<." d3c57839bf4900e64ea17f035382bb52
b64: 2144 16 6064a1000000008b .N.k]4.M4.O..M8. 46d100d435a37f89f3ab0ac6db3e9cac
hex: 2144 16 6064a1000000008b `d......@.%....f a2ae7b55955262ada177862ceb683977
a85: 2146 18 <6064a1000000008 T.v....q/:..I.#Y 76f1028223f9196c18ea2a0f64f304ef
b85: 2146 18 <6064a1000000008 .Z..p........?N. a890225330f60c170e2c588fa6e82f90
b64: 3640 16 48895C2408488974 ..=.-...<.....t. f0384b0c74e9402ab3f1aacc4a270ed3
hex: 3640 16 48895C2408488974 H.\$.H.t$.H.|$.U 3320b9e508862886d2c3f556cacc67ec
a85: 3642 18 <48895C240848897 T...?z.uHD.XE'X. 053a28dd24051ea7b2131a95c34b99ee
b85: 3642 18 <48895C240848897 .H6...DW......VR 5f57fa2b2288577e7844e28d6e9548ca
b64: 213024 22 5555555566666666 ..y..y........Z. c7ff5e1a5a01698f99f67fd3f0d20d6d
hex: 213024 22 5555555566666666 UUUUffffMZ...... 7a618dc1ef3f52693b8ec22dbe0300ec
a85: 213026 24 <555555556666666 T..[>...B...B... 5784f2d34e1881202f88a39dc2d64ffe
b85: 213026 24 <555555556666666 .Q?............. d227c899ea11f605aa1b78700cbc5f71
b64: 246816 22 5555555566666666 ..y..y........Z. bdec177dc760d4296aefbbdc4c47bcf2
hex: 246816 22 5555555566666666 UUUUffffMZ...... a69fa966087f38241544523d437f9a8b
a85: 246818 24 <555555556666666 T..[>...B...B... 1b925074b51c31454e68f50db73ea4a7
b85: 246818 24 <555555556666666 .Q?............. 80d002c19cc64ff2ab61c0cde01e6ddf
The list is sorted by increasing size.
Column Chars contains the number of unique characters found in the encoding.
The reason that sometimes 22 unique characters are found for the hex encoding in this example, and not 16 at most, is that this sample both uses lowercase and uppercase letters for hexadecimal.
Identical decoded content can be made unique with option -u.
base64dump can scan the content of the datastreams with YARA rules (the YARA Python module must be installed). You provide the YARA rules with option -y. You can provide one file with YARA rules, an at-file (@file containing the filenames of the YARA files) or a directory. In case of a directory, all files inside the directory are read as YARA files. Or you can provide the YARA rule with the option value if it starts with # (literal), #s# (string), #q# (quote), #h# (hexadecimal) or #b# (base64). Example: -y "#rule demo {strings: $a=\"demo\" condition: $a}"
Using #s#demo will instruct base64dump to generate a rule to search for string demo (rule string {strings: $a = "demo" ascii wide nocase condition: $a) and use that rule.
All datastreams are scanned with the provided YARA rules, you can not use option -s to select an individual datastream.
Example:
C:\Demo>base64dump.py -y contains_pe_file.yara malware.zip
Enc Size Encoded Decoded MD5 decoded
--- ---- ------- ------- -----------
ah: 256000 &H4d&H5a&H90&H00 MZ.............. 5cd40560a53fda5b32c35adfcdfca3e1
YARA rule: Contains_PE_File
In this example, you use YARA rule contains_pe_file.yara to find PE files (executables) in the decoded data.
If you want more information about what was detected by the YARA rule, use option --yarastrings like in this example:
C:\Demo>base64dump.py -y contains_pe_file.yara --yarastrings malware.zip
Enc Size Encoded Decoded MD5 decoded
--- ---- ------- ------- -----------
ah: 256000 &H4d&H5a&H90&H00 MZ.............. 5cd40560a53fda5b32c35adfcdfca3e1
YARA rule: Contains_PE_File
000000 $a:
4d5a
'MZ'
YARA rule contains_pe_file detects PE files by finding string MZ followed by string PE at the correct offset (AddressOfNewExeHeader).
The rule looks like this:
rule Contains_PE_File
{
meta:
author = "Didier Stevens (https://DidierStevens.com)"
description = "Detect a PE file inside a byte sequence"
method = "Find string MZ followed by string PE at the correct offset (AddressOfNewExeHeader)"
strings:
$a = "MZ"
condition:
for any i in (1..#a): (uint32(@a[i] + uint32(@a[i] + 0x3C)) == 0x00004550)
}
Dump options (-a, -x, -S, ...) can be used together with option -y: triggered rules will cause a dump of the data that was scanned.
When looking for traces of Windows executable code (PE files, shellcode, ...) with YARA rules, one must take into account the fact that the executable code might have been encoded (for example via XOR and a key) to evade detection.
To deal with this possibility, base64dump supports decoders. A decoder is a type of plugin, that will bruteforce a type of encoding on each datastream. For example, decoder_xor1 will encode each datastream via XOR and a key of 1 byte. So effectively, 256 different encodings of the datastream will be scanned by the YARA rules. 256 encodings because: XOR key 0x00, XOR key 0x01, XOR key 0x02, ..., XOR key 0xFF
Here is an example:
C:\Demo>base64dump.py -y contains_pe_file.yara -D decoder_xor1 malware.zip
Enc Size Encoded Decoded MD5 decoded
--- ---- ------- ------- -----------
ah: 256000 &H4d&H5a&H90&H00 MZ.............. 5cd40560a53fda5b32c35adfcdfca3e1
YARA rule (decoder: XOR 1 byte key 0x14): Contains_PE_File
000000 $a:
4d5a
'MZ'
You can specify decoders in exactly the same way as plugins, for example specifying more than one decoder separated by a comma ,.
If decoders are located in a different directory, you could specify it with the --decoderdir option.
Some decoders take options, to be provided with option --decoderoptions.
Some malformed encodings, like BASE64 strings with a number of digits that is not a multiple of 4, will not be detected by base64dump.py.
Such strings are detected using regular expressions, but are then discarded because they don't have the correct length (length of BASE64 strings must be a multiple of 4, length os hexadecimal strings must be a multiple of 2).
If you still need to decode such strings, you can use option -p (--process). This option allows you to specify a Python function (that will be evaluated, so use only trusted input for this option) to modify the detected string before it is decoded).
You can use builtin function L4 (Length 4) for example: function L4 takes a string as input and truncates it if necessary, by removing characters from the end so that the length becomes a multiple of 4.
Example:
C:\Demo>base64dump.py -p L4 malformed_base64.vir
You can also provide your own function, for example via a lambda expression:
C:\Demo>base64dump.py -p "lambda x: x[:-1]" malformed_base64.vir
In this example, the lambda expression will remove one character from the end of the detected string.
Option -P (postprocess) works just like option -p, but on the decoded data.
If an error occurs in the postprocess function, the error AND the encoded data will be ignored.
Built-in function UTF16_ASCII can be used to transform UTF16 to ASCII, and raise an error when the transformed string is not ASCII. This decoded data is then ignored.
This built-in function UTF16_ASCII allows us to search for encoded PowerShell commands that are pure ASCII.
This built-in function ExtractLongestString extracts the longest string (ASCII/UNICODE).
With option -T (--headtail), output can be truncated to the first 10 lines and last 10 lines of output.
With option --jsonoutput base64dump.py will output selected content as a JSON object that can be piped into other tools that support this JSON format.
With option --jsoninput base64dump.py will consume JSON output (produced by oledump.py, for example) and scan all items in the JSON input.
Option --sort can be used to sort base64dump's output (this option is ignored when option -y or -e all is used).
By default, encoded strings are listed in the order they appear in the input data.
Option --sort takes different flags that are separated by commas (,).
Flag e: sort by encoded data
Flag d: sort by decoded data
Flag do: sort by decoded original data
Flag l: sort by length
Flag r: reversed sort.
For example, --sort l,d will sort the output by length of the decoded data, from shortest to longest.
Option -c (--cut) allows for the partial selection of a datastream. Use this option to "cut out" part of the datastream.
The --cut option takes an argument to specify which section of bytes to select from the datastream. This argument is composed of 2 terms separated by a colon (:), like this:
termA:termB
termA and termB can be:
- nothing (an empty string)
- a positive decimal number; example: 10
- an hexadecimal number (to be preceded by 0x); example: 0x10
- a case sensitive string to search for (surrounded by square brackets and single quotes); example: ['MZ']
- an hexadecimal string to search for (surrounded by square brackets); example: [d0cf11e0]
If termA is nothing, then the cut section of bytes starts with the byte at position 0.
If termA is a number, then the cut section of bytes starts with the byte at the position given by the number (first byte has index 0).
If termA is a string to search for, then the cut section of bytes starts with the byte at the position where the string is first found. If the string is not found, the cut is empty (0 bytes).
If termB is nothing, then the cut section of bytes ends with the last byte.
If termB is a number, then the cut section of bytes ends with the byte at the position given by the number (first byte has index 0).
When termB is a number, it can have suffix letter l. This indicates that the number is a length (number of bytes), and not a position.
termB can also be a negative number (decimal or hexademical): in that case the position is counted from the end of the file. For example, :-5 selects the complete file except the last 5 bytes.
If termB is a string to search for, then the cut section of bytes ends with the last byte at the position where the string is first found. If the string is not found, the cut is empty (0 bytes).
No checks are made to assure that the position specified by termA is lower than the position specified by termB. This is left up to the user.
Search string expressions (ASCII and hexadecimal) can be followed by an instance (a number equal to 1 or greater) to indicate which instance needs to be taken. For example, ['ABC']2 will search for the second instance of string 'ABC'. If this instance is not found, then nothing is selected.
Search string expressions (ASCII and hexadecimal) can be followed by an offset (+ or - a number) to add (or substract) an offset to the found instance. For example, ['ABC']+3 will search for the first instance of string 'ABC' and then select the bytes after ABC (+ 3).
Finally, search string expressions (ASCII and hexadecimal) can be followed by an instance and an offset.
Examples:
This argument can be used to dump the first 256 bytes of a PE file located inside the datastream: ['MZ']:0x100l
This argument can be used to dump the OLE file located inside the datastream: [d0cf11e0]:
When this option is not used, the complete datastream is selected.
'''
for line in manual.split('\n'):
print(textwrap.fill(line, 78))
#Convert 2 Bytes If Python 3
def C2BIP3(string):
if sys.version_info[0] > 2:
return bytes([ord(x) for x in string])
else:
return string
# CIC: Call If Callable
def CIC(expression):
if callable(expression):
return expression()
else:
return expression
def P23Ord(value):
if type(value) == int:
return value
else:
return ord(value)
def P23Chr(value):
if type(value) == int:
return chr(value)
else:
return value
# IFF: IF Function
def IFF(expression, valueTrue, valueFalse):
if expression:
return CIC(valueTrue)
else:
return CIC(valueFalse)
def File2String(filename):
try:
f = open(filename, 'rb')
except:
return None
try:
return f.read()
except:
return None
finally:
f.close()
class cDump():
def __init__(self, data, prefix='', offset=0, dumplinelength=16):
self.data = data
self.prefix = prefix
self.offset = offset
self.dumplinelength = dumplinelength
def HexDump(self):
oDumpStream = self.cDumpStream(self.prefix)
hexDump = ''
for i, b in enumerate(self.data):
if i % self.dumplinelength == 0 and hexDump != '':
oDumpStream.Addline(hexDump)
hexDump = ''
hexDump += IFF(hexDump == '', '', ' ') + '%02X' % self.C2IIP2(b)
oDumpStream.Addline(hexDump)
return oDumpStream.Content()
def CombineHexAscii(self, hexDump, asciiDump):
if hexDump == '':
return ''
countSpaces = 3 * (self.dumplinelength - len(asciiDump))
if len(asciiDump) <= self.dumplinelength / 2:
countSpaces += 1
return hexDump + ' ' + (' ' * countSpaces) + asciiDump
def HexAsciiDump(self, rle=False):
oDumpStream = self.cDumpStream(self.prefix)
position = ''
hexDump = ''
asciiDump = ''
previousLine = None
countRLE = 0
for i, b in enumerate(self.data):
b = self.C2IIP2(b)
if i % self.dumplinelength == 0:
if hexDump != '':
line = self.CombineHexAscii(hexDump, asciiDump)
if not rle or line != previousLine:
if countRLE > 0:
oDumpStream.Addline('* %d 0x%02x' % (countRLE, countRLE * self.dumplinelength))
oDumpStream.Addline(position + line)
countRLE = 0
else:
countRLE += 1
previousLine = line
position = '%08X:' % (i + self.offset)
hexDump = ''
asciiDump = ''
if i % self.dumplinelength == self.dumplinelength / 2:
hexDump += ' '
hexDump += ' %02X' % b
asciiDump += IFF(b >= 32 and b < 128, chr(b), '.')
if countRLE > 0:
oDumpStream.Addline('* %d 0x%02x' % (countRLE, countRLE * self.dumplinelength))
oDumpStream.Addline(self.CombineHexAscii(position + hexDump, asciiDump))
return oDumpStream.Content()
def Base64Dump(self, nowhitespace=False):
encoded = binascii.b2a_base64(self.data)
if nowhitespace:
return encoded
oDumpStream = self.cDumpStream(self.prefix)
length = 64
for i in range(0, len(encoded), length):
oDumpStream.Addline(encoded[0+i:length+i])
return oDumpStream.Content()
class cDumpStream():
def __init__(self, prefix=''):
self.oStringIO = StringIO()
self.prefix = prefix
def Addline(self, line):
if line != '':
self.oStringIO.write(self.prefix + line + '\n')
def Content(self):
return self.oStringIO.getvalue()
@staticmethod
def C2IIP2(data):
if sys.version_info[0] > 2:
return data
else:
return ord(data)
def HexDump(data):
return cDump(data, dumplinelength=dumplinelength).HexDump()
def HexAsciiDump(data, rle=False):
return cDump(data, dumplinelength=dumplinelength).HexAsciiDump(rle=rle)
#Fix for http://bugs.python.org/issue11395
def StdoutWriteChunked(data):
if sys.version_info[0] > 2:
if isinstance(data, str):
sys.stdout.write(data)
else:
sys.stdout.buffer.write(data)
else:
while data != '':
sys.stdout.write(data[0:10000])
try:
sys.stdout.flush()
except IOError:
return
data = data[10000:]
def IfWIN32SetBinary(io):
if sys.platform == 'win32':
import msvcrt
msvcrt.setmode(io.fileno(), os.O_BINARY)
def File2Strings(filename):
try:
f = open(filename, 'r')
except:
return None
try:
return map(lambda line:line.rstrip('\n'), f.readlines())
except:
return None
finally:
f.close()
def ProcessAt(argument):
if argument.startswith('@'):
strings = File2Strings(argument[1:])
if strings == None:
raise Exception('Error reading %s' % argument)
else:
return strings
else:
return [argument]
def ExpandFilenameArguments(filenames):
return list(collections.OrderedDict.fromkeys(sum(map(glob.glob, sum(map(ProcessAt, filenames), [])), [])))
def AsciiDump(data):
return ''.join([IFF(P23Ord(b) >= 32 and P23Ord(b) < 127, P23Chr(b), '.') for b in data])
def Magic(data):
magicPrintable = ''
magicHex = ''
for iter in range(4):
if len(data) >= iter + 1:
if P23Ord(data[iter]) >= 0x20 and P23Ord(data[iter]) < 0x7F:
magicPrintable += P23Chr(data[iter])
else:
magicPrintable += '.'
magicHex += '%02x' % P23Ord(data[iter])
return magicPrintable, magicHex
def CalculateByteStatistics(dPrevalence=None, data=None):
longestString = 0
longestBASE64String = 0
longestHEXString = 0
base64digits = b'abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789+/'
hexdigits = b'abcdefABCDEF0123456789'
averageConsecutiveByteDifference = None
if dPrevalence == None:
dPrevalence = {iter: 0 for iter in range(0x100)}
sumDifferences = 0.0
previous = None
if len(data) > 1:
lengthString = 0
lengthBASE64String = 0
lengthHEXString = 0
for byte in data:
dPrevalence[byte] += 1
if previous != None:
sumDifferences += abs(byte - previous)
if byte >= 0x20 and byte < 0x7F:
lengthString += 1
else:
longestString = max(longestString, lengthString)
lengthString = 0
if byte in base64digits:
lengthBASE64String += 1
else:
longestBASE64String = max(longestBASE64String, lengthBASE64String)
lengthBASE64String = 0
if byte in hexdigits:
lengthHEXString += 1
else:
longestHEXString = max(longestHEXString, lengthHEXString)
lengthHEXString = 0
else:
if byte >= 0x20 and byte < 0x7F:
lengthString = 1
if byte in hexdigits:
lengthHEXString = 1
previous = byte
averageConsecutiveByteDifference = sumDifferences /float(len(data)-1)
longestString = max(longestString, lengthString)
longestBASE64String = max(longestBASE64String, lengthBASE64String)
longestHEXString = max(longestHEXString, lengthHEXString)
sumValues = sum(dPrevalence.values())
countNullByte = dPrevalence[0]
countControlBytes = 0
countWhitespaceBytes = 0
countUniqueBytes = 0
for iter in range(1, 0x21):
if chr(iter) in string.whitespace:
countWhitespaceBytes += dPrevalence[iter]
else:
countControlBytes += dPrevalence[iter]
countControlBytes += dPrevalence[0x7F]
countPrintableBytes = 0
for iter in range(0x21, 0x7F):
countPrintableBytes += dPrevalence[iter]
countHighBytes = 0
for iter in range(0x80, 0x100):
countHighBytes += dPrevalence[iter]
countHexadecimalBytes = 0
countBASE64Bytes = 0
for iter in range(0x30, 0x3A):
countHexadecimalBytes += dPrevalence[iter]
countBASE64Bytes += dPrevalence[iter]
for iter in range(0x41, 0x47):
countHexadecimalBytes += dPrevalence[iter]
for iter in range(0x61, 0x67):
countHexadecimalBytes += dPrevalence[iter]
for iter in range(0x41, 0x5B):
countBASE64Bytes += dPrevalence[iter]
for iter in range(0x61, 0x7B):
countBASE64Bytes += dPrevalence[iter]
countBASE64Bytes += dPrevalence[ord('+')] + dPrevalence[ord('/')] + dPrevalence[ord('=')]
entropy = 0.0
for iter in range(0x100):
if dPrevalence[iter] > 0:
prevalence = float(dPrevalence[iter]) / float(sumValues)
entropy += - prevalence * math.log(prevalence, 2)
countUniqueBytes += 1
return sumValues, entropy, countUniqueBytes, countNullByte, countControlBytes, countWhitespaceBytes, countPrintableBytes, countHighBytes, countHexadecimalBytes, countBASE64Bytes, averageConsecutiveByteDifference, longestString, longestHEXString, longestBASE64String
def CalculateFileMetaData(data):
fileSize, entropy, countUniqueBytes, countNullByte, countControlBytes, countWhitespaceBytes, countPrintableBytes, countHighBytes, countHexadecimalBytes, countBASE64Bytes, averageConsecutiveByteDifference, longestString, longestHEXString, longestBASE64String = CalculateByteStatistics(data=data)
magicPrintable, magicHex = Magic(data[0:4])
return CalculateChosenHash(data)[0], magicPrintable, magicHex, fileSize, entropy, countUniqueBytes, countNullByte, countControlBytes, countWhitespaceBytes, countPrintableBytes, countHighBytes, countHexadecimalBytes, countBASE64Bytes, averageConsecutiveByteDifference, longestString, longestHEXString, longestBASE64String
CUTTERM_NOTHING = 0
CUTTERM_POSITION = 1
CUTTERM_FIND = 2
CUTTERM_LENGTH = 3
def Replace(string, dReplacements):
if string in dReplacements:
return dReplacements[string]
else:
return string
def ParseInteger(argument):
sign = 1
if argument.startswith('+'):
argument = argument[1:]
elif argument.startswith('-'):
argument = argument[1:]
sign = -1
if argument.startswith('0x'):
return sign * int(argument[2:], 16)
else:
return sign * int(argument)
def ParseCutTerm(argument):
if argument == '':
return CUTTERM_NOTHING, None, ''
oMatch = re.match(r'\-?0x([0-9a-f]+)', argument, re.I)
if oMatch == None:
oMatch = re.match(r'\-?(\d+)', argument)
else:
value = int(oMatch.group(1), 16)
if argument.startswith('-'):
value = -value
return CUTTERM_POSITION, value, argument[len(oMatch.group(0)):]
if oMatch == None:
oMatch = re.match(r'\[([0-9a-f]+)\](\d+)?([+-](?:0x[0-9a-f]+|\d+))?', argument, re.I)
else:
value = int(oMatch.group(1))
if argument.startswith('-'):
value = -value
return CUTTERM_POSITION, value, argument[len(oMatch.group(0)):]
if oMatch == None:
oMatch = re.match(r"\[u?\'(.+?)\'\](\d+)?([+-](?:0x[0-9a-f]+|\d+))?", argument)
else:
if len(oMatch.group(1)) % 2 == 1:
raise Exception("Uneven length hexadecimal string")
else:
return CUTTERM_FIND, (binascii.a2b_hex(oMatch.group(1)), int(Replace(oMatch.group(2), {None: '1'})), ParseInteger(Replace(oMatch.group(3), {None: '0'}))), argument[len(oMatch.group(0)):]
if oMatch == None:
return None, None, argument
else:
if argument.startswith("[u'"):
# convert ascii to unicode 16 byte sequence
searchtext = oMatch.group(1).decode('unicode_escape').encode('utf16')[2:]
else:
searchtext = oMatch.group(1)
return CUTTERM_FIND, (searchtext, int(Replace(oMatch.group(2), {None: '1'})), ParseInteger(Replace(oMatch.group(3), {None: '0'}))), argument[len(oMatch.group(0)):]
def ParseCutArgument(argument):
type, value, remainder = ParseCutTerm(argument.strip())
if type == CUTTERM_NOTHING:
return CUTTERM_NOTHING, None, CUTTERM_NOTHING, None
elif type == None:
if remainder.startswith(':'):
typeLeft = CUTTERM_NOTHING
valueLeft = None
remainder = remainder[1:]
else:
return None, None, None, None
else:
typeLeft = type
valueLeft = value
if typeLeft == CUTTERM_POSITION and valueLeft < 0:
return None, None, None, None
if typeLeft == CUTTERM_FIND and valueLeft[1] == 0:
return None, None, None, None
if remainder.startswith(':'):
remainder = remainder[1:]
else:
return None, None, None, None
type, value, remainder = ParseCutTerm(remainder)
if type == CUTTERM_POSITION and remainder == 'l':
return typeLeft, valueLeft, CUTTERM_LENGTH, value
elif type == None or remainder != '':
return None, None, None, None
elif type == CUTTERM_FIND and value[1] == 0:
return None, None, None, None
else:
return typeLeft, valueLeft, type, value
def Find(data, value, nth, startposition=-1):
position = startposition
while nth > 0:
position = data.find(value, position + 1)
if position == -1:
return -1
nth -= 1
return position
def CutData(stream, cutArgument):
if cutArgument == '':
return [stream, None, None]
typeLeft, valueLeft, typeRight, valueRight = ParseCutArgument(cutArgument)
if typeLeft == None:
return [stream, None, None]
if typeLeft == CUTTERM_NOTHING:
positionBegin = 0
elif typeLeft == CUTTERM_POSITION:
positionBegin = valueLeft
elif typeLeft == CUTTERM_FIND:
positionBegin = Find(stream, valueLeft[0], valueLeft[1])
if positionBegin == -1:
return ['', None, None]
positionBegin += valueLeft[2]
else:
raise Exception("Unknown value typeLeft")
if typeRight == CUTTERM_NOTHING:
positionEnd = len(stream)
elif typeRight == CUTTERM_POSITION and valueRight < 0:
positionEnd = len(stream) + valueRight
elif typeRight == CUTTERM_POSITION:
positionEnd = valueRight + 1
elif typeRight == CUTTERM_LENGTH:
positionEnd = positionBegin + valueRight
elif typeRight == CUTTERM_FIND:
positionEnd = Find(stream, valueRight[0], valueRight[1], positionBegin)
if positionEnd == -1:
return ['', None, None]
else:
positionEnd += len(valueRight[0])
positionEnd += valueRight[2]
else:
raise Exception("Unknown value typeRight")
return [stream[positionBegin:positionEnd], positionBegin, positionEnd]
class cHashCRC32():
def __init__(self):
self.crc32 = None
def update(self, data):
self.crc32 = zlib.crc32(data)
def hexdigest(self):
return '%08x' % (self.crc32 & 0xffffffff)
class cHashChecksum8():
def __init__(self):
self.sum = 0
def update(self, data):
if sys.version_info[0] >= 3:
self.sum += sum(data)
else:
self.sum += sum(map(ord, data))
def hexdigest(self):
return '%08x' % (self.sum)
dSpecialHashes = {'crc32': cHashCRC32, 'checksum8': cHashChecksum8}
def GetHashObjects(algorithms):
global dSpecialHashes
dHashes = {}
if algorithms == '':
algorithms = os.getenv('DSS_DEFAULT_HASH_ALGORITHMS', 'md5')
if ',' in algorithms:
hashes = algorithms.split(',')
else:
hashes = algorithms.split(';')
for name in hashes:
if not name in dSpecialHashes.keys() and not name in hashlib.algorithms_available:
print('Error: unknown hash algorithm: %s' % name)
print('Available hash algorithms: ' + ' '.join([name for name in list(hashlib.algorithms_available)] + list(dSpecialHashes.keys())))
return [], {}
elif name in dSpecialHashes.keys():
dHashes[name] = dSpecialHashes[name]()
else:
dHashes[name] = hashlib.new(name)
return hashes, dHashes
def CalculateChosenHash(data):
hashes, dHashes = GetHashObjects('')
dHashes[hashes[0]].update(data)
return dHashes[hashes[0]].hexdigest(), hashes[0]
def ExtractStringsASCII(data):
regex = REGEX_STANDARD + b'{%d,}'
return re.findall(regex % 4, data)
def ExtractStringsUNICODE(data):
regex = b'((' + REGEX_STANDARD + b'\x00){%d,})'
return [foundunicodestring.replace(b'\x00', b'') for foundunicodestring, dummy in re.findall(regex % 4, data)]
def ExtractStrings(data):
return ExtractStringsASCII(data) + ExtractStringsUNICODE(data)
def ExtractLongestString(data):
return sorted([b''] + ExtractStrings(data), key=len)[-1]
def DumpFunctionStrings(data):
return ''.join([extractedstring + '\n' for extractedstring in ExtractStrings(data)])
def RemoveLeadingEmptyLines(data):
if data[0] == '':
return RemoveLeadingEmptyLines(data[1:])
else:
return data
def RemoveTrailingEmptyLines(data):
if data[-1] == '':
return RemoveTrailingEmptyLines(data[:-1])
else:
return data
def HeadTail(data, apply):
count = 10
if apply:
lines = RemoveTrailingEmptyLines(RemoveLeadingEmptyLines(data.split('\n')))
if len(lines) <= count * 2:
return data
else:
return '\n'.join(lines[0:count] + ['...'] + lines[-count:])
else:
return data
def Translate(expression):
return lambda x: x.decode(expression)
def UTF16_ASCII(data):
string = data.decode('utf16').encode()
for byte in string:
if byte > 127:
raise Exception('UTF16_ASCII')
return string
def DecodeDataBase64(items, PreProcessFunction, PostProcessFunction):
for item in items:
data = item['content']
for base64string in re.findall(b'[ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/]+={0,2}', data):
base64string = PreProcessFunction(base64string)
if len(base64string) % 4 == 0:
try:
decodedOriginal = binascii.a2b_base64(base64string)
decoded = PostProcessFunction(decodedOriginal)
yield (item, base64string, decoded, decodedOriginal)
except:
continue
def DecodeDataBase85RFC1924(items, PreProcessFunction, PostProcessFunction):
for item in items:
data = item['content']
for base85string in re.findall(b'[ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789!#$%&()*+\-;<=>?@^_`{|}~]+', data):
base85string = PreProcessFunction(base85string)
try:
decodedOriginal = base64.b85decode(base85string)
decoded = PostProcessFunction(decodedOriginal)
yield (item, base85string, decoded, decodedOriginal)
except:
continue
def DecodeDataAscii85(items, PreProcessFunction, PostProcessFunction):
for item in items:
data = item['content']
for ascii85string in re.findall(b'''[!"#$%&'()*+,./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\]^_`abcdefghijklmnopqrstuz-]+''', data):
ascii85string = PreProcessFunction(ascii85string)
try:
decodedOriginal = base64.a85decode(ascii85string)
decoded = PostProcessFunction(decodedOriginal)
yield (item, ascii85string, decoded, decodedOriginal)
except:
continue
def DecodeDataHex(items, PreProcessFunction, PostProcessFunction):
for item in items:
data = item['content']
for hexstring in re.findall(b'[ABCDEFabcdef0123456789]+', data):
hexstring = PreProcessFunction(hexstring)
if len(hexstring) % 2 == 0:
try:
decodedOriginal = binascii.a2b_hex(hexstring)
decoded = PostProcessFunction(decodedOriginal)
yield (item, hexstring, decoded, decodedOriginal)
except:
continue
def DecodeBU(data):
decoded = b''
while data != b'':
decoded += binascii.a2b_hex(data[4:6]) + binascii.a2b_hex(data[2:4])
data = data[6:]
return decoded
def DecodeDataBU(items, PreProcessFunction, PostProcessFunction):
for item in items:
data = item['content']
for bu in re.findall(br'(?:\\u[ABCDEFabcdef0123456789]{4})+', data):
bu = PreProcessFunction(bu)
try:
decodedOriginal = DecodeBU(bu)
decoded = PostProcessFunction(decodedOriginal)
yield (item, bu, decoded, decodedOriginal)
except:
continue
def DecodeDataPU(items, PreProcessFunction, PostProcessFunction):
for item in items:
data = item['content']
for bu in re.findall(br'(?:%u[ABCDEFabcdef0123456789]{4})+', data):
bu = PreProcessFunction(bu)
try:
decodedOriginal = DecodeBU(bu)
decoded = PostProcessFunction(decodedOriginal)
yield (item, bu, decoded, decodedOriginal)
except:
continue
def DecodeBX(data):
decoded = b''
while data != b'':
decoded += binascii.a2b_hex(data[2:4])
data = data[4:]
return decoded
def DecodeDataBX(items, PreProcessFunction, PostProcessFunction):
for item in items:
data = item['content']
for bx in re.findall(br'(?:\\x[ABCDEFabcdef0123456789]{2})+', data):
bx = PreProcessFunction(bx)
try:
decodedOriginal = DecodeBX(bx)
decoded = PostProcessFunction(decodedOriginal)
yield (item, bx, decoded, decodedOriginal)
except:
continue
def DecodeDataAH(items, PreProcessFunction, PostProcessFunction):
for item in items:
data = item['content']
for ah in re.findall(br'(?:&H[ABCDEFabcdef0123456789]{2})+', data):
ah = PreProcessFunction(ah)
try:
decodedOriginal = DecodeBX(ah)
decoded = PostProcessFunction(decodedOriginal)
yield (item, ah, decoded, decodedOriginal)
except:
continue
def ReverseCount(data, count):
result = b''
while data != b'':
part = data[0:count]
data = data[count:]
result = part + result
return result
def DecodeZXLittleEndian(data):
decoded = b''
for hex in data.split(b'0x'):
if hex == b'':
continue
if len(hex) % 2 == 1:
hex = b'0' + hex
decoded += binascii.a2b_hex(ReverseCount(hex, 2))
return decoded
def DecodeDataZXLittleEndian(items, PreProcessFunction, PostProcessFunction):
for item in items:
data = item['content']
for zx in re.findall(br'(?:0x[ABCDEFabcdef0123456789]{1,8})+', data):
zx = PreProcessFunction(zx)
try:
decodedOriginal = DecodeZXLittleEndian(zx)
decoded = PostProcessFunction(decodedOriginal)
yield (item, zx, decoded, decodedOriginal)
except:
continue
def DecodeZXBigEndian(data):
decoded = b''