Remove duplicates from MASSIVE wordlist, without sorting it (for dictionary-based password cracking)
Nowadays, password wordlist creation usually implies concatenating
multiple data sources.
Ideally, most probable passwords should stand at start of the wordlist,
so most common passwords are cracked instantly.
With existing dedupe tools you are forced to choose
if you prefer to preserve the order OR handle massive wordlists.
Unfortunately, wordlist creation requires both
So i wrote duplicut in highly optimized C to address this very specific need 🤓 💻
git clone https://github.com/nil0x42/duplicut # download ...
cd duplicut/ && make # compile ...
./duplicut wordlist.txt -o clean-wordlist.txt # dedupe !!!
-o, --outfile <FILE> Write result to <FILE>
-t, --threads <NUM> Max threads to use (default max)
-m, --memlimit <VALUE> Limit max used memory (default max)
-l, --line-max-size <NUM> Max line size (default 1024, max 4095)
-p, --printable Filter ascii printable lines
-c, --lowercase Convert wordlist to lowercase
-C, --uppercase Convert wordlist to uppercase
-D, --dupfile <FILE> Write dupes to <FILE> (slows down duplicut)
-h, --help Display this help and exit
-v, --version Output version information and exit
Features:
--line-max-size
option)--printable
option)--dupfile
option)--lowercase
or --uppercase
optionsImplementation:
Limitations:
--line-max-size
can’t exceed 4095An uint64
is enough to index lines in hashmap, by packing
size
info within pointer’s extra bits:
If whole file can’t fit in memory, it is split into
virtual chunks, in such way that each chunk uses as much RAM as possible.
Each chunk is then loaded into hashmap, deduped, and tested against
subsequent chunks.
That way, execution time decreases to at most th triangle number:
If you find a bug, or something doesn’t work as expected,
please compile duplicut in debug mode and post an issue with
attached output:
# debug level can be from 1 to 4
make debug level=1
./duplicut [OPTIONS] 2>&1 | tee /tmp/duplicut-debug.log