Large suppression files
How to handle massive suppression files appropriately
We recommend against uploading unnecessarily large suppression files directly into your platform.
Many advertisers now have suppression files that have grown to 1GB, 2GB, or even larger. The outcome of uploading files this large is unpredictable and contingent on many factors including the robustness of your hardware.
Here's the prompt I used on the free version of ChatGPT to generate a script that runs on my macbook pro that will process a large suppression file in md5sum format:
I need to create a python script that can run on osx which takes a plain text file and converts each line to an md5sum, then compares each newly md5sumed line of that file to each line of a second plain text file which already consists of md5sums, I then need the matching lines from the first file to be saved to a file in the lines original format before they were converted to md5 for the match. The pre-existing md5sums file is 812MB and over 23 million lines, so this script needs to handle large file sizes while completing quickly
Here's another example one liner with AWK in macOS terminal that can process two files of email addresses in regular plain unhashed format:
awk -F, 'FNR==NR {a[$1]; next}; $1 in a' suppression.csv contacts.csv
Last updated
Was this helpful?