github.com/bcampbell/scrapeomat@v0.0.0-20220820232205-23e64141c89e/TODO (about) 1 archive: use gzipped .warc files to save space 2 better setup instructions (particularly db creation) 3 artform should allow for underscored slugs (eg in http://www.lancashiretelegraph.co.uk) 4 show better error message for cascadia parse error in config files (ie file and linenum) 5 slurpserver: show proper IP address in log when behind proxy server 6 slurpserver: simple API token access (and show token in log) 7 scrapeomat: when scraping from list (-i), show how many articles already in database 8 summary API+tool: add option to batch by week 9 WARC archive read/write - should now use github.com/bcampbell/warc