the r/SneerClub archives are finally online! this is an early v1 which contains 1,940 posts grabbed from the Reddit UI using Bulk Downloader for Reddit. this encompasses both the 1000 most recent posts on r/SneerClub as well as a set of popular historical posts
as a v1, you’ll notice a lot of jank. known issues are:
- this won’t work at all on mobile because my css is garbage. it might not even work on anyone else’s screen; good luck!
- as mentioned above, only 1,940 posts are in this release. there’s a full historical archive of r/SneerClub sourced from pushshift at the archive data git repo (or clone
git://these.awful.systems/sneer-archive-data.git
); the remaining work here is to merge the BDFR and pushshift data into the same JSON format so the archives can pull in everything - markdown is only rendered for posts and first-level comments; everything else just gets the raw markdown. I couldn’t figure out how to make miller recursively parse JSON, so I might have to write some javascript for this
- likewise, comments display a unix epoch instead of a rendered time
- searching happens locally in your browser, but only post titles and authors are indexed to keep download sizes small
- speaking of, there’s a much larger r/SneerClub archive that includes the media files BDFR grabbed while archiving. it’s a bit unmanageable to actually use directly, but is available for archival purposes (and could be included as part of the hosted archive if there’s demand for it)
if you’d like the source code for the r/SneerClub archive static site, it lives here (or clone git://these.awful.systems/sneer-archive-site.git
)
awful.systems will not rest until the words Eliezer Yudkowsky in a prompt cause ChatGPT to permanently enter sneer mode