Iโm an admin of a private discourse forum and I would like to scrape it to get a static backup of the site.
Why? In a few years, the forum might not exist and I would like to share with all its users a local copy so we can remember the good old times .
I think a scraper(httrack/wget) can do it, but is there a better solution? If not, has anyone already used a scraper on discourse and could give me some pointers?
I donโt really understand your request. Why wouldnโt you just grab the default backup zip file? You can extract and render that data in a variety of ways, all of which would probably be easier than figuring out a way to scrape the site.
This is on my wishlist as well, since my sites involve people in developing countries who do not always have access to the Internet. A way to put Discourse on a flash drive for offline access would be super valuable. Would love to be able to hand it out as swag!
If anyone wants to work on this and has great ideas on how it might be done let me know.