CommentStreams:C53aed343d3ff3c1592deac84cad136a

Hi, Wiki Team.

In the context of a network isolation preparation project which will offer a rsync or torrent download for the people interested to have a compilation of essential free software components, I found that the Gentoo wiki may be a very important resource along with the already downloaded Gentoo mirror.

I thought that since scrapping your entire website may raise a few eyebrowns to ask first to iterate over the entire wiki's articles.

I want to limit the scrapper to as much as 5 simultaneous downloads at a time and I am (if I am granted permission.) going to use the Mojo::UserAgent Perl CPAN library.

The content to be downloaded is the source and name of every single article not in the User: namespace of the wiki, no multimedia content will be downloaded, no images and no video.

I would want first permission to do this and then I would like to ask for guidance to avoid to cause any problem, I read the FAQs about the privacy concerns which will be avoided cleaning any personal data and avoiding the User: namespace, but if any other but exists I would want to hear about it to avoid any problems that may arise.

I will develop the scrapper by myself, I am not going to use any existent tool to do it.

Regards, PerlDownloader.