A python library for fast parallel HTTP requests.
pip install git+https://github.com/legout/parallel-requests
...
Note The library should work fine without using random proxies. Using random proxies might be illegal.
You can set random-proxy=True
in any of the scraping functions. By default this uses free proxies*. In my experience, these proxies are not reliable, but maybe you are lucky.
I am using proxies from webshare.io. I am very happy with their service and the pricing. If you wanna use their service too, sign up (use the this link if you wanna support my work) and choose a plan that fits your needs. In the next step, go to Dashboard -> Proxy -> List -> Download and copy the download link. Set this download link as an environment variable WEBSHARE_PROXIES_URL
before importing any parallel-requests function.
Export WEBSHARE_PROXIES_URL in your linux shell
$ export WEBSHARE_PROXIES_URL="https://proxy.webshare.io/api/v2/proxy/list/download/abcdefg1234567/-/any/username/direct/-/"
You can also set this environment variable permanently in an .env
file (see the .env-exmaple
) in your home folder or current folder or in your command line config file (e.g. ~/.bashrc
).
Write WEBSHARE_PROXIES_URL into .env
WEBSHARE_PROXIES_URL="https://proxy.webshare.io/api/v2/proxy/list/download/abcdefg1234567/-/any/username/direct/-/"
or write WEBSHARE_PROXIES_URL into your shell config file (e.g. ~/.bashrc)
$ echo 'export WEBSHARE_PROXIES_URL="https://proxy.webshare.io/api/v2/proxy/list/download/abcdefg1234567/-/any/username/direct/-/"' >> ~/.bashrc
*Free Proxies are scraped from here:
- "http://www.free-proxy-list.net"
- "https://free-proxy-list.net/anonymous-proxy.html"
- "https://www.us-proxy.org/"
- "https://free-proxy-list.net/uk-proxy.html"
- "https://www.sslproxies.org/"
If you find this useful, you can buy me a coffee. Thanks!