Skip to content

AlejandroAkbal/proxy-scraper

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Proxy scraper and checker

Scrape more than 1K HTTP proxies in less than 2 seconds.

Scraping fresh public proxies from different sources:

Installation

Use this command to install dependencies.

python3 -m venv env

source env/bin/activate

pip3 install -r requirements.txt

Usage

For scraping:

python3 proxyScraper.py -p http
  • With -p or --proxy, You can choose your proxy type. Supported proxy types are: HTTP - HTTPS - Socks (Both 4 and 5) - Socks4 - Socks5
  • With -o or --output, create and write to a .txt file. (Default is output.txt)
  • With -v or --verbose, more details.
  • With -h or --help, Show help to who did't read this README.

For checking:

python3 proxyChecker.py -t 20 -s google.com -l output.txt
  • With -t or --timeout, dismiss the proxy after -t seconds (Default is 20)
  • With -p or --proxy, check HTTPS or HTTP proxies (Default is HTTP)
  • With -l or --list, path to your list.txt. (Default is output.txt)
  • With -s or --site, check with specific website like google.com. (Default is google.com)
  • With -v or --verbose, more details.
  • With -h or --help, Show help to who did't read this README.

Good to know

  • Dead proxies will be removed and just alive proxies will stay.
  • This script is also able to scrape Socks, but proxyChecker only check HTTP(S) proxies.

Contributing

Pull requests are welcome. For major changes, please open an issue first to discuss what you would like to change.

Credit

License

MIT

Languages

  • Python 100.0%