(A ProxyPool Retainer on Python)
- You set several urls.
- then CDRetainer will crawl free proxylist from network automatically.
- After filter function, make a proxylist for each target.
git clone https://github.com/okcd00/CDRetainer.git
cd CDRetainer
vim source_list.txt # Change conf by yourself, default 2 targets
vim conf/basic.conf # Change paras by yourself, default run once per 43200 seconds
nohup python CDRetainer.py &
# Wait and then get your results in "./data"
# You can view the progress in ""./log"
- Now only support Python3
BeautifulSoup4
andlxml
urllib2
Replaced withurllib.request
inpython3.x
ConfigParser
Replaced withconfigparser
inpython3.x
- Use shell or GUI to Switch Retainer
- Implement: clean dead address automatically
- Make a
Configure
out of Source Code - Custom
Sourcelist
for different aims
[2020/02/12] ver 1.0.5
- Remove deprecated websites
- Add a clean function for the old address list
- Better readability
- Reform the project into Python3
[2015/12/18] ver 1.0.4
- Remove
PCP
Function for Poor performance
- Add
Sup
(Support
orSubProcess
) Module for IssueErrno.110
[2015/12/07] ver 1.0.3
- Add
Try-Except
to protect and monitor - Add
PCP
Function as Formula - Add Check-and-Retry for Empty Return
[2015/11/25] ver 1.0.2
- Remove Access Rate
- Add switch for
PCP
(use Proxy to Capture Proxy)
[2015/11/11] ver 1.0.1
- Fix Division Operation for ZeroDivisionError
[2015/11/10] ver 1.0.0
- Add
Basic.conf
Config Function - Demo with
Logger
Complete