python scrapy 网络采集使用代理的方法,pythonscrapy,1.在Scrapy工程下
python scrapy 网络采集使用代理的方法,pythonscrapy,1.在Scrapy工程下
1.在Scrapy工程下新建“middlewares.py”```python
Importing base64 library because we'll need it ONLY in case if the proxy we are going to use requires authentication
import base64
Start your middleware class
class ProxyMiddleware(object): # overwrite process request def process_request(self, request, spider): # Set the location of the proxy request.meta['proxy'] = "http://YOUR_PROXY_IP:PORT"
# Use the following lines if your proxy requires authentication proxy_user_pass = "USERNAME:PASSWORD" # setup basic authentication for the proxy encoded_user_pass = base64.encodestring(proxy_user_pass) request.headers['Proxy-Authorization'] = 'Basic ' + encoded_user_pass
该代码片段来自于: http://www.sharejs.com/codes/python/8309
2.在项目配置文件里(./project_name/settings.py)添加
pythonDOWNLOADER_MIDDLEWARES = { 'scrapy.contrib.downloadermiddleware.httpproxy.HttpProxyMiddleware': 110, 'project_name.middlewares.ProxyMiddleware': 100,}
只要两步,现在请求就是通过代理的了。测试一下^_^
pythonfrom scrapy.spider import BaseSpiderfrom scrapy.contrib.spiders import CrawlSpider, Rulefrom scrapy.http import Request
class TestSpider(CrawlSpider): name = "test" domain_name = "whatismyip.com" # The following url is subject to change, you can get the last updated one from here : # http://www.whatismyip.com/faq/automation.asp start_urls = ["http://xujian.info"]
def parse(self, response): open('test.html', 'wb').write(response.body)
```
相关内容
- 使用libvirt抓取linux kvm虚拟机的缩略图,libvirtkvm,import
- 一组python用于验证数据的函数,一组python,这是一组py
- python设置进程的环境变量,python环境变量,import ospri
- python2.7自动创建虚拟机,python2.7虚拟机,# -*- coding
- 购物,,#!/usr/bin/e
- if elif else 使用,elifelse,1 #!/usr/bin
- sys.argv经典代码,sys.argv经典,import osimp
- python编写的一个通过多线程扫描端口的代码,python多线
- web小工具,web工具,# -*- coding
- python寻找1~n以内(包括n)所有素数,python素数,def f
评论关闭