python 中的代理设置问题

2018-10-15 16:57:47 [scrapy.core.engine] INFO: Spider opened
2018-10-15 16:57:47 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min)
2018-10-15 16:57:47 [scrapy.extensions.telnet] DEBUG: Telnet console listening on 127.0.0.1:6034
2018-10-15 16:57:47 [scrapy.core.scraper] ERROR: Error downloading <POST https://www.lagou.com/gongsi/0-1-0-0.json>
Traceback (most recent call last):
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/twisted/internet/defer.py", line 1386, in _inlineCallbacks
    result = g.send(result)
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/scrapy/core/downloader/middleware.py", line 37, in process_request
    response = yield method(request=request, spider=spider)
TypeError: process_request() got an unexpected keyword argument 'spider'
2018-10-15 16:57:47 [scrapy.core.scraper] ERROR: Error downloading <POST https://www.lagou.com/gongsi/0-1-0-0.json>
Traceback (most recent call last):
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/twisted/internet/defer.py", line 1386, in _inlineCallbacks
    result = g.send(result)
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/scrapy/core/downloader/middleware.py", line 37, in process_request
    response = yield method(request=request, spider=spider)
TypeError: process_request() got an unexpected keyword argument 'spider'
2018-10-15 16:57:47 [scrapy.core.scraper] ERROR: Error downloading <POST https://www.lagou.com/gongsi/0-1,2-0-0.json>
Traceback (most recent call last):
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/twisted/internet/defer.py", line 1386, in _inlineCallbacks
    result = g.send(result)
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/scrapy/core/downloader/middleware.py", line 37, in process_request
    response = yield method(request=request, spider=spider)
TypeError: process_request() got an unexpected keyword argument 'spider'
2018-10-15 16:57:47 [scrapy.core.scraper] ERROR: Error downloading <POST https://www.lagou.com/gongsi/0-1,2-0-0.json>
Traceback (most recent call last):
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/twisted/internet/defer.py", line 1386, in _inlineCallbacks
    result = g.send(result)
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/scrapy/core/downloader/middleware.py", line 37, in process_request
    response = yield method(request=request, spider=spider)
TypeError: process_request() got an unexpected keyword argument 'spider'
2018-10-15 16:57:47 [scrapy.core.scraper] ERROR: Error downloading <POST https://www.lagou.com/gongsi/0-1,2,3-0-0.json>
Traceback (most recent call last):
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/twisted/internet/defer.py", line 1386, in _inlineCallbacks
    result = g.send(result)
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/scrapy/core/downloader/middleware.py", line 37, in process_request
    response = yield method(request=request, spider=spider)
TypeError: process_request() got an unexpected keyword argument 'spider'
2018-10-15 16:57:47 [scrapy.core.scraper] ERROR: Error downloading <POST https://www.lagou.com/gongsi/0-1,2,3-0-0.json>
Traceback (most recent call last):
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/twisted/internet/defer.py", line 1386, in _inlineCallbacks
    result = g.send(result)
  File "/home/shenjianlin/.local/lib/python3.4/site-packages/scrapy/core/downloader/middleware.py", line 37, in process_request
    response = yield method(request=request, spider=spider)
TypeError: process_request() got an unexpected keyword argument 'spider'
2018-10-15 16:57:47 [scrapy.core.engine] INFO: Closing spider (finished)
2018-10-15 16:57:47 [scrapy.statscollectors] INFO: Dumping Scrapy stats:
{'downloader/exception_count': 6,





class LagouMiddlerware():

def __init__(self):
    proxy_host = "w5.t.16yun.cn"
    proxy_port = "6469"
    self.username = "16ISMSBW"
    self.password = "162267"
    self.proxies = {"https": "https://{}:{}/".format(proxy_host, proxy_port)}
    self.proxy_server = 'https://w5.t.16yun.cn:6469'
    self.proxy_authorization = 'Basic ' + base64.urlsafe_b64encode(
        bytes((self.username + ':' + self.password), 'ascii')).decode('utf8')


def process_request(self, request):
    request.meta['proxy'] = self.proxy_server
    request.headers['Proxy-Authorization'] = self.proxy_authorization
    

我在这里设置了动态代理,但是总是报错?求原因?

阅读 3.9k
1 个回答

你在 middlewares.py 文件中设置动态代理后,还需要在 settings.py 的 SPIDER_MIDDLEWARES 中声明一下,然后再试试。

撰写回答
你尚未登录,登录后可以
  • 和开发者交流问题的细节
  • 关注并接收问题和回答的更新提醒
  • 参与内容的编辑和改进,让解决方法与时俱进
推荐问题