123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301 |
- 2023-09-11 11:05:39 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:06:13 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:07:21 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:07:21 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:07:59 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:07:59 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:10:23 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:10:23 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:10:48 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:11:01 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:11:01 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:11:31 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:13:10 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:13:10 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:13:44 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:13:44 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:13:44 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:14:35 [scrapy.core.scraper] ERROR: Spider error processing <GET https://mp.weixin.qq.com/cgi-bin/appmsg> (referer: https://mp.weixin.qq.com/cgi-bin/searchbiz)
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/utils/defer.py", line 260, in iter_errback
- yield next(it)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/utils/python.py", line 336, in __next__
- return next(self.data)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/utils/python.py", line 336, in __next__
- return next(self.data)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/spidermw.py", line 106, in process_sync
- for r in iterable:
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/spidermiddlewares/offsite.py", line 28, in <genexpr>
- return (r for r in result or () if self._filter(r, spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/spidermw.py", line 106, in process_sync
- for r in iterable:
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/spidermiddlewares/referer.py", line 352, in <genexpr>
- return (self._set_referer(r, response) for r in result or ())
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/spidermw.py", line 106, in process_sync
- for r in iterable:
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/spidermiddlewares/urllength.py", line 27, in <genexpr>
- return (r for r in result or () if self._filter(r, spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/spidermw.py", line 106, in process_sync
- for r in iterable:
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/spidermiddlewares/depth.py", line 31, in <genexpr>
- return (r for r in result or () if self._filter(r, response, spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/spidermw.py", line 106, in process_sync
- for r in iterable:
- File "/Users/luojunhui/cyber/gzh_spider/gzh_spider/spiders/gzh_author.py", line 150, in parse_video
- item['video_url'] = functions.find_video_url(article_url)
- File "/Users/luojunhui/cyber/gzh_spider/gzh_spider/functions/get_video_url.py", line 54, in find_video_url
- video_url = get_tencent_video_url(video_id)
- File "/Users/luojunhui/cyber/gzh_spider/gzh_spider/functions/get_video_url.py", line 72, in get_tencent_video_url
- url = response["vl"]["vi"][0]["ul"]["ui"][0]["url"]
- KeyError: 'vl'
- 2023-09-11 11:14:35 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:15:46 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:16:26 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:16:26 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:16:46 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:21:00 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:21:00 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:21:27 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:21:27 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:23:52 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:24:21 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:25:57 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:26:03 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 11:44:19 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:44:19 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:44:19 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:44:51 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:44:51 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:44:51 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:50:31 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:50:31 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:51:07 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 11:51:07 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 12:00:01 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 12:00:01 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 12:00:01 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 12:00:34 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 12:00:34 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 12:00:34 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 1693, in _inlineCallbacks
- result = context.run(
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/python/failure.py", line 518, in throwExceptionIntoGenerator
- return g.throw(self.type, self.value, self.tb)
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/twisted/internet/defer.py", line 892, in _runCallbacks
- current.result = callback( # type: ignore[misc]
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/handlers/http11.py", line 397, in _cb_timeout
- raise TimeoutError(f"Getting {url} took longer than {timeout} seconds.")
- twisted.internet.error.TimeoutError: User timeout caused connection failure: Getting https://mp.weixin.qq.com/cgi-bin/searchbiz took longer than 180.0 seconds..
- 2023-09-11 12:03:03 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 12:03:03 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 12:03:03 [scrapy.downloadermiddlewares.retry] ERROR: Gave up retrying <GET https://mp.weixin.qq.com/cgi-bin/searchbiz> (failed 3 times): [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 12:03:42 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 12:03:42 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
- 2023-09-11 12:03:42 [scrapy.core.scraper] ERROR: Error downloading <GET https://mp.weixin.qq.com/cgi-bin/searchbiz>
- Traceback (most recent call last):
- File "/Users/luojunhui/miniconda3/envs/Spider/lib/python3.9/site-packages/scrapy/core/downloader/middleware.py", line 54, in process_request
- return (yield download_func(request=request, spider=spider))
- twisted.web._newclient.ResponseNeverReceived: [<twisted.python.failure.Failure twisted.internet.error.ConnectionLost: Connection to the other side was lost in a non-clean fashion: Connection lost.>]
|