Scrapy gzip errors -
i seem getting gzip errors scrapy quite lot. example:
2016-11-04 16:18:25 [scrapy] error: error downloading <get http://www.laithwaites.co.uk/wines/grenache/_/n-1z141xy?icamp=nav-browse-grenache&no=0#page-1> traceback (most recent call last): file "/path/to/my/proj/venv/local/lib/python2.7/site-packages/twisted/internet/defer.py", line 1185, in _inlinecallbacks result = g.send(result) file "/path/to/my/proj/venv/local/lib/python2.7/site-packages/scrapy/core/downloader/middleware.py", line 53, in process_response spider=spider) file "/path/to/my/proj/venv/local/lib/python2.7/site-packages/scrapy/downloadermiddlewares/httpcompression.py", line 30, in process_response decoded_body = self._decode(response.body, encoding.lower()) file "/path/to/my/proj/venv/local/lib/python2.7/site-packages/scrapy/downloadermiddlewares/httpcompression.py", line 46, in _decode body = gunzip(body) file "/path/to/my/proj/venv/local/lib/python2.7/site-packages/scrapy/utils/gz.py", line 37, in gunzip chunk = read1(f, 8196) file "/path/to/my/proj/venv/local/lib/python2.7/site-packages/scrapy/utils/gz.py", line 21, in read1 return gzf.read(size) file "/usr/lib/python2.7/gzip.py", line 261, in read self._read(readsize) file "/usr/lib/python2.7/gzip.py", line 312, in _read uncompress = self.decompress.decompress(buf) error: error -3 while decompressing: invalid distance far
when later try fetch url in shell there no problem.
i read elsewhere maybe removing httpcompression
middleware solution, i'm worried in case has other implications.
Comments
Post a Comment