Desktop version Jobs
  • Analysis
  • Categories
  • Charts
  • Logs
  • Crawler.stats
  • projectsourcing_v2
    spiderbca.uk
    job2026-03-30T10_28_25
    first_log_time2026-03-30 10:28:31
    latest_log_time2026-03-30 11:09:57
    runtime0:41:26
    crawled_pages 1039
    scraped_items 1455
    shutdown_reasonN/A
    finish_reasonfinished
    log_critical_count0
    log_error_count142
    log_warning_count387
    log_redirect_count9977
    log_retry_count0
    log_ignore_count0
    latest_crawl
    latest_scrape
    latest_log
    current_time
    latest_itemN/A
    • WARNING+

    • error_logs
      last 10 of 142

      2026-03-30 10:59:41 [crawlers.middlewares.monitoring_spider_middleware] ERROR: 'YH14%20FTD'
      Traceback (most recent call last):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/depth.py", line 35, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/photo_download_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/report_download_middleware.py", line 36, in process_spider_output
          item._scrape_type = spider.identifier_scrape_type[identifier]
                              ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
      KeyError: 'YH14%20FTD'
      2026-03-30 10:59:41 [crawlers.middlewares.monitoring_spider_middleware] ERROR: 'YH14%20FTD'
      Traceback (most recent call last):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/urllength.py", line 30, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/depth.py", line 35, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/photo_download_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/report_download_middleware.py", line 36, in process_spider_output
          item._scrape_type = spider.identifier_scrape_type[identifier]
                              ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
      KeyError: 'YH14%20FTD'
      2026-03-30 10:59:41 [crawlers.middlewares.monitoring_spider_middleware] ERROR: 'YH14%20FTD'
      Traceback (most recent call last):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/referer.py", line 355, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/urllength.py", line 30, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/depth.py", line 35, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/photo_download_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/report_download_middleware.py", line 36, in process_spider_output
          item._scrape_type = spider.identifier_scrape_type[identifier]
                              ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
      KeyError: 'YH14%20FTD'
      2026-03-30 10:59:41 [scrapy.core.scraper] ERROR: Spider error processing <GET https://bcamediaprod.blob.core.windows.net/private/pdfs/InspectionBase/GB/YH14FTD/42700264?sv=2020-08-04&st=2026-03-30T10%3A59%3A34Z&se=2026-03-30T11%3A24%3A34Z&sr=b&sp=r&sig=ZAPfqdk15Zsobve0jgJ4SsoUYeT0XFrR6dV5D7TwGXo%3D> (referer: https://www.bca.co.uk/)
      Traceback (most recent call last):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/defer.py", line 295, in aiter_errback
          yield await it.__anext__()
                ^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/python.py", line 374, in __anext__
          return await self.data.__anext__()
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/python.py", line 355, in _async_chain
          async for o in as_async_generator(it):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/asyncgen.py", line 14, in as_async_generator
          async for r in it:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/python.py", line 374, in __anext__
          return await self.data.__anext__()
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/python.py", line 355, in _async_chain
          async for o in as_async_generator(it):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/asyncgen.py", line 14, in as_async_generator
          async for r in it:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy_zyte_api/_middlewares.py", line 206, in process_spider_output_async
          async for item_or_request in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/referer.py", line 355, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/urllength.py", line 30, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/depth.py", line 35, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/photo_download_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/report_download_middleware.py", line 36, in process_spider_output
          item._scrape_type = spider.identifier_scrape_type[identifier]
                              ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
      KeyError: 'YH14%20FTD'
      2026-03-30 10:59:41 [crawlers.middlewares.monitoring_spider_middleware] ERROR: 'SF69%20NHK'
      Traceback (most recent call last):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/report_download_middleware.py", line 36, in process_spider_output
          item._scrape_type = spider.identifier_scrape_type[identifier]
                              ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
      KeyError: 'SF69%20NHK'
      2026-03-30 10:59:41 [crawlers.middlewares.monitoring_spider_middleware] ERROR: 'SF69%20NHK'
      Traceback (most recent call last):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/photo_download_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/report_download_middleware.py", line 36, in process_spider_output
          item._scrape_type = spider.identifier_scrape_type[identifier]
                              ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
      KeyError: 'SF69%20NHK'
      2026-03-30 10:59:41 [crawlers.middlewares.monitoring_spider_middleware] ERROR: 'SF69%20NHK'
      Traceback (most recent call last):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/depth.py", line 35, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/photo_download_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/report_download_middleware.py", line 36, in process_spider_output
          item._scrape_type = spider.identifier_scrape_type[identifier]
                              ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
      KeyError: 'SF69%20NHK'
      2026-03-30 10:59:41 [crawlers.middlewares.monitoring_spider_middleware] ERROR: 'SF69%20NHK'
      Traceback (most recent call last):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/urllength.py", line 30, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/depth.py", line 35, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/photo_download_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/report_download_middleware.py", line 36, in process_spider_output
          item._scrape_type = spider.identifier_scrape_type[identifier]
                              ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
      KeyError: 'SF69%20NHK'
      2026-03-30 10:59:42 [crawlers.middlewares.monitoring_spider_middleware] ERROR: 'SF69%20NHK'
      Traceback (most recent call last):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/referer.py", line 355, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/urllength.py", line 30, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/depth.py", line 35, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/photo_download_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/report_download_middleware.py", line 36, in process_spider_output
          item._scrape_type = spider.identifier_scrape_type[identifier]
                              ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
      KeyError: 'SF69%20NHK'
      2026-03-30 10:59:42 [scrapy.core.scraper] ERROR: Spider error processing <GET https://bcamediaprod.blob.core.windows.net/private/pdfs/InspectionBase/GB/SF69NHK/42794010?sv=2020-08-04&st=2026-03-30T10%3A59%3A34Z&se=2026-03-30T11%3A24%3A34Z&sr=b&sp=r&sig=yDB3zttIOrfRkZohgm5t6owhWuvZx7%2BHB5OgRlUZZ%2Bo%3D> (referer: https://www.bca.co.uk/)
      Traceback (most recent call last):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/defer.py", line 295, in aiter_errback
          yield await it.__anext__()
                ^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/python.py", line 374, in __anext__
          return await self.data.__anext__()
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/python.py", line 355, in _async_chain
          async for o in as_async_generator(it):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/asyncgen.py", line 14, in as_async_generator
          async for r in it:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/python.py", line 374, in __anext__
          return await self.data.__anext__()
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/python.py", line 355, in _async_chain
          async for o in as_async_generator(it):
        File "/usr/local/lib/python3.11/dist-packages/scrapy/utils/asyncgen.py", line 14, in as_async_generator
          async for r in it:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy_zyte_api/_middlewares.py", line 206, in process_spider_output_async
          async for item_or_request in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/referer.py", line 355, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/urllength.py", line 30, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/spidermiddlewares/depth.py", line 35, in process_spider_output_async
          async for r in result or ():
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/photo_download_middleware.py", line 42, in process_spider_output
          async for item in result:
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 121, in process_async
          exception_result = self._process_spider_exception(
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 150, in _process_spider_exception
          result = method(response=response, exception=exception, spider=spider)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/usr/src/app/crawlers/middlewares/monitoring_spider_middleware.py", line 73, in process_spider_exception
          raise exception
        File "/usr/local/lib/python3.11/dist-packages/scrapy/core/spidermw.py", line 118, in process_async
          async for r in iterable:
        File "/usr/src/app/crawlers/middlewares/report_download_middleware.py", line 36, in process_spider_output
          item._scrape_type = spider.identifier_scrape_type[identifier]
                              ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
      KeyError: 'SF69%20NHK'
    • warning_logs
      last 10 of 387

      2026-03-30 11:09:37 [scrapy_zyte_api._params] WARNING: Request <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=125&sort=MostRecentlyAdded> defines the Zyte API customHttpRequestHeaders parameter, overriding Request.headers. Use Request.headers instead.
      2026-03-30 11:09:39 [scrapy_zyte_api._params] WARNING: Request <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=126&sort=MostRecentlyAdded> defines the Zyte API customHttpRequestHeaders parameter, overriding Request.headers. Use Request.headers instead.
      2026-03-30 11:09:41 [scrapy_zyte_api._params] WARNING: Request <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=126&sort=MostRecentlyAdded> defines the Zyte API customHttpRequestHeaders parameter, overriding Request.headers. Use Request.headers instead.
      2026-03-30 11:09:41 [scrapy_zyte_api._params] WARNING: Request <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=126&sort=MostRecentlyAdded> defines the Zyte API customHttpRequestHeaders parameter, overriding Request.headers. Use Request.headers instead.
      2026-03-30 11:09:42 [scrapy_zyte_api._params] WARNING: Request <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=127&sort=MostRecentlyAdded> defines the Zyte API customHttpRequestHeaders parameter, overriding Request.headers. Use Request.headers instead.
      2026-03-30 11:09:45 [scrapy_zyte_api._params] WARNING: Request <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=127&sort=MostRecentlyAdded> defines the Zyte API customHttpRequestHeaders parameter, overriding Request.headers. Use Request.headers instead.
      2026-03-30 11:09:45 [scrapy_zyte_api._params] WARNING: Request <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=127&sort=MostRecentlyAdded> defines the Zyte API customHttpRequestHeaders parameter, overriding Request.headers. Use Request.headers instead.
      2026-03-30 11:09:50 [scrapy_zyte_api._params] WARNING: Request <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=128&sort=MostRecentlyAdded> defines the Zyte API customHttpRequestHeaders parameter, overriding Request.headers. Use Request.headers instead.
      2026-03-30 11:09:51 [scrapy_zyte_api._params] WARNING: Request <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=128&sort=MostRecentlyAdded> defines the Zyte API customHttpRequestHeaders parameter, overriding Request.headers. Use Request.headers instead.
      2026-03-30 11:09:51 [scrapy_zyte_api._params] WARNING: Request <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=128&sort=MostRecentlyAdded> defines the Zyte API customHttpRequestHeaders parameter, overriding Request.headers. Use Request.headers instead.

      INFO

    • redirect_logs
      last 10 of 9977

      2026-03-30 11:08:44 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/AE65YCK/699608538/600> from <GET https://www1.bcaimage.com/Document?DocType=VehicleImage&width=600&docId=699608538>
      2026-03-30 11:08:44 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/EO62FCN/699621183/600> from <GET https://www1.bcaimage.com/Document?DocType=VehicleImage&width=600&docId=699621183>
      2026-03-30 11:08:44 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/AE65YCK/699608566/600> from <GET https://www1.bcaimage.com/Document?DocType=VehicleImage&width=600&docId=699608566>
      2026-03-30 11:08:44 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/EO62FCN/699621218/600> from <GET https://www1.bcaimage.com/Document?DocType=VehicleImage&width=600&docId=699621218>
      2026-03-30 11:08:44 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/AE65YCK/699608522/600> from <GET https://www1.bcaimage.com/Document?DocType=VehicleImage&width=600&docId=699608522>
      2026-03-30 11:08:45 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/EO62FCN/699621373/600> from <GET https://www1.bcaimage.com/Document?DocType=VehicleImage&width=600&docId=699621373>
      2026-03-30 11:08:45 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/EO62FCN/699621330/600> from <GET https://www1.bcaimage.com/Document?DocType=VehicleImage&width=600&docId=699621330>
      2026-03-30 11:08:45 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/EO62FCN/699621287/600> from <GET https://www1.bcaimage.com/Document?DocType=VehicleImage&width=600&docId=699621287>
      2026-03-30 11:08:45 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/AE65YCK/699608592/600> from <GET https://www1.bcaimage.com/Document?DocType=VehicleImage&width=600&docId=699608592>
      2026-03-30 11:08:45 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/EO62FCN/699621251/600> from <GET https://www1.bcaimage.com/Document?DocType=VehicleImage&width=600&docId=699621251>

      DEBUG

    • scrapy_version

      2.11.2
    • telnet_console

      127.0.0.1:6024
    • telnet_password

      015baff6453dcd36
    • latest_crawl

      2026-03-30 11:09:52 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.bca.co.uk/api/search?q=&pageSize=100&page=128&sort=MostRecentlyAdded> (referer: https://www.bca.co.uk/search) ['zyte-api']
    • latest_scrape

      2026-03-30 11:08:49 [scrapy.core.scraper] DEBUG: Scraped from <200 https://bcamediaprod.blob.core.windows.net/public/images/vehicle/GB/AE65YCK/699608592/600>
    • latest_stat

      2026-03-30 11:09:32 [scrapy.extensions.logstats] INFO: Crawled 1034 pages (at 11 pages/min), scraped 1455 items (at 2 items/min)
    • Head

      2026-03-30 10:28:31 [scrapy.utils.log] INFO: Scrapy 2.11.2 started (bot: SourcingV2)
      2026-03-30 10:28:31 [scrapy.utils.log] INFO: Versions: lxml 5.2.2.0, libxml2 2.12.6, cssselect 1.2.0, parsel 1.9.1, w3lib 2.1.2, Twisted 24.3.0, Python 3.11.2 (main, Mar 13 2023, 12:18:29) [GCC 12.2.0], pyOpenSSL 24.1.0 (OpenSSL 3.2.1 30 Jan 2024), cryptography 42.0.7, Platform Linux-5.15.0-1098-azure-x86_64-with-glibc2.36
      2026-03-30 10:28:31 [bca.uk] INFO: Starting spider bca.uk
      2026-03-30 10:28:31 [azure.identity._credentials.environment] INFO: Incomplete environment configuration for EnvironmentCredential. These variables are set: AZURE_CLIENT_ID, AZURE_TENANT_ID
      2026-03-30 10:28:31 [urllib3.connectionpool] DEBUG: Starting new HTTPS connection (1): login.microsoftonline.com:443
      2026-03-30 10:28:31 [urllib3.connectionpool] DEBUG: https://login.microsoftonline.com:443 "POST /8ea908c1-4e85-4692-bc3f-3646b9b40891/oauth2/v2.0/token HTTP/1.1" 200 2122
      2026-03-30 10:28:31 [azure.identity._credentials.chained] INFO: DefaultAzureCredential acquired a token from WorkloadIdentityCredential
      2026-03-30 10:28:31 [urllib3.connectionpool] DEBUG: Starting new HTTPS connection (1): alxsourcingstorageprod.table.core.windows.net:443
      2026-03-30 10:28:32 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "POST /Tables HTTP/1.1" 409 None
      2026-03-30 10:28:32 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "GET /SpiderData()?$filter=PartitionKey%20eq%20%27BCAUk%27%20and%20RowKey%20eq%20%27cookies%27 HTTP/1.1" 200 None
      2026-03-30 10:28:32 [bca.uk] INFO: No cached cookies found, will perform fresh login
      2026-03-30 10:28:32 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "POST /Tables HTTP/1.1" 409 None
      2026-03-30 10:28:32 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "GET /SpiderData()?$filter=PartitionKey%20eq%20%27BCAUk%27%20and%20RowKey%20eq%20%27auth_token%27 HTTP/1.1" 200 None
      2026-03-30 10:28:32 [bca.uk] INFO: Loaded cached Auth0 token from Azure Tables (length=1492)
      2026-03-30 10:28:32 [scrapy.addons] INFO: Enabled addons:
      []
      2026-03-30 10:28:32 [asyncio] DEBUG: Using selector: EpollSelector
      2026-03-30 10:28:32 [scrapy.utils.log] DEBUG: Using reactor: twisted.internet.asyncioreactor.AsyncioSelectorReactor
      2026-03-30 10:28:32 [scrapy.utils.log] DEBUG: Using asyncio event loop: asyncio.unix_events._UnixSelectorEventLoop
      2026-03-30 10:28:32 [scrapy.extensions.telnet] INFO: Telnet Password: 015baff6453dcd36
      2026-03-30 10:28:32 [scrapy.middleware] INFO: Enabled extensions:
      ['scrapy.extensions.corestats.CoreStats',
       'scrapy.extensions.telnet.TelnetConsole',
       'scrapy.extensions.memusage.MemoryUsage',
       'scrapy.extensions.feedexport.FeedExporter',
       'scrapy.extensions.logstats.LogStats',
       'scrapy.extensions.closespider.CloseSpider']
      2026-03-30 10:28:32 [scrapy.crawler] INFO: Overridden settings:
      {'BOT_NAME': 'SourcingV2',
       'CLOSESPIDER_TIMEOUT': 7200,
       'DOWNLOAD_MAXSIZE': 52428800,
       'DOWNLOAD_WARNSIZE': 10485760,
       'FEED_EXPORT_ENCODING': 'utf-8',
       'LOG_FILE': '/var/log/scrapyd/logs/sourcing_v2/bca.uk/2026-03-30T10_28_25.log',
       'LOG_FORMATTER': 'crawlers.log_formatter.SourcingLogFormatter',
       'MEMUSAGE_LIMIT_MB': 2048,
       'MEMUSAGE_WARNING_MB': 1536,
       'NEWSPIDER_MODULE': 'spiders',
       'REQUEST_FINGERPRINTER_CLASS': 'scrapy_zyte_api.ScrapyZyteAPIRequestFingerprinter',
       'REQUEST_FINGERPRINTER_IMPLEMENTATION': '2.7',
       'SPIDER_MODULES': ['spiders', 'auth_check'],
       'TWISTED_REACTOR': 'twisted.internet.asyncioreactor.AsyncioSelectorReactor',
       'USER_AGENT': ''}
      2026-03-30 10:28:32 [scrapy_zyte_api.handler] INFO: Using a Zyte API key starting with '5857011'
      2026-03-30 10:28:32 [scrapy_zyte_api.handler] INFO: Using a Zyte API key starting with '5857011'
      2026-03-30 10:28:32 [scrapy.middleware] INFO: Enabled downloader middlewares:
      ['scrapy.downloadermiddlewares.offsite.OffsiteMiddleware',
       'scrapy.downloadermiddlewares.httpauth.HttpAuthMiddleware',
       'scrapy.downloadermiddlewares.downloadtimeout.DownloadTimeoutMiddleware',
       'scrapy.downloadermiddlewares.defaultheaders.DefaultHeadersMiddleware',
       'scrapy.downloadermiddlewares.useragent.UserAgentMiddleware',
       'scrapy_zyte_api.ScrapyZyteAPIDownloaderMiddleware',
       'scrapy.downloadermiddlewares.retry.RetryMiddleware',
       'scrapy.downloadermiddlewares.redirect.MetaRefreshMiddleware',
       'scrapy.downloadermiddlewares.httpcompression.HttpCompressionMiddleware',
       'scrapy.downloadermiddlewares.redirect.RedirectMiddleware',
       'scrapy.downloadermiddlewares.cookies.CookiesMiddleware',
       'scrapy.downloadermiddlewares.httpproxy.HttpProxyMiddleware',
       'scrapy.downloadermiddlewares.stats.DownloaderStats']
      2026-03-30 10:28:32 [crawlers.middlewares.id_gen_middleware] INFO: Setting up IdGenerationMiddleware
      2026-03-30 10:28:32 [scrapy.middleware] INFO: Enabled spider middlewares:
      ['scrapy.spidermiddlewares.httperror.HttpErrorMiddleware',
       'scrapy_zyte_api.ScrapyZyteAPISpiderMiddleware',
       'crawlers.middlewares.monitoring_spider_middleware.MonitoringSpiderMiddleware',
       'scrapy.spidermiddlewares.referer.RefererMiddleware',
       'scrapy.spidermiddlewares.urllength.UrlLengthMiddleware',
       'scrapy.spidermiddlewares.depth.DepthMiddleware',
       'crawlers.middlewares.photo_download_middleware.PhotoDownloadMiddleware',
       'crawlers.middlewares.report_download_middleware.ReportDownloadMiddleware',
       'crawlers.middlewares.id_gen_middleware.IdGenMiddleware']
      2026-03-30 10:28:32 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "POST /Tables HTTP/1.1" 409 None
      2026-03-30 10:28:32 [crawlers.pipelines.translation_pipeline] INFO: Loading translations for language: auto
      2026-03-30 10:28:32 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "GET /Translations()?$filter=PartitionKey%20eq%20%27auto%27%20and%20RowKey%20eq%20%27auto%27 HTTP/1.1" 200 None
      2026-03-30 10:28:32 [crawlers.pipelines.item_rules_pipeline] INFO: Setting up ItemRules Pipeline
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: invalid_location_for_country.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: damaged_cars_from_auction_title.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: puretech_for_country.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: puretech_for_fr.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: missing_photos.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: damaged_from_info.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: makes_not_allowed.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: not_operable_from_info.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: makes_models_not_allowed.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: keywords_from_title.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: imported_cars.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: invalid_currency.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: missing_mileage.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: keywords_from_auction_title.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: invalid_country_of_origin.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: puretech_for_pt.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: electric_cars.json
      2026-03-30 10:28:32 [crawlers.filter_rules.rules_loader] INFO: Loaded rule: missing_color.json
      2026-03-30 10:28:32 [crawlers.pipelines.post_to_api] INFO: Setting up PostToApi Pipeline pointing to https://api.app.infinit.cc/api/command/Alx.Cars.Contracts.Internal.Sourcing.AddListing
      2026-03-30 10:28:32 [scrapy.middleware] INFO: Enabled item pipelines:
      ['crawlers.pipelines.translation_pipeline.TranslationPipeline',
       'crawlers.pipelines.item_rules_pipeline.ItemRulesPipeline',
       'crawlers.pipelines.post_to_api.PostToApiPipeline']
      2026-03-30 10:28:32 [scrapy.core.engine] INFO: Spider opened
      2026-03-30 10:28:32 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min)
      2026-03-30 10:28:32 [scrapy.extensions.telnet] INFO: Telnet console listening on 127.0.0.1:6024
    • Tail

      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SA20%2520HCP') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SA68%2520MJU') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SB13%2520FWT') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SB65%2520CME') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SB73%2520JKY') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SC19%2520OEY') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SC65%2520AUO') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SC68%2520OGG') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SC71%2520KVP') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SC71%2520MZO') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SD67%2520HPO') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SE17%2520GXP') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SE66%2520NYH') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SF65%2520VBR') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SG18%2520PPF') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SH62%2520YPY') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SH70%2520KXT') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SJ61%2520XUA') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SK20%2520WNX') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SK24%2520XBO') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SL21%2520ZNU') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SL21%2520ZPE') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SM21%2520DHK') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SN72%2520GVM') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SO21%2520EJX') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SO21%2520HHN') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SP17%2520ZNA') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SP64%2520ZVD') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='ST15%2520VAA') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='ST17%2520JYS') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SV62%2520UGF') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SV67%2520YXK') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SW20%2520OJN') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SW70%2520SVT') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SX19%2520HYN') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SX66%2520FNF') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SY13%2520ZBC') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='SY20%2520LAE') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VA19%2520OND') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VE62%2520UPH') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VE71%2520XWR') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VK17%2520UKN') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VK63%2520NWA') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VK69%2520EZX') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VK70%2520FNH') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VK71%2520NFP') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VO14%2520GMX') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VO18%2520ZYK') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VO20%2520CEN') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VU18%2520KUB') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VU21%2520TYW') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VU63%2520VXH') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VX66%2520OCO') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='VX71%2520LRE') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WA17%2520USN') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WA69%2520WUU') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WA73%2520VFZ') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WD17%2520LZH') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WD17%2520RVZ') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WD25%2520LLN') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WF13%2520FBC') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WF17%2520AYH') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WF20%2520UKD') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WF57%2520NXU') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WF69%2520CKG') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WF70%2520ZFG') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WG18%2520OML') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WG23%2520UFX') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WG64%2520ZWA') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WG71%2520AWX') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WG71%2520BNV') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WG73%2520OGB') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WH22%2520BZR') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WJ11%2520HZE') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WJ17%2520JVH') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WJ22%2520BKL') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WJ64%2520ZHV') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WJ67%2520LBG') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WJ67%2520YVG') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WJ72%2520MPU') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WK06%2520GGE') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WK16%2520ZYM') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WK63%2520YLS') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WK66%2520HFS') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WL19%2520NZS') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WM71%2520OAO') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WM72%2520XSP') HTTP/1.1" 204 0
      2026-03-30 11:09:56 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WN14%2520UAV') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WN18%2520XCD') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WN71%2520KYJ') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WO15%2520CKF') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WR23%2520SBZ') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WR71%2520JLO') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WR71%2520JNO') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WU15%2520GMZ') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WV17%2520GDJ') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WV18%2520BKK') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WV18%2520ZDK') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='WV71%2520UZD') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='X7%2520GKR') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YB67%2520TSZ') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YB68%2520FFT') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YC12%2520PLJ') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YC67%2520EMF') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YC67%2520EVH') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YE10%2520CHG') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YE12%2520UNY') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YE20%2520PLV') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YE72%2520JWX') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YF24%2520LVW') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YF65%2520DVJ') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YG17%2520GSV') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YH17%2520OOV') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YH22%2520KCC') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YH24%2520KVO') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YJ21%2520KMB') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YJ67%2520WLN') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YL19%2520ZKS') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YL67%2520FKT') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YM20%2520NRY') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YM23%2520XHR') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YN72%2520JOU') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YP73%2520HJG') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YR18%2520XHU') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YR21%2520LFM') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YR22%2520DBY') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YR58%2520XPT') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YS16%2520RVW') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YT69%2520XJE') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YY15%2520HAO') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YY23%2520HAE') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [urllib3.connectionpool] DEBUG: https://alxsourcingstorageprod.table.core.windows.net:443 "DELETE /ScrapedListings(PartitionKey='bca.uk',RowKey='YY59%2520VRZ') HTTP/1.1" 204 0
      2026-03-30 11:09:57 [scrapy.extensions.feedexport] INFO: Stored jsonlines feed (1455 items) in: file:///var/lib/scrapyd/items/sourcing_v2/bca.uk/2026-03-30T10_28_25.jl
      2026-03-30 11:09:57 [scrapy.statscollectors] INFO: Dumping Scrapy stats:
      {'downloader/exception_count': 17,
       'downloader/exception_type_count/scrapy.core.downloader.handlers.http11.TunnelError': 17,
       'downloader/request_bytes': 17090473,
       'downloader/request_count': 20157,
       'downloader/request_method_count/GET': 20157,
       'downloader/response_bytes': 851395033,
       'downloader/response_count': 20140,
       'downloader/response_status_count/200': 10163,
       'downloader/response_status_count/302': 9977,
       'elapsed_time_seconds': 2484.855176,
       'feedexport/success_count/FileFeedStorage': 1,
       'finish_reason': 'finished',
       'finish_time': datetime.datetime(2026, 3, 30, 11, 9, 57, 410335, tzinfo=datetime.timezone.utc),
       'item_dropped_count': 28,
       'item_dropped_reasons_count/DropItem': 28,
       'item_scraped_count': 1455,
       'log_count/DEBUG': 131399,
       'log_count/ERROR': 142,
       'log_count/INFO': 106659,
       'log_count/WARNING': 387,
       'memusage/max': 215912448,
       'memusage/startup': 150077440,
       'photo_download_count': 9124,
       'pipeline/dropped_expired': 28,
       'playwright/context_count': 2,
       'playwright/context_count/max_concurrent': 1,
       'playwright/context_count/persistent/False': 2,
       'playwright/context_count/remote/False': 2,
       'playwright/page_count': 0,
       'request_depth_max': 128,
       'response_received_count': 1039,
       'scheduler/dequeued': 20157,
       'scheduler/dequeued/memory': 20157,
       'scheduler/enqueued': 20157,
       'scheduler/enqueued/memory': 20157,
       'scrape_type/new': 986,
       'scrape_type/price_update': 498,
       'scrape_type/skipped': 11008,
       'scrapy-zyte-api/429': 0,
       'scrapy-zyte-api/attempts': 130,
       'scrapy-zyte-api/error_ratio': 0.007692307692307693,
       'scrapy-zyte-api/errors': 1,
       "scrapy-zyte-api/exception_types/<class 'aiohttp.client_exceptions.ClientConnectorError'>": 1,
       'scrapy-zyte-api/fatal_errors': 0,
       'scrapy-zyte-api/mean_connection_seconds': 5.762903231997476,
       'scrapy-zyte-api/mean_response_seconds': 6.174709865635681,
       'scrapy-zyte-api/processed': 129,
       'scrapy-zyte-api/request_args/actions': 1,
       'scrapy-zyte-api/request_args/browserHtml': 1,
       'scrapy-zyte-api/request_args/customHttpRequestHeaders': 128,
       'scrapy-zyte-api/request_args/experimental.requestCookies': 128,
       'scrapy-zyte-api/request_args/experimental.responseCookies': 129,
       'scrapy-zyte-api/request_args/httpResponseBody': 128,
       'scrapy-zyte-api/request_args/httpResponseHeaders': 128,
       'scrapy-zyte-api/request_args/sessionContext': 1,
       'scrapy-zyte-api/request_args/url': 129,
       'scrapy-zyte-api/status_codes/0': 1,
       'scrapy-zyte-api/status_codes/200': 129,
       'scrapy-zyte-api/success': 129,
       'scrapy-zyte-api/success_ratio': 1.0,
       'scrapy-zyte-api/throttle_ratio': 0.0,
       'source/items_encountered': 12623,
       'spider_exceptions/KeyError': 21,
       'spider_exceptions/ValueError': 2,
       'start_time': datetime.datetime(2026, 3, 30, 10, 28, 32, 555159, tzinfo=datetime.timezone.utc)}
      2026-03-30 11:09:57 [scrapy.core.engine] INFO: Spider closed (finished)
    • Log

      /1/log/utf8/sourcing_v2/bca.uk/2026-03-30T10_28_25/?job_finished=True&ui=mobile

    • Source

      http://scrapyd-0:6800/logs/sourcing_v2/bca.uk/2026-03-30T10_28_25.log

  • sourcelog
    last update time2026-03-30 11:09:57
    last update timestamp1774868997
    downloader / exception count17
    downloader / exception type count / scrapy.core.downloader.handlers.http11.TunnelError17
    downloader / request bytes17090473
    downloader / request count20157
    downloader / request method count / GET20157
    downloader / response bytes851395033
    downloader / response count20140
    downloader / response status count / 20010163
    downloader / response status count / 3029977
    elapsed time seconds2484.855176
    feedexport / success count / FileFeedStorage1
    finish reasonfinished
    finish timedatetime.datetime(2026, 3, 30, 11, 9, 57, 410335, tzinfo=datetime.timezone.utc)
    item dropped count28
    item dropped reasons count / DropItem28
    item scraped count1455
    log count / DEBUG131399
    log count / ERROR142
    log count / INFO106659
    log count / WARNING387
    memusage / max215912448
    memusage / startup150077440
    photo download count9124
    pipeline / dropped expired28
    playwright / context count2
    playwright / context count / max concurrent1
    playwright / context count / persistent / False2
    playwright / context count / remote / False2
    playwright / page count0
    request depth max128
    response received count1039
    scheduler / dequeued20157
    scheduler / dequeued / memory20157
    scheduler / enqueued20157
    scheduler / enqueued / memory20157
    scrape type / new986
    scrape type / price update498
    scrape type / skipped11008
    scrapy-zyte-api / 4290
    scrapy-zyte-api / attempts130
    scrapy-zyte-api / error ratio0.007692307692307693
    scrapy-zyte-api / errors1
    scrapy-zyte-api / exception types / <class aiohttp.client exceptions.ClientConnectorError >1
    scrapy-zyte-api / fatal errors0
    scrapy-zyte-api / mean connection seconds5.762903231997476
    scrapy-zyte-api / mean response seconds6.174709865635681
    scrapy-zyte-api / processed129
    scrapy-zyte-api / request args / actions1
    scrapy-zyte-api / request args / browserHtml1
    scrapy-zyte-api / request args / customHttpRequestHeaders128
    scrapy-zyte-api / request args / experimental.requestCookies128
    scrapy-zyte-api / request args / experimental.responseCookies129
    scrapy-zyte-api / request args / httpResponseBody128
    scrapy-zyte-api / request args / httpResponseHeaders128
    scrapy-zyte-api / request args / sessionContext1
    scrapy-zyte-api / request args / url129
    scrapy-zyte-api / status codes / 01
    scrapy-zyte-api / status codes / 200129
    scrapy-zyte-api / success129
    scrapy-zyte-api / success ratio1.0
    scrapy-zyte-api / throttle ratio0.0
    source / items encountered12623
    spider exceptions / KeyError21
    spider exceptions / ValueError2
    start timedatetime.datetime(2026, 3, 30, 10, 28, 32, 555159, tzinfo=datetime.timezone.utc)