howawong / hong_kong_dioxin_level_summary

Hong Kong Summary of Ambient Dioxin Level


This is a scraper that runs on Morph. To get started see the documentation

Contributors howawong

Last run completed successfully .

Console output of last run

Injecting configuration and compiling...  -----> Python app detected  ! The latest version of Python 2 is python-2.7.14 (you are using python-2.7.9, which is unsupported).  ! We recommend upgrading by specifying the latest version (python-2.7.14).  Learn More: https://devcenter.heroku.com/articles/python-runtimes -----> Installing python-2.7.9 -----> Installing pip -----> Installing requirements with pip  DEPRECATION: Python 2.7 reached the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 is no longer maintained. pip 21.0 will drop support for Python 2.7 in January 2021. More details about Python 2 support in pip can be found at https://pip.pypa.io/en/latest/development/release-process/#python-2-support pip 21.0 will remove support for this functionality.  Obtaining scraperwiki from git+http://github.com/openaustralia/scraperwiki-python.git@morph_defaults#egg=scraperwiki (from -r /tmp/build/requirements.txt (line 6))  Cloning http://github.com/openaustralia/scraperwiki-python.git (to revision morph_defaults) to /app/.heroku/src/scraperwiki  Running command git clone -q http://github.com/openaustralia/scraperwiki-python.git /app/.heroku/src/scraperwiki  Running command git checkout -b morph_defaults --track origin/morph_defaults  Switched to a new branch 'morph_defaults'  Branch morph_defaults set up to track remote branch morph_defaults from origin.  Collecting lxml==3.4.4  Downloading lxml-3.4.4.tar.gz (3.5 MB)  Collecting cssselect==0.9.1  Downloading cssselect-0.9.1.tar.gz (32 kB)  Collecting Scrapy==1.0.3  Downloading Scrapy-1.0.3-py2-none-any.whl (290 kB)  Collecting dumptruck>=0.1.2  Downloading dumptruck-0.1.6.tar.gz (15 kB)  Collecting requests  Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB)  Collecting service-identity  Downloading service_identity-21.1.0-py2.py3-none-any.whl (12 kB)  Collecting Twisted>=10.0.0  Downloading Twisted-20.3.0-cp27-cp27m-manylinux1_x86_64.whl (3.2 MB)  Collecting queuelib  Downloading queuelib-1.6.1-py2.py3-none-any.whl (12 kB)  Collecting w3lib>=1.8.0  Downloading w3lib-1.22.0-py2.py3-none-any.whl (20 kB)  Collecting pyOpenSSL  Downloading pyOpenSSL-21.0.0-py2.py3-none-any.whl (55 kB)  Collecting six>=1.5.2  Downloading six-1.16.0-py2.py3-none-any.whl (11 kB)  Collecting idna<3,>=2.5; python_version < "3"  Downloading idna-2.10-py2.py3-none-any.whl (58 kB)  Collecting certifi>=2017.4.17  Downloading certifi-2021.10.8-py2.py3-none-any.whl (149 kB)  Collecting chardet<5,>=3.0.2; python_version < "3"  Downloading chardet-4.0.0-py2.py3-none-any.whl (178 kB)  Collecting urllib3<1.27,>=1.21.1  Downloading urllib3-1.26.9-py2.py3-none-any.whl (138 kB)  Collecting pyasn1  Downloading pyasn1-0.4.8-py2.py3-none-any.whl (77 kB)  Collecting attrs>=19.1.0  Downloading attrs-21.4.0-py2.py3-none-any.whl (60 kB)  Collecting pyasn1-modules  Downloading pyasn1_modules-0.2.8-py2.py3-none-any.whl (155 kB)  Collecting ipaddress; python_version < "3.3"  Downloading ipaddress-1.0.23-py2.py3-none-any.whl (18 kB)  Collecting cryptography  Downloading cryptography-3.3.2-cp27-cp27m-manylinux2010_x86_64.whl (2.6 MB)  Collecting incremental>=16.10.1  Downloading incremental-21.3.0-py2.py3-none-any.whl (15 kB)  Collecting Automat>=0.3.0  Downloading Automat-20.2.0-py2.py3-none-any.whl (31 kB)  Collecting zope.interface>=4.4.2  Downloading zope.interface-5.4.0-cp27-cp27m-manylinux2010_x86_64.whl (247 kB)  Collecting hyperlink>=17.1.1  Downloading hyperlink-21.0.0-py2.py3-none-any.whl (74 kB)  Collecting constantly>=15.1  Downloading constantly-15.1.0-py2.py3-none-any.whl (7.9 kB)  Collecting PyHamcrest!=1.10.0,>=1.9.0  Downloading PyHamcrest-1.10.1.tar.gz (43 kB)  Collecting enum34; python_version < "3"  Downloading enum34-1.1.10-py2-none-any.whl (11 kB)  Collecting cffi>=1.12  Downloading cffi-1.15.0-cp27-cp27m-manylinux1_x86_64.whl (393 kB)  Collecting typing; python_version < "3.5"  Downloading typing-3.10.0.0-py2-none-any.whl (26 kB)  Collecting pycparser  Downloading pycparser-2.21-py2.py3-none-any.whl (118 kB)  Building wheels for collected packages: lxml, cssselect, dumptruck, PyHamcrest  Building wheel for lxml (setup.py): started  Building wheel for lxml (setup.py): still running...  Building wheel for lxml (setup.py): finished with status 'done'  Created wheel for lxml: filename=lxml-3.4.4-cp27-cp27m-linux_x86_64.whl size=2989859 sha256=d81da08df57fc5e36ae665a22e0e0711c955b26072977e928c86673af75b8cf4  Stored in directory: /tmp/pip-ephem-wheel-cache-0cbA1H/wheels/d6/de/81/11ae6edd05c75aac677e67dd154c85da758ba6f3e8e80e962e  Building wheel for cssselect (setup.py): started  Building wheel for cssselect (setup.py): finished with status 'done'  Created wheel for cssselect: filename=cssselect-0.9.1-py2-none-any.whl size=26992 sha256=f49d91b372893cf61ddfb922e95fd4f4b2fe62c4d5044fe3b831828a8048a83e  Stored in directory: /tmp/pip-ephem-wheel-cache-0cbA1H/wheels/85/fe/00/b94036d8583cec9791d8cda24c184f2d2ac1397822f7f0e8d4  Building wheel for dumptruck (setup.py): started  Building wheel for dumptruck (setup.py): finished with status 'done'  Created wheel for dumptruck: filename=dumptruck-0.1.6-py2-none-any.whl size=11844 sha256=f818aba20be09430b7a671a922d13643d037fc480be615cd09c0c39c9a15f753  Stored in directory: /tmp/pip-ephem-wheel-cache-0cbA1H/wheels/dc/75/e9/1e61c4080c73e7bda99614549591f83b53bcc2d682f26fce62  Building wheel for PyHamcrest (setup.py): started  Building wheel for PyHamcrest (setup.py): finished with status 'done'  Created wheel for PyHamcrest: filename=PyHamcrest-1.10.1-py2-none-any.whl size=48898 sha256=b0dd9084ca8ac179114a638f80da74ed0cdad8136d39a278a91858cab09f6807  Stored in directory: /tmp/pip-ephem-wheel-cache-0cbA1H/wheels/f5/8c/e2/f0cea19d340270166bbfd4a2e9d8b8c132e26ef7e1376a0890  Successfully built lxml cssselect dumptruck PyHamcrest  Installing collected packages: dumptruck, idna, certifi, chardet, urllib3, requests, scraperwiki, lxml, cssselect, pyasn1, attrs, pyasn1-modules, ipaddress, six, enum34, pycparser, cffi, cryptography, service-identity, incremental, Automat, zope.interface, typing, hyperlink, constantly, PyHamcrest, Twisted, queuelib, w3lib, pyOpenSSL, Scrapy  Running setup.py develop for scraperwiki  Successfully installed Automat-20.2.0 PyHamcrest-1.10.1 Scrapy-1.0.3 Twisted-20.3.0 attrs-21.4.0 certifi-2021.10.8 cffi-1.15.0 chardet-4.0.0 constantly-15.1.0 cryptography-3.3.2 cssselect-0.9.1 dumptruck-0.1.6 enum34-1.1.10 hyperlink-21.0.0 idna-2.10 incremental-21.3.0 ipaddress-1.0.23 lxml-3.4.4 pyOpenSSL-21.0.0 pyasn1-0.4.8 pyasn1-modules-0.2.8 pycparser-2.21 queuelib-1.6.1 requests-2.27.1 scraperwiki service-identity-21.1.0 six-1.16.0 typing-3.10.0.0 urllib3-1.26.9 w3lib-1.22.0 zope.interface-5.4.0 DEPRECATION: Python 2.7 reached the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 is no longer maintained. pip 21.0 will drop support for Python 2.7 in January 2021. More details about Python 2 support in pip can be found at https://pip.pypa.io/en/latest/development/release-process/#python-2-support pip 21.0 will remove support for this functionality.    -----> Discovering process types  Procfile declares types -> scraper Injecting scraper and running... /app/.heroku/python/lib/python2.7/site-packages/OpenSSL/crypto.py:14: CryptographyDeprecationWarning: Python 2 is no longer supported by the Python core team. Support for it is now deprecated in cryptography, and will be removed in the next release. from cryptography import utils, x509 2022-05-23 16:34:11 [scrapy] INFO: Scrapy 1.0.3 started (bot: scrapybot) 2022-05-23 16:34:11 [scrapy] INFO: Optional features available: ssl, http11 2022-05-23 16:34:11 [scrapy] INFO: Overridden settings: {} 2022-05-23 16:34:11 [scrapy] INFO: Enabled extensions: LogStats, CoreStats, CloseSpider, TelnetConsole, SpiderState Unhandled error in Deferred: 2022-05-23 16:34:11 [twisted] CRITICAL: Unhandled error in Deferred: Traceback (most recent call last): File "scraper.py", line 59, in <module> process.crawl(DioxinSpider) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/crawler.py", line 153, in crawl d = crawler.crawl(*args, **kwargs) File "/app/.heroku/python/lib/python2.7/site-packages/twisted/internet/defer.py", line 1613, in unwindGenerator return _cancellableInlineCallbacks(gen) File "/app/.heroku/python/lib/python2.7/site-packages/twisted/internet/defer.py", line 1529, in _cancellableInlineCallbacks _inlineCallbacks(None, g, status) --- <exception caught here> --- File "/app/.heroku/python/lib/python2.7/site-packages/twisted/internet/defer.py", line 1418, in _inlineCallbacks result = g.send(result) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/crawler.py", line 71, in crawl self.engine = self._create_engine() File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/crawler.py", line 83, in _create_engine return ExecutionEngine(self, lambda _: self.stop()) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/core/engine.py", line 64, in __init__ self.scheduler_cls = load_object(self.settings['SCHEDULER']) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/utils/misc.py", line 44, in load_object mod = import_module(module) File "/app/.heroku/python/lib/python2.7/importlib/__init__.py", line 37, in import_module __import__(name) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/core/scheduler.py", line 6, in <module> from queuelib import PriorityQueue File "/app/.heroku/python/lib/python2.7/site-packages/queuelib/__init__.py", line 1, in <module> from queuelib.queue import FifoDiskQueue, LifoDiskQueue File "/app/.heroku/python/lib/python2.7/site-packages/queuelib/queue.py", line 7, in <module> from contextlib import suppress exceptions.ImportError: cannot import name suppress 2022-05-23 16:34:11 [twisted] CRITICAL: Traceback (most recent call last): File "/app/.heroku/python/lib/python2.7/site-packages/twisted/internet/defer.py", line 1418, in _inlineCallbacks result = g.send(result) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/crawler.py", line 71, in crawl self.engine = self._create_engine() File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/crawler.py", line 83, in _create_engine return ExecutionEngine(self, lambda _: self.stop()) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/core/engine.py", line 64, in __init__ self.scheduler_cls = load_object(self.settings['SCHEDULER']) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/utils/misc.py", line 44, in load_object mod = import_module(module) File "/app/.heroku/python/lib/python2.7/importlib/__init__.py", line 37, in import_module __import__(name) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/core/scheduler.py", line 6, in <module> from queuelib import PriorityQueue File "/app/.heroku/python/lib/python2.7/site-packages/queuelib/__init__.py", line 1, in <module> from queuelib.queue import FifoDiskQueue, LifoDiskQueue File "/app/.heroku/python/lib/python2.7/site-packages/queuelib/queue.py", line 7, in <module> from contextlib import suppress ImportError: cannot import name suppress

Statistics

Average successful run time: 2 minutes

Total run time: 23 days

Total cpu time used: 36 minutes

Total disk space used: 44.2 KB

History

  • Auto ran revision 1eb02b48 and completed successfully .
    nothing changed in the database
  • Auto ran revision 1eb02b48 and completed successfully .
    nothing changed in the database
  • Auto ran revision 1eb02b48 and completed successfully .
    nothing changed in the database
  • Auto ran revision 1eb02b48 and completed successfully .
    nothing changed in the database
  • Auto ran revision 1eb02b48 and completed successfully .
    nothing changed in the database
  • ...
  • Created on morph.io

Show complete history