howawong / hong_kong_current_consultation_pages

Current Consultation Papers from http://www.gov.hk/en/residents/government/publication/consultation/current.htm


This is a scraper that runs on Morph. To get started see the documentation

Contributors howawong

Last run completed successfully .

Console output of last run

Injecting configuration and compiling...  -----> Python app detected  ! The latest version of Python 2 is python-2.7.14 (you are using python-2.7.9, which is unsupported).  ! We recommend upgrading by specifying the latest version (python-2.7.14).  Learn More: https://devcenter.heroku.com/articles/python-runtimes -----> Installing python-2.7.9 -----> Installing pip -----> Installing requirements with pip  DEPRECATION: Python 2.7 reached the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 is no longer maintained. pip 21.0 will drop support for Python 2.7 in January 2021. More details about Python 2 support in pip can be found at https://pip.pypa.io/en/latest/development/release-process/#python-2-support pip 21.0 will remove support for this functionality.  Obtaining scraperwiki from git+http://github.com/openaustralia/scraperwiki-python.git@morph_defaults#egg=scraperwiki (from -r /tmp/build/requirements.txt (line 6))  Cloning http://github.com/openaustralia/scraperwiki-python.git (to revision morph_defaults) to /app/.heroku/src/scraperwiki  Running command git clone -q http://github.com/openaustralia/scraperwiki-python.git /app/.heroku/src/scraperwiki  Running command git checkout -b morph_defaults --track origin/morph_defaults  Switched to a new branch 'morph_defaults'  Branch morph_defaults set up to track remote branch morph_defaults from origin.  Collecting lxml==3.4.4  Downloading lxml-3.4.4.tar.gz (3.5 MB)  Collecting cssselect==0.9.1  Downloading cssselect-0.9.1.tar.gz (32 kB)  Collecting Scrapy==1.0.3  Downloading Scrapy-1.0.3-py2-none-any.whl (290 kB)  Collecting dumptruck>=0.1.2  Downloading dumptruck-0.1.6.tar.gz (15 kB)  Collecting requests  Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB)  Collecting service-identity  Downloading service_identity-21.1.0-py2.py3-none-any.whl (12 kB)  Collecting Twisted>=10.0.0  Downloading Twisted-20.3.0-cp27-cp27m-manylinux1_x86_64.whl (3.2 MB)  Collecting queuelib  Downloading queuelib-1.6.1-py2.py3-none-any.whl (12 kB)  Collecting w3lib>=1.8.0  Downloading w3lib-1.22.0-py2.py3-none-any.whl (20 kB)  Collecting pyOpenSSL  Downloading pyOpenSSL-21.0.0-py2.py3-none-any.whl (55 kB)  Collecting six>=1.5.2  Downloading six-1.16.0-py2.py3-none-any.whl (11 kB)  Collecting idna<3,>=2.5; python_version < "3"  Downloading idna-2.10-py2.py3-none-any.whl (58 kB)  Collecting certifi>=2017.4.17  Downloading certifi-2021.10.8-py2.py3-none-any.whl (149 kB)  Collecting chardet<5,>=3.0.2; python_version < "3"  Downloading chardet-4.0.0-py2.py3-none-any.whl (178 kB)  Collecting urllib3<1.27,>=1.21.1  Downloading urllib3-1.26.9-py2.py3-none-any.whl (138 kB)  Collecting pyasn1  Downloading pyasn1-0.4.8-py2.py3-none-any.whl (77 kB)  Collecting attrs>=19.1.0  Downloading attrs-21.4.0-py2.py3-none-any.whl (60 kB)  Collecting pyasn1-modules  Downloading pyasn1_modules-0.2.8-py2.py3-none-any.whl (155 kB)  Collecting ipaddress; python_version < "3.3"  Downloading ipaddress-1.0.23-py2.py3-none-any.whl (18 kB)  Collecting cryptography  Downloading cryptography-3.3.2-cp27-cp27m-manylinux2010_x86_64.whl (2.6 MB)  Collecting incremental>=16.10.1  Downloading incremental-21.3.0-py2.py3-none-any.whl (15 kB)  Collecting Automat>=0.3.0  Downloading Automat-20.2.0-py2.py3-none-any.whl (31 kB)  Collecting zope.interface>=4.4.2  Downloading zope.interface-5.4.0-cp27-cp27m-manylinux2010_x86_64.whl (247 kB)  Collecting hyperlink>=17.1.1  Downloading hyperlink-21.0.0-py2.py3-none-any.whl (74 kB)  Collecting constantly>=15.1  Downloading constantly-15.1.0-py2.py3-none-any.whl (7.9 kB)  Collecting PyHamcrest!=1.10.0,>=1.9.0  Downloading PyHamcrest-1.10.1.tar.gz (43 kB)  Collecting enum34; python_version < "3"  Downloading enum34-1.1.10-py2-none-any.whl (11 kB)  Collecting cffi>=1.12  Downloading cffi-1.15.0-cp27-cp27m-manylinux1_x86_64.whl (393 kB)  Collecting typing; python_version < "3.5"  Downloading typing-3.10.0.0-py2-none-any.whl (26 kB)  Collecting pycparser  Downloading pycparser-2.21-py2.py3-none-any.whl (118 kB)  Building wheels for collected packages: lxml, cssselect, dumptruck, PyHamcrest  Building wheel for lxml (setup.py): started  Building wheel for lxml (setup.py): still running...  Building wheel for lxml (setup.py): finished with status 'done'  Created wheel for lxml: filename=lxml-3.4.4-cp27-cp27m-linux_x86_64.whl size=2989844 sha256=0c2cffe0082be6e9197b597baffdf1cc466e3c34322d98e75ffa847b3b9f889d  Stored in directory: /tmp/pip-ephem-wheel-cache-vgJvCI/wheels/d6/de/81/11ae6edd05c75aac677e67dd154c85da758ba6f3e8e80e962e  Building wheel for cssselect (setup.py): started  Building wheel for cssselect (setup.py): finished with status 'done'  Created wheel for cssselect: filename=cssselect-0.9.1-py2-none-any.whl size=26992 sha256=072aee63873d24e1b820e664819f77bc0566a06732c81293be174364389f3cc9  Stored in directory: /tmp/pip-ephem-wheel-cache-vgJvCI/wheels/85/fe/00/b94036d8583cec9791d8cda24c184f2d2ac1397822f7f0e8d4  Building wheel for dumptruck (setup.py): started  Building wheel for dumptruck (setup.py): finished with status 'done'  Created wheel for dumptruck: filename=dumptruck-0.1.6-py2-none-any.whl size=11844 sha256=f058193eb322a38a146e0b3045b752be915561b434a55a6c64b1e6075ed68f4f  Stored in directory: /tmp/pip-ephem-wheel-cache-vgJvCI/wheels/dc/75/e9/1e61c4080c73e7bda99614549591f83b53bcc2d682f26fce62  Building wheel for PyHamcrest (setup.py): started  Building wheel for PyHamcrest (setup.py): finished with status 'done'  Created wheel for PyHamcrest: filename=PyHamcrest-1.10.1-py2-none-any.whl size=48898 sha256=555bf6501421242a317be1c01b1320ab0defac64b8c4e27a1fe4f12e14c828eb  Stored in directory: /tmp/pip-ephem-wheel-cache-vgJvCI/wheels/f5/8c/e2/f0cea19d340270166bbfd4a2e9d8b8c132e26ef7e1376a0890  Successfully built lxml cssselect dumptruck PyHamcrest  Installing collected packages: dumptruck, idna, certifi, chardet, urllib3, requests, scraperwiki, lxml, cssselect, pyasn1, attrs, pyasn1-modules, ipaddress, six, enum34, pycparser, cffi, cryptography, service-identity, incremental, Automat, zope.interface, typing, hyperlink, constantly, PyHamcrest, Twisted, queuelib, w3lib, pyOpenSSL, Scrapy  Running setup.py develop for scraperwiki  Successfully installed Automat-20.2.0 PyHamcrest-1.10.1 Scrapy-1.0.3 Twisted-20.3.0 attrs-21.4.0 certifi-2021.10.8 cffi-1.15.0 chardet-4.0.0 constantly-15.1.0 cryptography-3.3.2 cssselect-0.9.1 dumptruck-0.1.6 enum34-1.1.10 hyperlink-21.0.0 idna-2.10 incremental-21.3.0 ipaddress-1.0.23 lxml-3.4.4 pyOpenSSL-21.0.0 pyasn1-0.4.8 pyasn1-modules-0.2.8 pycparser-2.21 queuelib-1.6.1 requests-2.27.1 scraperwiki service-identity-21.1.0 six-1.16.0 typing-3.10.0.0 urllib3-1.26.9 w3lib-1.22.0 zope.interface-5.4.0 DEPRECATION: Python 2.7 reached the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 is no longer maintained. pip 21.0 will drop support for Python 2.7 in January 2021. More details about Python 2 support in pip can be found at https://pip.pypa.io/en/latest/development/release-process/#python-2-support pip 21.0 will remove support for this functionality.    -----> Discovering process types  Procfile declares types -> scraper Injecting scraper and running... /app/.heroku/python/lib/python2.7/site-packages/OpenSSL/crypto.py:14: CryptographyDeprecationWarning: Python 2 is no longer supported by the Python core team. Support for it is now deprecated in cryptography, and will be removed in the next release. from cryptography import utils, x509 2022-05-19 04:44:30 [scrapy] INFO: Scrapy 1.0.3 started (bot: scrapybot) 2022-05-19 04:44:30 [scrapy] INFO: Optional features available: http11, ssl 2022-05-19 04:44:30 [scrapy] INFO: Overridden settings: {} 2022-05-19 04:44:30 [scrapy] INFO: Enabled extensions: SpiderState, LogStats, TelnetConsole, CoreStats, CloseSpider Unhandled error in Deferred: 2022-05-19 04:44:30 [twisted] CRITICAL: Unhandled error in Deferred: Traceback (most recent call last): File "scraper.py", line 28, in <module> process.crawl(ConsultationSpider) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/crawler.py", line 153, in crawl d = crawler.crawl(*args, **kwargs) File "/app/.heroku/python/lib/python2.7/site-packages/twisted/internet/defer.py", line 1613, in unwindGenerator return _cancellableInlineCallbacks(gen) File "/app/.heroku/python/lib/python2.7/site-packages/twisted/internet/defer.py", line 1529, in _cancellableInlineCallbacks _inlineCallbacks(None, g, status) --- <exception caught here> --- File "/app/.heroku/python/lib/python2.7/site-packages/twisted/internet/defer.py", line 1418, in _inlineCallbacks result = g.send(result) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/crawler.py", line 71, in crawl self.engine = self._create_engine() File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/crawler.py", line 83, in _create_engine return ExecutionEngine(self, lambda _: self.stop()) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/core/engine.py", line 64, in __init__ self.scheduler_cls = load_object(self.settings['SCHEDULER']) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/utils/misc.py", line 44, in load_object mod = import_module(module) File "/app/.heroku/python/lib/python2.7/importlib/__init__.py", line 37, in import_module __import__(name) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/core/scheduler.py", line 6, in <module> from queuelib import PriorityQueue File "/app/.heroku/python/lib/python2.7/site-packages/queuelib/__init__.py", line 1, in <module> from queuelib.queue import FifoDiskQueue, LifoDiskQueue File "/app/.heroku/python/lib/python2.7/site-packages/queuelib/queue.py", line 7, in <module> from contextlib import suppress exceptions.ImportError: cannot import name suppress 2022-05-19 04:44:30 [twisted] CRITICAL: Traceback (most recent call last): File "/app/.heroku/python/lib/python2.7/site-packages/twisted/internet/defer.py", line 1418, in _inlineCallbacks result = g.send(result) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/crawler.py", line 71, in crawl self.engine = self._create_engine() File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/crawler.py", line 83, in _create_engine return ExecutionEngine(self, lambda _: self.stop()) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/core/engine.py", line 64, in __init__ self.scheduler_cls = load_object(self.settings['SCHEDULER']) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/utils/misc.py", line 44, in load_object mod = import_module(module) File "/app/.heroku/python/lib/python2.7/importlib/__init__.py", line 37, in import_module __import__(name) File "/app/.heroku/python/lib/python2.7/site-packages/scrapy/core/scheduler.py", line 6, in <module> from queuelib import PriorityQueue File "/app/.heroku/python/lib/python2.7/site-packages/queuelib/__init__.py", line 1, in <module> from queuelib.queue import FifoDiskQueue, LifoDiskQueue File "/app/.heroku/python/lib/python2.7/site-packages/queuelib/queue.py", line 7, in <module> from contextlib import suppress ImportError: cannot import name suppress

Data

Downloaded 142 times by howawong

To download data sign in with GitHub

Download table (as CSV) Download SQLite database (25 KB) Use the API

rows 10 / 64

link title lang date
Review of Statutory Minimum Wage Rate Public Consultation
en
2016-05-22T00:00:00+00:00
法定最低工資水平檢討公眾諮詢
tc
2016-05-22T00:00:00+00:00
法定最低工资水平检讨公众咨询
sc
2016-05-22T00:00:00+00:00
選舉委員會界別分組選舉活動建議指引公眾諮詢
tc
2016-06-09T00:00:00+00:00
香港將與格魯吉亞和馬爾代夫談判自由貿易協定
tc
2016-06-06T00:00:00+00:00
Public consultation on the Proposed Guidelines on Election-related Activities in respect of the Election Committee Subsector Elections
en
2016-06-09T00:00:00+00:00
Hong Kong to Negotiate Free Trade Agreements with Georgia and Maldives
en
2016-06-06T00:00:00+00:00
选举委员会界别分组选举活动建议指引公众咨询
sc
2016-06-09T00:00:00+00:00
香港将与格鲁吉亚和马尔代夫谈判自由贸易协定
sc
2016-06-06T00:00:00+00:00
《職業介紹所實務守則》草擬本諮詢
tc
2016-06-17T00:00:00+00:00

Statistics

Average successful run time: 2 minutes

Total run time: 20 days

Total cpu time used: 32 minutes

Total disk space used: 46.5 KB

History

  • Auto ran revision 053233cb and completed successfully .
    nothing changed in the database
  • Auto ran revision 053233cb and completed successfully .
    nothing changed in the database
  • Auto ran revision 053233cb and completed successfully .
    nothing changed in the database
  • Auto ran revision 053233cb and completed successfully .
    nothing changed in the database
  • Auto ran revision 053233cb and completed successfully .
    nothing changed in the database
  • ...
  • Created on morph.io

Show complete history