This is a scraper that runs on Morph. To get started see the documentation

Contributors woodbine blablupcom

Last run completed successfully .

Console output of last run

Injecting configuration and compiling...  -----> Python app detected  ! The latest version of Python 2 is python-2.7.14 (you are using python-2.7.9, which is unsupported).  ! We recommend upgrading by specifying the latest version (python-2.7.14).  Learn More: https://devcenter.heroku.com/articles/python-runtimes -----> Installing python-2.7.9 -----> Installing pip -----> Installing requirements with pip  DEPRECATION: Python 2.7 reached the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 is no longer maintained. pip 21.0 will drop support for Python 2.7 in January 2021. More details about Python 2 support in pip can be found at https://pip.pypa.io/en/latest/development/release-process/#python-2-support pip 21.0 will remove support for this functionality.  Collecting beautifulsoup4==4.2.0  Downloading beautifulsoup4-4.2.0.tar.gz (63 kB)  Obtaining scraperwiki from git+http://github.com/openaustralia/scraperwiki-python.git@morph_defaults#egg=scraperwiki (from -r /tmp/build/requirements.txt (line 10))  Cloning http://github.com/openaustralia/scraperwiki-python.git (to revision morph_defaults) to /app/.heroku/src/scraperwiki  Running command git clone -q http://github.com/openaustralia/scraperwiki-python.git /app/.heroku/src/scraperwiki  Running command git checkout -b morph_defaults --track origin/morph_defaults  Switched to a new branch 'morph_defaults'  Branch morph_defaults set up to track remote branch morph_defaults from origin.  Collecting lxml==3.4.4  Downloading lxml-3.4.4.tar.gz (3.5 MB)  Collecting cssselect==0.9.1  Downloading cssselect-0.9.1.tar.gz (32 kB)  Collecting dumptruck>=0.1.2  Downloading dumptruck-0.1.6.tar.gz (15 kB)  Collecting requests  Downloading requests-2.25.1-py2.py3-none-any.whl (61 kB)  Collecting chardet<5,>=3.0.2  Downloading chardet-4.0.0-py2.py3-none-any.whl (178 kB)  Collecting certifi>=2017.4.17  Downloading certifi-2020.12.5-py2.py3-none-any.whl (147 kB)  Collecting urllib3<1.27,>=1.21.1  Downloading urllib3-1.26.4-py2.py3-none-any.whl (153 kB)  Collecting idna<3,>=2.5  Downloading idna-2.10-py2.py3-none-any.whl (58 kB)  Building wheels for collected packages: beautifulsoup4, lxml, cssselect, dumptruck  Building wheel for beautifulsoup4 (setup.py): started  Building wheel for beautifulsoup4 (setup.py): finished with status 'done'  Created wheel for beautifulsoup4: filename=beautifulsoup4-4.2.0-py2-none-any.whl size=71804 sha256=1925037238becaedeaf8445f1fbbf9fae456a0676b4fa63bf7e0133cbd3c21a8  Stored in directory: /tmp/pip-ephem-wheel-cache-0wmpIZ/wheels/dc/69/36/2edde9ec10080447fc56d1a3d1235ddba3dd4b0dc9fff29134  Building wheel for lxml (setup.py): started  Building wheel for lxml (setup.py): still running...  Building wheel for lxml (setup.py): finished with status 'done'  Created wheel for lxml: filename=lxml-3.4.4-cp27-cp27m-linux_x86_64.whl size=2989845 sha256=037eabb0c16d2bb383c7a76f3809caea0c97187c35a0eda32c1a1ffb57c50a5a  Stored in directory: /tmp/pip-ephem-wheel-cache-0wmpIZ/wheels/d6/de/81/11ae6edd05c75aac677e67dd154c85da758ba6f3e8e80e962e  Building wheel for cssselect (setup.py): started  Building wheel for cssselect (setup.py): finished with status 'done'  Created wheel for cssselect: filename=cssselect-0.9.1-py2-none-any.whl size=26993 sha256=84114297c20140c0f7b2aab831e26fde134d0681f59ed19c17790e02ffea1cd4  Stored in directory: /tmp/pip-ephem-wheel-cache-0wmpIZ/wheels/85/fe/00/b94036d8583cec9791d8cda24c184f2d2ac1397822f7f0e8d4  Building wheel for dumptruck (setup.py): started  Building wheel for dumptruck (setup.py): finished with status 'done'  Created wheel for dumptruck: filename=dumptruck-0.1.6-py2-none-any.whl size=11844 sha256=33be984358cccb3596869b684a8703a0cfd7fa8ea411414fc41a2f0955873e31  Stored in directory: /tmp/pip-ephem-wheel-cache-0wmpIZ/wheels/dc/75/e9/1e61c4080c73e7bda99614549591f83b53bcc2d682f26fce62  Successfully built beautifulsoup4 lxml cssselect dumptruck  Installing collected packages: beautifulsoup4, dumptruck, chardet, certifi, urllib3, idna, requests, scraperwiki, lxml, cssselect  Running setup.py develop for scraperwiki  Successfully installed beautifulsoup4-4.2.0 certifi-2020.12.5 chardet-4.0.0 cssselect-0.9.1 dumptruck-0.1.6 idna-2.10 lxml-3.4.4 requests-2.25.1 scraperwiki urllib3-1.26.4 DEPRECATION: Python 2.7 reached the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 is no longer maintained. pip 21.0 will drop support for Python 2.7 in January 2021. More details about Python 2 support in pip can be found at https://pip.pypa.io/en/latest/development/release-process/#python-2-support pip 21.0 will remove support for this functionality.    -----> Discovering process types  Procfile declares types -> scraper Injecting scraper and running... E1821_WCC_gov_2019_12 E1821_WCC_gov_2019_11 E1821_WCC_gov_2019_10 E1821_WCC_gov_2019_09 E1821_WCC_gov_2019_08 E1821_WCC_gov_2019_07 E1821_WCC_gov_2019_06 E1821_WCC_gov_2019_05 E1821_WCC_gov_2019_04 E1821_WCC_gov_2019_03 E1821_WCC_gov_2019_02 E1821_WCC_gov_2019_01 E1821_WCC_gov_2018_12 E1821_WCC_gov_2018_11 E1821_WCC_gov_2018_10 E1821_WCC_gov_2018_09 E1821_WCC_gov_2018_08 E1821_WCC_gov_2018_07 E1821_WCC_gov_2018_06 E1821_WCC_gov_2018_05 E1821_WCC_gov_2018_04 E1821_WCC_gov_2018_03 E1821_WCC_gov_2018_02 E1821_WCC_gov_2018_01 E1821_WCC_gov_2017_12 E1821_WCC_gov_2017_11 E1821_WCC_gov_2017_10 E1821_WCC_gov_2017_09 E1821_WCC_gov_2017_08 E1821_WCC_gov_2017_07 E1821_WCC_gov_2017_06 E1821_WCC_gov_2017_05 E1821_WCC_gov_2017_03 E1821_WCC_gov_2017_04 E1821_WCC_gov_2017_02 E1821_WCC_gov_2017_01 E1821_WCC_gov_2016_12 E1821_WCC_gov_2016_11 E1821_WCC_gov_2016_10 E1821_WCC_gov_2016_09 E1821_WCC_gov_2016_08 E1821_WCC_gov_2016_07 E1821_WCC_gov_2016_06 E1821_WCC_gov_2016_05 E1821_WCC_gov_2016_04 E1821_WCC_gov_2016_03 E1821_WCC_gov_2016_02 E1821_WCC_gov_2016_01 E1821_WCC_gov_2015_12 E1821_WCC_gov_2015_11 E1821_WCC_gov_2015_10 E1821_WCC_gov_2015_09 E1821_WCC_gov_2015_08 E1821_WCC_gov_2015_07 E1821_WCC_gov_2015_06 E1821_WCC_gov_2015_05 E1821_WCC_gov_2015_04 E1821_WCC_gov_2015_03 E1821_WCC_gov_2015_02 E1821_WCC_gov_2015_01 E1821_WCC_gov_2014_12 E1821_WCC_gov_2014_11 E1821_WCC_gov_2014_10 E1821_WCC_gov_2014_09 E1821_WCC_gov_2014_08 E1821_WCC_gov_2014_07 E1821_WCC_gov_2014_06 E1821_WCC_gov_2014_05 E1821_WCC_gov_2014_04 E1821_WCC_gov_2014_03 E1821_WCC_gov_2014_02 E1821_WCC_gov_2014_01 E1821_WCC_gov_2013_12 E1821_WCC_gov_2013_11 E1821_WCC_gov_2013_10 E1821_WCC_gov_2013_09 E1821_WCC_gov_2013_08 E1821_WCC_gov_2013_07 E1821_WCC_gov_2013_06 E1821_WCC_gov_2013_05 E1821_WCC_gov_2013_04 E1821_WCC_gov_2013_03 E1821_WCC_gov_2013_02 E1821_WCC_gov_2013_01 E1821_WCC_gov_2012_12 E1821_WCC_gov_2012_11 E1821_WCC_gov_2012_10 E1821_WCC_gov_2012_09 E1821_WCC_gov_2012_08 E1821_WCC_gov_2012_07 E1821_WCC_gov_2012_06 E1821_WCC_gov_2012_05 E1821_WCC_gov_2012_04 E1821_WCC_gov_2012_03 E1821_WCC_gov_2012_02 E1821_WCC_gov_2012_01 E1821_WCC_gov_2011_12 E1821_WCC_gov_2011_11 E1821_WCC_gov_2011_10 E1821_WCC_gov_2011_09 E1821_WCC_gov_2011_08 E1821_WCC_gov_2011_07 E1821_WCC_gov_2011_06 E1821_WCC_gov_2011_05 E1821_WCC_gov_2011_04 E1821_WCC_gov_2011_03 E1821_WCC_gov_2011_02 E1821_WCC_gov_2011_01 E1821_WCC_gov_2010_12 E1821_WCC_gov_2010_11 E1821_WCC_gov_2010_10 E1821_WCC_gov_2010_09 E1821_WCC_gov_2010_08 E1821_WCC_gov_2010_07 E1821_WCC_gov_2010_06 E1821_WCC_gov_2010_05 E1821_WCC_gov_2010_04

Data

Downloaded 863 times by SimKennedy MikeRalphson woodbine

To download data sign in with GitHub

Download table (as CSV) Download SQLite database (102 KB) Use the API

rows 10 / 253

Statistics

Average successful run time: 8 minutes

Total run time: 26 days

Total cpu time used: about 2 hours

Total disk space used: 140 KB

History

  • Auto ran revision 50cff8fa and completed successfully .
    117 records added, 117 records removed in the database
  • Auto ran revision 50cff8fa and completed successfully .
    117 records added, 117 records removed in the database
  • Auto ran revision 50cff8fa and completed successfully .
    117 records added, 117 records removed in the database
  • Auto ran revision 50cff8fa and completed successfully .
    117 records added, 117 records removed in the database
  • Auto ran revision 50cff8fa and completed successfully .
    117 records added, 117 records removed in the database
  • ...
  • Created on morph.io

Show complete history

Scraper code

Python

sp_E1821_WCC_gov / scraper.py