This is a scraper that runs on Morph. To get started see the documentation

Contributors woodbine blablupcom

Last run completed successfully .

Console output of last run

Injecting configuration and compiling...  -----> Python app detected  ! The latest version of Python 2 is python-2.7.14 (you are using python-2.7.9, which is unsupported).  ! We recommend upgrading by specifying the latest version (python-2.7.14).  Learn More: https://devcenter.heroku.com/articles/python-runtimes -----> Installing python-2.7.9 -----> Installing pip -----> Installing requirements with pip  DEPRECATION: Python 2.7 will reach the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 won't be maintained after that date. A future version of pip will drop support for Python 2.7. More details about Python 2 support in pip, can be found at https://pip.pypa.io/en/latest/development/release-process/#python-2-support  Obtaining scraperwiki from git+http://github.com/openaustralia/scraperwiki-python.git@morph_defaults#egg=scraperwiki (from -r /tmp/build/requirements.txt (line 1))  Cloning http://github.com/openaustralia/scraperwiki-python.git (to revision morph_defaults) to /app/.heroku/src/scraperwiki  Running command git clone -q http://github.com/openaustralia/scraperwiki-python.git /app/.heroku/src/scraperwiki  Running command git checkout -b morph_defaults --track origin/morph_defaults  Switched to a new branch 'morph_defaults'  Branch morph_defaults set up to track remote branch morph_defaults from origin.  Collecting lxml==3.4.4  Downloading https://files.pythonhosted.org/packages/63/c7/4f2a2a4ad6c6fa99b14be6b3c1cece9142e2d915aa7c43c908677afc8fa4/lxml-3.4.4.tar.gz (3.5MB)  Collecting cssselect==0.9.1  Downloading https://files.pythonhosted.org/packages/aa/e5/9ee1460d485b94a6d55732eb7ad5b6c084caf73dd6f9cb0bb7d2a78fafe8/cssselect-0.9.1.tar.gz  Collecting beautifulsoup4  Downloading https://files.pythonhosted.org/packages/f9/d9/183705a87492249b212d88eef740995f55076195bcf45ed59306c146e42d/beautifulsoup4-4.8.1-py2-none-any.whl (102kB)  Collecting dumptruck>=0.1.2  Downloading https://files.pythonhosted.org/packages/15/27/3330a343de80d6849545b6c7723f8c9a08b4b104de964ac366e7e6b318df/dumptruck-0.1.6.tar.gz  Collecting requests  Downloading https://files.pythonhosted.org/packages/51/bd/23c926cd341ea6b7dd0b2a00aba99ae0f828be89d72b2190f27c11d4b7fb/requests-2.22.0-py2.py3-none-any.whl (57kB)  Collecting soupsieve>=1.2  Downloading https://files.pythonhosted.org/packages/81/94/03c0f04471fc245d08d0a99f7946ac228ca98da4fa75796c507f61e688c2/soupsieve-1.9.5-py2.py3-none-any.whl  Collecting certifi>=2017.4.17  Downloading https://files.pythonhosted.org/packages/18/b0/8146a4f8dd402f60744fa380bc73ca47303cccf8b9190fd16a827281eac2/certifi-2019.9.11-py2.py3-none-any.whl (154kB)  Collecting urllib3!=1.25.0,!=1.25.1,<1.26,>=1.21.1  Downloading https://files.pythonhosted.org/packages/b4/40/a9837291310ee1ccc242ceb6ebfd9eb21539649f193a7c8c86ba15b98539/urllib3-1.25.7-py2.py3-none-any.whl (125kB)  Collecting idna<2.9,>=2.5  Downloading https://files.pythonhosted.org/packages/14/2c/cd551d81dbe15200be1cf41cd03869a46fe7226e7450af7a6545bfc474c9/idna-2.8-py2.py3-none-any.whl (58kB)  Collecting chardet<3.1.0,>=3.0.2  Downloading https://files.pythonhosted.org/packages/bc/a9/01ffebfb562e4274b6487b4bb1ddec7ca55ec7510b22e4c51f14098443b8/chardet-3.0.4-py2.py3-none-any.whl (133kB)  Collecting backports.functools-lru-cache; python_version < "3"  Downloading https://files.pythonhosted.org/packages/da/d1/080d2bb13773803648281a49e3918f65b31b7beebf009887a529357fd44a/backports.functools_lru_cache-1.6.1-py2.py3-none-any.whl  Building wheels for collected packages: lxml, cssselect, dumptruck  Building wheel for lxml (setup.py): started  Building wheel for lxml (setup.py): still running...  Building wheel for lxml (setup.py): finished with status 'done'  Created wheel for lxml: filename=lxml-3.4.4-cp27-cp27m-linux_x86_64.whl size=2989817 sha256=6c3e707f124b4c2d12aa66be038436ac02016339e45e1accaaa4108b1c65433c  Stored in directory: /tmp/pip-ephem-wheel-cache-nHkmxg/wheels/f6/df/7b/af9cace9baf95a6e4a2b5790e30da55fc780ddee598314d1ed  Building wheel for cssselect (setup.py): started  Building wheel for cssselect (setup.py): finished with status 'done'  Created wheel for cssselect: filename=cssselect-0.9.1-cp27-none-any.whl size=26994 sha256=a1b4ea4e7bfb1a514740c9cba796490cb7b63542462dfc794ff359a9a9c29386  Stored in directory: /tmp/pip-ephem-wheel-cache-nHkmxg/wheels/45/25/d7/5a3b06d22b1ffb616f868a74729a5a002bcc04d45109b4f223  Building wheel for dumptruck (setup.py): started  Building wheel for dumptruck (setup.py): finished with status 'done'  Created wheel for dumptruck: filename=dumptruck-0.1.6-cp27-none-any.whl size=11845 sha256=57f4854dca6e01060b4ddb1cf7b349ced1edeb64cd38a6b4784be2d6bee01692  Stored in directory: /tmp/pip-ephem-wheel-cache-nHkmxg/wheels/57/df/83/32654ae89119876c7a7db66829bbdb646caa151589dbaf226e  Successfully built lxml cssselect dumptruck  Installing collected packages: dumptruck, certifi, urllib3, idna, chardet, requests, scraperwiki, lxml, cssselect, backports.functools-lru-cache, soupsieve, beautifulsoup4  Running setup.py develop for scraperwiki  Successfully installed backports.functools-lru-cache-1.6.1 beautifulsoup4-4.8.1 certifi-2019.9.11 chardet-3.0.4 cssselect-0.9.1 dumptruck-0.1.6 idna-2.8 lxml-3.4.4 requests-2.22.0 scraperwiki soupsieve-1.9.5 urllib3-1.25.7 DEPRECATION: Python 2.7 will reach the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 won't be maintained after that date. A future version of pip will drop support for Python 2.7. More details about Python 2 support in pip, can be found at https://pip.pypa.io/en/latest/development/release-process/#python-2-support    -----> Discovering process types  Procfile declares types -> scraper Injecting scraper and running... E3720_WCC_gov_2018_07 E3720_WCC_gov_2011_04 E3720_WCC_gov_2012_04 E3720_WCC_gov_2013_04 E3720_WCC_gov_2014_04 E3720_WCC_gov_2015_04 E3720_WCC_gov_2016_04 E3720_WCC_gov_2017_04 E3720_WCC_gov_2018_04 E3720_WCC_gov_2019_04 E3720_WCC_gov_2011_08 E3720_WCC_gov_2012_08 E3720_WCC_gov_2013_08 E3720_WCC_gov_2014_08 E3720_WCC_gov_2015_08 E3720_WCC_gov_2016_08 E3720_WCC_gov_2017_08 E3720_WCC_gov_2018_08 E3720_WCC_gov_2019_08 E3720_WCC_gov_2010_12 E3720_WCC_gov_2011_12 E3720_WCC_gov_2011_12 E3720_WCC_gov_2012_12 E3720_WCC_gov_2013_12 E3720_WCC_gov_2014_12 E3720_WCC_gov_2015_12 E3720_WCC_gov_2016_12 E3720_WCC_gov_2017_12 E3720_WCC_gov_2018_12 E3720_WCC_gov_2011_02 E3720_WCC_gov_2012_02 E3720_WCC_gov_2012_02 E3720_WCC_gov_2013_02 E3720_WCC_gov_2014_02 E3720_WCC_gov_2015_02 E3720_WCC_gov_2016_02 E3720_WCC_gov_2017_02 E3720_WCC_gov_2018_02 E3720_WCC_gov_2019_02 E3720_WCC_gov_2011_01 E3720_WCC_gov_2012_01 E3720_WCC_gov_2012_01 E3720_WCC_gov_2013_01 E3720_WCC_gov_2014_01 E3720_WCC_gov_2015_01 E3720_WCC_gov_2016_01 E3720_WCC_gov_2017_01 E3720_WCC_gov_2018_01 E3720_WCC_gov_2019_01 E3720_WCC_gov_2011_07 E3720_WCC_gov_2012_07 E3720_WCC_gov_2013_07 E3720_WCC_gov_2014_07 E3720_WCC_gov_2014_07 E3720_WCC_gov_2015_07 E3720_WCC_gov_2016_07 E3720_WCC_gov_2017_07 E3720_WCC_gov_2019_07 E3720_WCC_gov_2011_06 E3720_WCC_gov_2012_06 E3720_WCC_gov_2013_06 E3720_WCC_gov_2014_06 E3720_WCC_gov_2015_06 E3720_WCC_gov_2016_06 E3720_WCC_gov_2017_06 E3720_WCC_gov_2018_06 E3720_WCC_gov_2019_06 E3720_WCC_gov_2011_03 E3720_WCC_gov_2012_03 E3720_WCC_gov_2012_03 E3720_WCC_gov_2013_03 E3720_WCC_gov_2014_03 E3720_WCC_gov_2015_03 E3720_WCC_gov_2016_03 E3720_WCC_gov_2017_03 E3720_WCC_gov_2018_03 E3720_WCC_gov_2019_03 E3720_WCC_gov_2011_05 E3720_WCC_gov_2012_05 E3720_WCC_gov_2013_05 E3720_WCC_gov_2014_05 E3720_WCC_gov_2015_05 E3720_WCC_gov_2016_05 E3720_WCC_gov_2017_05 E3720_WCC_gov_2018_05 E3720_WCC_gov_2019_05 E3720_WCC_gov_2010_11 E3720_WCC_gov_2011_11 E3720_WCC_gov_2011_11 E3720_WCC_gov_2012_11 E3720_WCC_gov_2013_11 E3720_WCC_gov_2014_11 E3720_WCC_gov_2015_11 E3720_WCC_gov_2016_11 E3720_WCC_gov_2017_11 E3720_WCC_gov_2018_11 E3720_WCC_gov_2011_10 E3720_WCC_gov_2011_10 E3720_WCC_gov_2012_10

Data

Downloaded 833 times by SimKennedy woodbine MikeRalphson blablupcom

To download data sign in with GitHub

Download table (as CSV) Download SQLite database (64 KB) Use the API

rows 10 / 212

d f l
2018-02-08 23:31:53.251748
E3720_WCC_gov_2015_02
2018-02-08 23:32:27.337050
E3720_WCC_gov_2014_07
https://s3-eu-west-1.amazonaws.com/opendata/supplierpayments/all-july-2014.csv
2018-11-26 02:56:21.216661
E3720_WCC_gov_2016_09
2018-11-26 02:56:24.384195
E3720_WCC_gov_2017_09
2019-02-04 10:11:44.285298
E3720_WCC_gov_2018_07
2019-02-04 10:11:46.296672
E3720_WCC_gov_2011_04
2019-02-04 10:11:48.467975
E3720_WCC_gov_2012_04
2019-02-04 10:11:50.421670
E3720_WCC_gov_2013_04
2019-02-04 10:11:52.671064
E3720_WCC_gov_2014_04
2019-02-04 10:11:54.774276
E3720_WCC_gov_2015_04

Statistics

Average successful run time: 9 minutes

Total run time: about 1 month

Total cpu time used: about 8 hours

Total disk space used: 102 KB

History

  • Auto ran revision 45d3ebf9 and completed successfully .
    98 records added, 98 records removed in the database
  • Auto ran revision 45d3ebf9 and completed successfully .
    98 records added, 98 records removed in the database
  • Auto ran revision 45d3ebf9 and completed successfully .
    98 records added, 98 records removed in the database
  • Auto ran revision 45d3ebf9 and completed successfully .
    98 records added, 98 records removed in the database
  • Auto ran revision 45d3ebf9 and completed successfully .
    98 records added, 98 records removed in the database
  • ...
  • Created on morph.io

Show complete history

Scraper code

Python

sp_E3720_WCC_gov / scraper.py