woodbine / sp_E3820_WSCC_gov

Scrapes www.westsussex.gov.uk

Find information for residents, visitors and businesses on the local authority website for West Sussex, the county serving over 750,000 people.


This is a scraper that runs on Morph. To get started see the documentation

Contributors woodbine blablupcom

Last run completed successfully .

Console output of last run

Injecting configuration and compiling...  -----> Python app detected  ! The latest version of Python 2 is python-2.7.14 (you are using python-2.7.9, which is unsupported).  ! We recommend upgrading by specifying the latest version (python-2.7.14).  Learn More: https://devcenter.heroku.com/articles/python-runtimes -----> Installing python-2.7.9 -----> Installing pip -----> Installing requirements with pip  Collecting beautifulsoup4==4.2.0 (from -r /tmp/build/requirements.txt (line 5))  Downloading https://files.pythonhosted.org/packages/b3/fe/6888f755c0b9f66d4cc1f9ee0077a613a09e211acca3242a7335ffe7ed06/beautifulsoup4-4.2.0.tar.gz (63kB)  Obtaining scraperwiki from git+http://github.com/openaustralia/scraperwiki-python.git@morph_defaults#egg=scraperwiki (from -r /tmp/build/requirements.txt (line 8))  Cloning http://github.com/openaustralia/scraperwiki-python.git (to revision morph_defaults) to /app/.heroku/src/scraperwiki  Collecting lxml==3.4.4 (from -r /tmp/build/requirements.txt (line 10))  Downloading https://files.pythonhosted.org/packages/63/c7/4f2a2a4ad6c6fa99b14be6b3c1cece9142e2d915aa7c43c908677afc8fa4/lxml-3.4.4.tar.gz (3.5MB)  Collecting cssselect==0.9.1 (from -r /tmp/build/requirements.txt (line 11))  Downloading https://files.pythonhosted.org/packages/aa/e5/9ee1460d485b94a6d55732eb7ad5b6c084caf73dd6f9cb0bb7d2a78fafe8/cssselect-0.9.1.tar.gz  Collecting dumptruck>=0.1.2 (from scraperwiki->-r /tmp/build/requirements.txt (line 8))  Downloading https://files.pythonhosted.org/packages/15/27/3330a343de80d6849545b6c7723f8c9a08b4b104de964ac366e7e6b318df/dumptruck-0.1.6.tar.gz  Collecting requests (from scraperwiki->-r /tmp/build/requirements.txt (line 8))  Downloading https://files.pythonhosted.org/packages/65/47/7e02164a2a3db50ed6d8a6ab1d6d60b69c4c3fdf57a284257925dfc12bda/requests-2.19.1-py2.py3-none-any.whl (91kB)  Collecting idna<2.8,>=2.5 (from requests->scraperwiki->-r /tmp/build/requirements.txt (line 8))  Downloading https://files.pythonhosted.org/packages/4b/2a/0276479a4b3caeb8a8c1af2f8e4355746a97fab05a372e4a2c6a6b876165/idna-2.7-py2.py3-none-any.whl (58kB)  Collecting certifi>=2017.4.17 (from requests->scraperwiki->-r /tmp/build/requirements.txt (line 8))  Downloading https://files.pythonhosted.org/packages/7c/e6/92ad559b7192d846975fc916b65f667c7b8c3a32bea7372340bfe9a15fa5/certifi-2018.4.16-py2.py3-none-any.whl (150kB)  Collecting urllib3<1.24,>=1.21.1 (from requests->scraperwiki->-r /tmp/build/requirements.txt (line 8))  Downloading https://files.pythonhosted.org/packages/bd/c9/6fdd990019071a4a32a5e7cb78a1d92c53851ef4f56f62a3486e6a7d8ffb/urllib3-1.23-py2.py3-none-any.whl (133kB)  Collecting chardet<3.1.0,>=3.0.2 (from requests->scraperwiki->-r /tmp/build/requirements.txt (line 8))  Downloading https://files.pythonhosted.org/packages/bc/a9/01ffebfb562e4274b6487b4bb1ddec7ca55ec7510b22e4c51f14098443b8/chardet-3.0.4-py2.py3-none-any.whl (133kB)  Installing collected packages: beautifulsoup4, dumptruck, idna, certifi, urllib3, chardet, requests, scraperwiki, lxml, cssselect  Running setup.py install for beautifulsoup4: started  Running setup.py install for beautifulsoup4: finished with status 'done'  Running setup.py install for dumptruck: started  Running setup.py install for dumptruck: finished with status 'done'  Running setup.py develop for scraperwiki  Running setup.py install for lxml: started  Running setup.py install for lxml: still running...  Running setup.py install for lxml: finished with status 'done'  Running setup.py install for cssselect: started  Running setup.py install for cssselect: finished with status 'done'  Successfully installed beautifulsoup4-4.2.0 certifi-2018.4.16 chardet-3.0.4 cssselect-0.9.1 dumptruck-0.1.6 idna-2.7 lxml-3.4.4 requests-2.19.1 scraperwiki urllib3-1.23   -----> Discovering process types  Procfile declares types -> scraper Injecting scraper and running...

Data

Downloaded 521 times by SimKennedy woodbine MikeRalphson

To download data sign in with GitHub

Download table (as CSV) Download SQLite database (17 KB) Use the API

rows 10 / 50

f d l
E3820_WSCC_gov_2016_03
2016-06-16 10:51:46.802345
E3820_WSCC_gov_2016_02
2016-06-16 10:51:48.024810
E3820_WSCC_gov_2016_01
2016-06-16 10:51:49.395262
E3820_WSCC_gov_2015_12
2016-06-16 10:51:50.946524
E3820_WSCC_gov_2015_11
2016-06-16 10:51:52.333700
E3820_WSCC_gov_2015_10
2016-06-16 10:51:53.519136
E3820_WSCC_gov_2015_09
2016-06-16 10:51:54.835552
E3820_WSCC_gov_2015_08
2016-06-16 10:51:56.757327
E3820_WSCC_gov_2015_07
2016-06-16 10:51:57.788285
E3820_WSCC_gov_2015_06
2016-06-16 10:51:58.945084

Statistics

Average successful run time: 15 minutes

Total run time: about 1 month

Total cpu time used: 15 minutes

Total disk space used: 46.6 KB

History

  • Auto ran revision bbabae65 and completed successfully .
    nothing changed in the database
    1 page scraped
  • Auto ran revision bbabae65 and completed successfully .
    nothing changed in the database
    1 page scraped
  • Auto ran revision bbabae65 and completed successfully .
    nothing changed in the database
    1 page scraped
  • Auto ran revision bbabae65 and completed successfully .
    nothing changed in the database
    1 page scraped
  • Auto ran revision bbabae65 and completed successfully .
    nothing changed in the database
  • ...
  • Created on morph.io

Show complete history

Scraper code

Python

sp_E3820_WSCC_gov / scraper.py