danwainwright / scraper_15_tutorial

album data for tutorial


Contributors danwainwright

Last run failed with status code 1.

Console output of last run

Injecting configuration and compiling...  -----> Python app detected -----> Installing python-2.7.9  $ pip install -r requirements.txt  Obtaining scraperwiki from git+http://github.com/openaustralia/scraperwiki-python.git@morph_defaults#egg=scraperwiki (from -r /tmp/build/requirements.txt (line 6))  Cloning http://github.com/openaustralia/scraperwiki-python.git (to morph_defaults) to /app/.heroku/src/scraperwiki  Collecting lxml==3.4.4 (from -r /tmp/build/requirements.txt (line 8))  Downloading lxml-3.4.4.tar.gz (3.5MB)  Collecting cssselect==0.9.1 (from -r /tmp/build/requirements.txt (line 9))  Downloading cssselect-0.9.1.tar.gz  Collecting dumptruck>=0.1.2 (from scraperwiki->-r /tmp/build/requirements.txt (line 6))  Downloading dumptruck-0.1.6.tar.gz  Collecting requests (from scraperwiki->-r /tmp/build/requirements.txt (line 6))  Downloading requests-2.18.1-py2.py3-none-any.whl (88kB)  Collecting urllib3<1.22,>=1.21.1 (from requests->scraperwiki->-r /tmp/build/requirements.txt (line 6))  Downloading urllib3-1.21.1-py2.py3-none-any.whl (131kB)  Collecting idna<2.6,>=2.5 (from requests->scraperwiki->-r /tmp/build/requirements.txt (line 6))  Downloading idna-2.5-py2.py3-none-any.whl (55kB)  Collecting certifi>=2017.4.17 (from requests->scraperwiki->-r /tmp/build/requirements.txt (line 6))  Downloading certifi-2017.4.17-py2.py3-none-any.whl (375kB)  Collecting chardet<3.1.0,>=3.0.2 (from requests->scraperwiki->-r /tmp/build/requirements.txt (line 6))  Downloading chardet-3.0.4-py2.py3-none-any.whl (133kB)  Installing collected packages: dumptruck, urllib3, idna, certifi, chardet, requests, scraperwiki, lxml, cssselect  Running setup.py install for dumptruck: started  Running setup.py install for dumptruck: finished with status 'done'  Running setup.py develop for scraperwiki  Running setup.py install for lxml: started  Running setup.py install for lxml: still running...  Running setup.py install for lxml: finished with status 'done'  Running setup.py install for cssselect: started  Running setup.py install for cssselect: finished with status 'done'  Successfully installed certifi-2017.4.17 chardet-3.0.4 cssselect-0.9.1 dumptruck-0.1.6 idna-2.5 lxml-3.4.4 requests-2.18.1 scraperwiki urllib3-1.21.1   -----> Discovering process types  Procfile declares types -> scraper Injecting scraper and running... Traceback (most recent call last): File "scraper.py", line 48, in <module> scrape_and_look_for_next_link(starting_url) File "scraper.py", line 31, in scrape_and_look_for_next_link html = scraperwiki.scrape(url) File "/app/.heroku/src/scraperwiki/scraperwiki/utils.py", line 31, in scrape f = urllib2.urlopen(req) File "/app/.heroku/python/lib/python2.7/urllib2.py", line 154, in urlopen return opener.open(url, data, timeout) File "/app/.heroku/python/lib/python2.7/urllib2.py", line 437, in open response = meth(req, response) File "/app/.heroku/python/lib/python2.7/urllib2.py", line 550, in http_response 'http', request, response, code, msg, hdrs) File "/app/.heroku/python/lib/python2.7/urllib2.py", line 475, in error return self._call_chain(*args) File "/app/.heroku/python/lib/python2.7/urllib2.py", line 409, in _call_chain result = func(*args) File "/app/.heroku/python/lib/python2.7/urllib2.py", line 558, in http_error_default raise HTTPError(req.get_full_url(), code, msg, hdrs, fp) urllib2.HTTPError: HTTP Error 404: Not Found

Statistics

Total run time: 2 minutes

Total cpu time used: less than 5 seconds

Total disk space used: 22.6 KB

History

  • Manually ran revision 07623b96 and failed .
    nothing changed in the database
  • Created on morph.io

Scraper code

scraper_15_tutorial