randomgitman / cn_data_ro

Scrapes data.gov.ro

Bun venit - data.gov.ro


Last run failed with status code 1.

Console output of last run

Injecting configuration and compiling... [1G [1G-----> Python app detected [1G-----> Installing python-2.7.6 [1G $ pip install -r requirements.txt [1G Obtaining scraperwiki from git+http://github.com/openaustralia/scraperwiki-python.git@morph_defaults#egg=scraperwiki (from -r requirements.txt (line 2)) [1G Cloning http://github.com/openaustralia/scraperwiki-python.git (to morph_defaults) to ./.heroku/src/scraperwiki [1G Collecting lxml==3.4.4 (from -r requirements.txt (line 4)) [1G /app/.heroku/python/lib/python2.7/site-packages/pip-8.1.2-py2.7.egg/pip/_vendor/requests/packages/urllib3/util/ssl_.py:318: SNIMissingWarning: An HTTPS request has been made, but the SNI (Subject Name Indication) extension to TLS is not available on this platform. This may cause the server to present an incorrect TLS certificate, which can cause validation failures. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#snimissingwarning. [1G SNIMissingWarning [1G /app/.heroku/python/lib/python2.7/site-packages/pip-8.1.2-py2.7.egg/pip/_vendor/requests/packages/urllib3/util/ssl_.py:122: InsecurePlatformWarning: A true SSLContext object is not available. This prevents urllib3 from configuring SSL appropriately and may cause certain SSL connections to fail. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#insecureplatformwarning. [1G InsecurePlatformWarning [1G Downloading lxml-3.4.4.tar.gz (3.5MB) [1G Collecting cssselect==0.9.1 (from -r requirements.txt (line 5)) [1G Downloading cssselect-0.9.1.tar.gz [1G Collecting beautifulsoup4 (from -r requirements.txt (line 6)) [1G Downloading beautifulsoup4-4.5.1-py2-none-any.whl (83kB) [1G Collecting python-dateutil (from -r requirements.txt (line 7)) [1G Downloading python_dateutil-2.6.0-py2.py3-none-any.whl (194kB) [1G Collecting selenium (from -r requirements.txt (line 8)) [1G Downloading selenium-3.0.2-py2.py3-none-any.whl (915kB) [1G Collecting splinter>=0.7.3 (from -r requirements.txt (line 9)) [1G Downloading splinter-0.7.5.tar.gz [1G Collecting dumptruck>=0.1.2 (from scraperwiki->-r requirements.txt (line 2)) [1G Downloading dumptruck-0.1.6.tar.gz [1G Collecting requests (from scraperwiki->-r requirements.txt (line 2)) [1G Downloading requests-2.12.3-py2.py3-none-any.whl (575kB) [1G Collecting six>=1.5 (from python-dateutil->-r requirements.txt (line 7)) [1G Downloading six-1.10.0-py2.py3-none-any.whl [1G Installing collected packages: dumptruck, requests, scraperwiki, lxml, cssselect, beautifulsoup4, six, python-dateutil, selenium, splinter [1G Running setup.py install for dumptruck: started [1G Running setup.py install for dumptruck: finished with status 'done' [1G Running setup.py develop for scraperwiki [1G Running setup.py install for lxml: started [1G Running setup.py install for lxml: still running... [1G Running setup.py install for lxml: finished with status 'done' [1G Running setup.py install for cssselect: started [1G Running setup.py install for cssselect: finished with status 'done' [1G Running setup.py install for splinter: started [1G Running setup.py install for splinter: finished with status 'done' [1G Successfully installed beautifulsoup4-4.5.1 cssselect-0.9.1 dumptruck-0.1.6 lxml-3.4.4 python-dateutil-2.6.0 requests-2.12.3 scraperwiki selenium-3.0.2 six-1.10.0 splinter-0.7.5 [1G [1G ! Hello! It looks like your application is using an outdated version of Python. [1G ! This caused the security warning you saw above during the 'pip install' step. [1G ! We recommend 'python-2.7.12', which you can specify in a 'runtime.txt' file. [1G ! -- Much Love, Heroku. [1G [1G [1G-----> Discovering process types [1G Procfile declares types -> scraper Injecting scraper and running... Traceback (most recent call last): File "scraper.py", line 46, in <module> links = get_links(portal) File "scraper.py", line 19, in get_links table = soup.find('ul', {"class":"resource-list"}).findAll('a') AttributeError: 'NoneType' object has no attribute 'findAll'

Data

Downloaded 2 times by MikeRalphson

To download data sign in with GitHub

Download table (as CSV) Download SQLite database (11 KB) Use the API

rows 10 / 14

country_code data_url todays_date tender_url
ro
2016-06-01 00:24:31.354156
ro
2016-06-01 00:24:31.354156
ro
2016-06-01 00:24:31.354156
ro
2016-06-01 00:24:31.354156
ro
2016-06-01 00:24:31.354156
ro
2016-06-01 00:24:31.354156
ro
2016-06-01 00:24:31.354156
ro
2016-06-01 00:24:31.354156
ro
2016-06-01 00:24:31.354156
ro
2016-06-01 00:24:31.354156

Statistics

Average successful run time: 12 minutes

Total run time: about 14 hours

Total cpu time used: 3 minutes

Total disk space used: 30.5 KB

History

  • Auto ran revision f0ccefae and failed .
    nothing changed in the database
    2 pages scraped
  • Auto ran revision f0ccefae and failed .
    nothing changed in the database
    2 pages scraped
  • Auto ran revision f0ccefae and failed .
    nothing changed in the database
  • Auto ran revision f0ccefae and failed .
    nothing changed in the database
    2 pages scraped
  • Auto ran revision f0ccefae and failed .
    nothing changed in the database
    2 pages scraped
  • ...
  • Created on morph.io

Show complete history

Scraper code

cn_data_ro