Pipeline ID: activities/social_services/social_service_reports
Pipeline ID: activities/social_services/social_service_suppliers
Pipeline ID: activities/social_services/social_service_tenders
Pipeline ID: activities/social_services/social_services
scraper
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/pyquery/pyquery.py", line 77, in fromstring
result = getattr(etree, meth)(context)
File "src/lxml/etree.pyx", line 3428, in lxml.etree.fromstring
File "src/lxml/parser.pxi", line 2066, in lxml.etree._parseMemoryDocument
File "src/lxml/parser.pxi", line 1921, in lxml.etree._parseDoc
File "src/lxml/parser.pxi", line 1937, in lxml.etree._parseDoc_unicode
File "src/lxml/parser.pxi", line 1136, in lxml.etree._BaseParser._parseUnicodeDoc
File "src/lxml/parser.pxi", line 647, in lxml.etree._ParserContext._handleParseResultDoc
File "src/lxml/parser.pxi", line 765, in lxml.etree._handleParseResult
File "src/lxml/parser.pxi", line 689, in lxml.etree._raiseParseError
File "<string>", line 1
lxml.etree.XMLSyntaxError: Document is empty, line 1, column 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/pipelines/bank_of_israel/scraper.py", line 124, in <module>
series_sum = get_all_series()
File "/datapackage_pipelines_budgetkey/pipelines/bank_of_israel/scraper.py", line 23, in get_all_series
series = get_series(sid)
File "/datapackage_pipelines_budgetkey/pipelines/bank_of_israel/scraper.py", line 36, in get_series
resp = pq(data.decode(response.encoding or 'utf8'))
File "/usr/local/lib/python3.9/site-packages/pyquery/pyquery.py", line 230, in __init__
elements = fromstring(context, self.parser)
File "/usr/local/lib/python3.9/site-packages/pyquery/pyquery.py", line 81, in fromstring
result = getattr(lxml.html, meth)(context)
File "/usr/local/lib/python3.9/site-packages/lxml/html/__init__.py", line 854, in fromstring
doc = document_fromstring(html, parser=parser, base_url=base_url, **kw)
File "/usr/local/lib/python3.9/site-packages/lxml/html/__init__.py", line 742, in document_fromstring
raise etree.ParserError(
lxml.etree.ParserError: Document is empty
Pipeline ID: bank_of_israel/bank_of_israel
Pipeline ID: budget/municipal/fetcher
Pipeline ID: budget/national/changes/explanations/all
Pipeline ID: budget/national/changes/original/committee-zipfile
Pipeline ID: budget/national/changes/original/current-year-fixes
Pipeline ID: budget/national/changes/original/national-budget-changes
Pipeline ID: budget/national/changes/processed/national-budget-changes-aggregated
Pipeline ID: budget/national/changes/processed/transactions
We get the data the Ministry of Finance publishes in data.gov.il every year. It comes in XLS format, with one row per TAKANA and phase (original, approved, executed). In this pipeline we create from each triplet a single row that has all the data.
Pipeline ID: budget/national/original/national-budgets
This pipeline joins the different phases of the budget (allocated, revised and executed). In the original file there's a separate row for each of the phases. We like it better as a single row with all phase info. Another thing this pipeline does is to rename the column titles - to more friendly English names. Also, it create rows for all hierarchies - where upper hierarchies (2, 4 & 6 digits) are plain aggregations of the 8-digit items they contain.
Pipeline ID: budget/national/processed/aggregated-yearly
Pipeline ID: budget/national/processed/category-explanations
This pipeline joins budget items that span across years.
Pipeline ID: budget/national/processed/connected-items-explained
This pipeline joins budget items that span across years.
Pipeline ID: budget/national/processed/connected-national-budgets
Pipeline ID: budget/national/processed/just-the-total
Pipeline ID: budget/national/processed/roof-names
This pipeline joins the budget data to itself so that each item has a list of its immediate children.
Pipeline ID: budget/national/processed/with-extras
Pipeline ID: budgetkey/analysis/spending/publisher_entity_analysis
Pipeline ID: budgetkey/analysis/spending/publisher_foa_analysis
Pipeline ID: budgetkey/analysis/spending/united
Pipeline ID: budgetkey/budget-functional-aggregates
Pipeline ID: budgetkey/documents
Pipeline ID: budgetkey/elasticsearch/index_activities
Pipeline ID: budgetkey/elasticsearch/index_budget
Invalid dependency :Cannot run until dependency passes validation: ./procurement/calls_for_bids/calls-for-bids-all From ./procurement/calls_for_bids/calls-for-bids-all :['Invalid dependency', 'Cannot run until dependency passes validation: ./procurement/calls_for_bids/jobiz-calls-for-bids'] From ./procurement/calls_for_bids/calls-for-bids-all :['From ./procurement/calls_for_bids/jobiz-calls-for-bids', ['Invalid dependency', 'Cannot run until dependency passes validation: ./procurement/calls_for_bids/jobiz-all']] From ./procurement/calls_for_bids/calls-for-bids-all :['From ./procurement/calls_for_bids/jobiz-calls-for-bids', ['From ./procurement/calls_for_bids/jobiz-all', ['Dependency unsuccessful', 'Cannot run until dependency "./procurement/calls_for_bids/jobiz-scraper" is successfully executed']]]
Pipeline ID: budgetkey/elasticsearch/index_calls_for_bids
Pipeline ID: budgetkey/elasticsearch/index_contract_spending
Pipeline ID: budgetkey/elasticsearch/index_entities
dump_to_es
ERROR :DUMP TO ES ERROR ApiError(413, 'None')
ERROR :TB
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/processors/dump_to_es.py", line 146, in <module>
spew_flow(flow(ctx.parameters), ctx)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 181, in __exit__
spew(self.datapackage, self.resource_iterator, stats=self.stats)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 75, in spew
for rec in res:
File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/dumper_base.py", line 69, in row_counter
for row in iterator:
File "/usr/local/lib/python3.9/site-packages/tableschema_elasticsearch/storage.py", line 226, in write
for result, row in iter:
File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 453, in streaming_bulk
for data, (ok, info) in zip(
File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 359, in _process_bulk_chunk
yield from gen
File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 293, in _process_bulk_chunk_error
raise error
File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 343, in _process_bulk_chunk
resp = client.bulk(*args, operations=bulk_actions, **kwargs) # type: ignore[arg-type]
File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/utils.py", line 452, in wrapped
return api(*args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/__init__.py", line 828, in bulk
return self.perform_request( # type: ignore[return-value]
File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/_base.py", line 271, in perform_request
response = self._perform_request(
File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/_base.py", line 351, in _perform_request
raise HTTP_EXCEPTIONS.get(meta.status, ApiError)(
elasticsearch.ApiError: ApiError(413, 'None')
ERROR :Traceback (most recent call last):
ERROR :File "/datapackage_pipelines_budgetkey/processors/dump_to_es.py", line 146, in <module>
ERROR :spew_flow(flow(ctx.parameters), ctx)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 181, in __exit__
ERROR :spew(self.datapackage, self.resource_iterator, stats=self.stats)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 75, in spew
ERROR :for rec in res:
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/dumper_base.py", line 69, in row_counter
ERROR :for row in iterator:
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_elasticsearch/storage.py", line 226, in write
ERROR :for result, row in iter:
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 453, in streaming_bulk
ERROR :for data, (ok, info) in zip(
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 359, in _process_bulk_chunk
ERROR :yield from gen
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 293, in _process_bulk_chunk_error
ERROR :raise error
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 343, in _process_bulk_chunk
ERROR :resp = client.bulk(*args, operations=bulk_actions, **kwargs) # type: ignore[arg-type]
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/utils.py", line 452, in wrapped
ERROR :return api(*args, **kwargs)
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/__init__.py", line 828, in bulk
ERROR :return self.perform_request( # type: ignore[return-value]
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/_base.py", line 271, in perform_request
ERROR :response = self._perform_request(
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/_base.py", line 351, in _perform_request
ERROR :raise HTTP_EXCEPTIONS.get(meta.status, ApiError)(
ERROR :elasticsearch
ERROR :.
ERROR :ApiError
ERROR ::
ERROR :ApiError(413, 'None')
Pipeline ID: budgetkey/elasticsearch/index_gov_decisions
Pipeline ID: budgetkey/elasticsearch/index_muni_budgets
Pipeline ID: budgetkey/elasticsearch/index_muni_tenders
Pipeline ID: budgetkey/elasticsearch/index_national_budget_changes
Invalid dependency :Cannot run until dependency passes validation: ./people/aggregated From ./people/aggregated :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/all'] From ./people/aggregated :['From ./people/all', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/all-appointments']] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/media/themarker/themarker" is successfully executed']]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-nominations']]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']]]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]]]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]]]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-resignations']]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']]]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]]]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]]]] From ./people/aggregated :['From ./people/all', ['Dependency unsuccessful', 'Cannot run until dependency "./people/political_candidates/election_candidates21" is successfully executed']]
Pipeline ID: budgetkey/elasticsearch/index_people
Pipeline ID: budgetkey/elasticsearch/index_reports
Dependency unsuccessful :Cannot run until dependency "./supports/criteria/support-criteria" is successfully executed
Pipeline ID: budgetkey/elasticsearch/index_support_criteria
Pipeline ID: budgetkey/elasticsearch/index_support_programs
Pipeline ID: budgetkey/elasticsearch/index_supports
Pipeline ID: budgetkey/elasticsearch/index_tenders
Pipeline ID: budgetkey/elasticsearch/index_units
Dependency unsuccessful :Cannot run until dependency "./budgetkey/elasticsearch/index_gov_decisions" is successfully executed
Pipeline ID: budgetkey/elasticsearch/sitemaps
Pipeline ID: budgetkey/emails/emails
Pipeline ID: donations/candidates
Pipeline ID: donations/parties
get_transactions
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/pipelines/donations/get_transactions.py", line 150, in <module>
spew(dp, process_resources(res_iter))
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 75, in spew
for rec in res:
File "/datapackage_pipelines_budgetkey/pipelines/donations/get_transactions.py", line 109, in get_transactions
for resp in self.get_for_candidate(cid):
File "/datapackage_pipelines_budgetkey/pipelines/donations/get_transactions.py", line 95, in get_for_candidate
resp = self.get_for_range(cid, range_start, range_end)
File "/datapackage_pipelines_budgetkey/pipelines/donations/get_transactions.py", line 70, in get_for_range
assert len(resp) == 6
AssertionError
Pipeline ID: donations/transactions
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/education/tochniyot/scraper.py", line 103, in flow
return Flow(scrape(),
File "/datapackage_pipelines_budgetkey/pipelines/education/tochniyot/scraper.py", line 65, in scrape
education_programs_count = get_education_programs_count()
File "/datapackage_pipelines_budgetkey/pipelines/education/tochniyot/scraper.py", line 56, in get_education_programs_count
res = send_tochniyot_request()
File "/datapackage_pipelines_budgetkey/pipelines/education/tochniyot/scraper.py", line 80, in send_tochniyot_request
resp.raise_for_status()
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 1021, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: https://apps.education.gov.il/TyhNet/ClientWs/TochnitCh.asmx/IturTochnitChByMeafyenim
Pipeline ID: education/tochniyot/tochniyot
Pipeline ID: entities/all
Pipeline ID: entities/associations/guidestar/guidestar
Pipeline ID: entities/associations/guidestar/guidestar-scraper
רשימת כלל העמותות הפעילות, מתעדכן שבועית.
Pipeline ID: entities/associations/registrar/registry
Pipeline ID: entities/companies/registrar/details
רשימת כלל החברות הפעילות, מתעדכן שבועית.
Pipeline ID: entities/companies/registrar/scraper
רשימת אגודות שיתופיות, מתוך ממשק נתונים של מאגר של משרד הכלכלה / האגף לאיגוד שיתופי, מתעדכן שבועית
Pipeline ID: entities/cooperatives/cooperatives
Pipeline ID: entities/endowments/endowments
Pipeline ID: entities/entities
Pipeline ID: entities/fingerprints
Pipeline ID: entities/foi_orgs/foi_orgs
scraper
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/pipelines/entities/moital/scraper.py", line 79, in <module>
spew(datapackage, [scrape()])
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 75, in spew
for rec in res:
File "/datapackage_pipelines_budgetkey/pipelines/entities/moital/scraper.py", line 32, in scrape
desired_capabilities=DesiredCapabilities.PHANTOMJS)
AttributeError: type object 'DesiredCapabilities' has no attribute 'PHANTOMJS'
Pipeline ID: entities/moital/moital_service_providers
רשימת אגודות עותומניות, מתוך תשובה לבקשת חופש מידע
Pipeline ID: entities/ottoman/ottoman-association-registry
Pipeline ID: entities/partnerships/partnerships
stream_remote_resources
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 174, in _new_conn
conn = connection.create_connection(
File "/usr/local/lib/python3.9/site-packages/urllib3/util/connection.py", line 72, in create_connection
for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
File "/usr/local/lib/python3.9/socket.py", line 954, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno -2] Name or service not known
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 714, in urlopen
httplib_response = self._make_request(
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 415, in _make_request
conn.request(method, url, **httplib_request_kw)
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 244, in request
super(HTTPConnection, self).request(method, url, body=body, headers=headers)
File "/usr/local/lib/python3.9/http/client.py", line 1285, in request
self._send_request(method, url, body, headers, encode_chunked)
File "/usr/local/lib/python3.9/http/client.py", line 1331, in _send_request
self.endheaders(body, encode_chunked=encode_chunked)
File "/usr/local/lib/python3.9/http/client.py", line 1280, in endheaders
self._send_output(message_body, encode_chunked=encode_chunked)
File "/usr/local/lib/python3.9/http/client.py", line 1040, in _send_output
self.send(msg)
File "/usr/local/lib/python3.9/http/client.py", line 980, in send
self.connect()
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 205, in connect
conn = self._new_conn()
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 186, in _new_conn
raise NewConnectionError(
urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPConnection object at 0x7f5cec45c580>: Failed to establish a new connection: [Errno -2] Name or service not known
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 486, in send
resp = conn.urlopen(
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 798, in urlopen
retries = retries.increment(
File "/usr/local/lib/python3.9/site-packages/urllib3/util/retry.py", line 592, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPConnectionPool(host='employment.molsa.gov.il', port=80): Max retries exceeded with url: /Employment/SafetyAndHealth/Enforcement/Documents/%D7%A6%D7%95%D7%95%D7%99%20%D7%91%D7%98%D7%99%D7%97%D7%95%D7%AA.xlsx (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f5cec45c580>: Failed to establish a new connection: [Errno -2] Name or service not known'))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 63, in load
bytes = _RemoteStream(source, self.__http_session, self.__http_timeout).open()
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 121, in open
self.seek(0)
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 142, in seek
self.__response = self.__session.get(self.__source, stream=True, timeout=self.__timeout)
File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 602, in get
return self.request("GET", url, **kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 589, in request
resp = self.send(prep, **send_kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 703, in send
r = adapter.send(request, **kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 519, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPConnectionPool(host='employment.molsa.gov.il', port=80): Max retries exceeded with url: /Employment/SafetyAndHealth/Enforcement/Documents/%D7%A6%D7%95%D7%95%D7%99%20%D7%91%D7%98%D7%99%D7%97%D7%95%D7%AA.xlsx (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f5cec45c580>: Failed to establish a new connection: [Errno -2] Name or service not known'))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 244, in <module>
rows = stream_reader(resource, url, ignore_missing or url == "", limit_rows,
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 199, in stream_reader
schema, headers, columns, stream, close = get_opener(url, _resource)()
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 175, in opener
_stream.open()
File "/usr/local/lib/python3.9/site-packages/tabulator/stream.py", line 425, in open
self.__parser.open(source, encoding=self.__encoding)
File "/usr/local/lib/python3.9/site-packages/tabulator/parsers/xlsx.py", line 81, in open
source_bytes = self.__loader.load(source, mode="b", encoding=encoding)
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 72, in load
raise exceptions.HTTPError(str(exception))
tabulator.exceptions.HTTPError: HTTPConnectionPool(host='employment.molsa.gov.il', port=80): Max retries exceeded with url: /Employment/SafetyAndHealth/Enforcement/Documents/%D7%A6%D7%95%D7%95%D7%99%20%D7%91%D7%98%D7%99%D7%97%D7%95%D7%AA.xlsx (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f5cec45c580>: Failed to establish a new connection: [Errno -2] Name or service not known'))
Pipeline ID: entities/safety-orders/safety-orders
Dependency unsuccessful :Cannot run until dependency "./entities/safety-orders/safety-orders" is successfully executed
Pipeline ID: entities/safety-orders/safety-orders-by-entity
Pipeline ID: entities/special/registry
Pipeline ID: ezvonot/ezvonot_accounts
Pipeline ID: ezvonot/ezvonot_cases
Pipeline ID: ezvonot/ezvonot_crontab
Pipeline ID: ezvonot/ezvonot_decisions
Pipeline ID: ezvonot/ezvonot_decisions_join
Pipeline ID: ezvonot/ezvonot_deposits
Pipeline ID: ezvonot/ezvonot_purposes
Pipeline ID: facilities/all/facilities_all
Pipeline ID: facilities/education/facilities_education
Pipeline ID: facilities/labor/facilities_labor
Pipeline ID: facilities/welfare/facilities_welfare
Pipeline ID: government_decisions/government_decisions
Pipeline ID: government_decisions/scraper
Pipeline ID: knesset/ethics_committee_decisions
Pipeline ID: knesset/knesset_committee_decisions
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/knesset/knesset_legal_advisor.py", line 44, in flow
document = gcl.download(url, use_curl=True, outfile=filename)
File "/datapackage_pipelines_budgetkey/common/google_chrome.py", line 162, in download
assert False, 'Failed to download file, %r' % downloads
AssertionError: Failed to download file, ['']
DEBUG :[chan 16] Max packet in: 32768 bytes
DEBUG :[chan 16] Max packet out: 32768 bytes
DEBUG :Secsh channel 16 opened.
DEBUG :[chan 16] Sesch channel 16 request ok
DEBUG :[chan 16] EOF received (16)
DEBUG :[chan 16] EOF sent (16)
DEBUG :Dropping user packet because connection is dead.
DEBUG :Dropping user packet because connection is dead.
Pipeline ID: knesset/knesset_legal_advisor
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/knesset/knesset_legal_advisor_letters.py", line 44, in flow
document = gcl.download(url, use_curl=True, outfile=filename)
File "/datapackage_pipelines_budgetkey/common/google_chrome.py", line 162, in download
assert False, 'Failed to download file, %r' % downloads
AssertionError: Failed to download file, ['']
DEBUG :[chan 16] Max packet in: 32768 bytes
DEBUG :[chan 16] Max packet out: 32768 bytes
DEBUG :Secsh channel 16 opened.
DEBUG :[chan 16] Sesch channel 16 request ok
DEBUG :[chan 16] EOF received (16)
DEBUG :[chan 16] EOF sent (16)
DEBUG :Dropping user packet because connection is dead.
Pipeline ID: knesset/knesset_legal_advisor_letters
Pipeline ID: lamas/municipal-data
Pipeline ID: lamas/municipality-to-city
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-bonds-and-options-holdings-changes
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-company-director-appointments
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-company-officer-list
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-company-stakeholder-list
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-company-vip-appointments
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates From ./maya/maya-notification-parse-updates :['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']
Pipeline ID: maya/maya-complete-notification-list
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-holdings-changes
Dependency unsuccessful :Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed
Pipeline ID: maya/maya-notification-parse-updates
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-reported-academic-degrees
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-reported-work-record
Dependency unsuccessful :Cannot run until dependency "./maya/scrape-maya-tase-companies" is successfully executed
Pipeline ID: maya/maya-tase-companies-current-management
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 79, in _process
self.datapackage = self.process_datapackage(self.datapackage)
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 112, in process_datapackage
raise self.exc
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 60, in describe
sample = list(itertools.islice(self.iterable, self.SAMPLE_SIZE))
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 89, in handle_iterable
for x in self.iterable:
File "/datapackage_pipelines_budgetkey/pipelines/maya/scrape_maya_notification_list.py", line 153, in scrape_maya_notification_list
first_date, last_date = get_existing_records()
File "/datapackage_pipelines_budgetkey/pipelines/maya/scrape_maya_notification_list.py", line 31, in get_existing_records
rs = con.execute(text("SELECT min(date), max(date) FROM maya_notifications where parser_version=:v"), v=PARSER_VERSION)
TypeError: execute() got an unexpected keyword argument 'v'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 15, in <module>
spew_flow(flow, ctx)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/utilities/flow_utils.py", line 46, in spew_flow
datastream = flow.datastream()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/flow.py", line 19, in datastream
return self._chain(ds)._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 86, in _process
self.raise_exception(exception)
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 96, in raise_exception
raise error from cause
dataflows.base.exceptions.ProcessorError: Errored in processor iterable_loader in position #9: execute() got an unexpected keyword argument 'v'
Pipeline ID: maya/scrape-maya-notification-list
dump_to_path
ERROR :Data Package validation error: Descriptor validation error: [] is too short at "schema/fields" in descriptor and at "properties/schema/properties/fields/minItems" in profile
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/dump_to_path.py", line 22, in <module>
with ingest() as ctx:
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 186, in ingest
params, datapackage, resource_iterator = _ingest(debug=debug)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 44, in _ingest
datapackage, resource_iterator, dependency_dp = process_input(sys.stdin, validate, debug)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/input_processor.py", line 88, in process_input
datapackage.validate(dp_to_validate)
File "/usr/local/lib/python3.9/site-packages/datapackage/validate.py", line 28, in validate
Package(descriptor, strict=True)
File "/usr/local/lib/python3.9/site-packages/datapackage/package.py", line 126, in __init__
self.__build()
File "/usr/local/lib/python3.9/site-packages/datapackage/package.py", line 534, in __build
updated_resource = Resource(descriptor,
File "/usr/local/lib/python3.9/site-packages/datapackage/resource.py", line 90, in __init__
self.__build()
File "/usr/local/lib/python3.9/site-packages/datapackage/resource.py", line 648, in __build
raise exception
File "/usr/local/lib/python3.9/site-packages/datapackage/resource.py", line 643, in __build
self.__profile.validate(self.__current_descriptor)
File "/usr/local/lib/python3.9/site-packages/datapackage/profile.py", line 89, in validate
raise exceptions.ValidationError(message, errors=errors)
datapackage.exceptions.ValidationError: There are 1 validation errors (see exception.errors)
Pipeline ID: maya/scrape-maya-tase-companies
Invalid dependency :Cannot run until dependency passes validation: ./people/all From ./people/all :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/all-appointments'] From ./people/all :['From ./people/company_appointments/all-appointments', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/media/themarker/themarker" is successfully executed']] From ./people/all :['From ./people/company_appointments/all-appointments', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-nominations']] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']]] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]]] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]]] From ./people/all :['From ./people/company_appointments/all-appointments', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-resignations']] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']]] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]]] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]]] From ./people/all :['Dependency unsuccessful', 'Cannot run until dependency "./people/political_candidates/election_candidates21" is successfully executed']
Pipeline ID: people/aggregated
Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/all-appointments From ./people/company_appointments/all-appointments :['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/media/themarker/themarker" is successfully executed'] From ./people/company_appointments/all-appointments :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-nominations'] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-nominations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]] From ./people/company_appointments/all-appointments :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-resignations'] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-resignations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]] Dependency unsuccessful :Cannot run until dependency "./people/political_candidates/election_candidates21" is successfully executed
Pipeline ID: people/all
Pipeline ID: people/association_founders/association-founders
Dependency unsuccessful :Cannot run until dependency "./people/company_appointments/media/themarker/themarker" is successfully executed Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-nominations From ./people/company_appointments/maya/maya-nominations :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list'] From ./people/company_appointments/maya/maya-nominations :['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']] From ./people/company_appointments/maya/maya-nominations :['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]] Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-resignations From ./people/company_appointments/maya/maya-resignations :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list'] From ./people/company_appointments/maya/maya-resignations :['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']] From ./people/company_appointments/maya/maya-resignations :['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]
Pipeline ID: people/company_appointments/all-appointments
Dependency unsuccessful :Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed
Pipeline ID: people/company_appointments/maya/collect-maya-forms-debug-stats
Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-nominations From ./people/company_appointments/maya/maya-nominations :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list'] From ./people/company_appointments/maya/maya-nominations :['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']] From ./people/company_appointments/maya/maya-nominations :['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]] Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-resignations From ./people/company_appointments/maya/maya-resignations :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list'] From ./people/company_appointments/maya/maya-resignations :['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']] From ./people/company_appointments/maya/maya-resignations :['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]
Pipeline ID: people/company_appointments/maya/correlate
Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates From ./people/company_appointments/maya/maya-notification-updates :['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']
Pipeline ID: people/company_appointments/maya/maya-list
Name, job and company description.
Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list From ./people/company_appointments/maya/maya-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates'] From ./people/company_appointments/maya/maya-list :['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]
Pipeline ID: people/company_appointments/maya/maya-nominations
scrape-notification-list
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 971, in json
return complexjson.loads(self.text, **kwargs)
File "/usr/local/lib/python3.9/json/__init__.py", line 346, in loads
return _default_decoder.decode(s)
File "/usr/local/lib/python3.9/json/decoder.py", line 337, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/usr/local/lib/python3.9/json/decoder.py", line 355, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/pipelines/people/company_appointments/maya/scrape-notification-list.py", line 55, in _maya_api_call
return res.json()
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 975, in json
raise RequestsJSONDecodeError(e.msg, e.doc, e.pos)
requests.exceptions.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/pipelines/people/company_appointments/maya/scrape-notification-list.py", line 130, in <module>
spew(datapackage, [collect()])
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 75, in spew
for rec in res:
File "/datapackage_pipelines_budgetkey/pipelines/people/company_appointments/maya/scrape-notification-list.py", line 113, in collect
yield from _collect_date_range(year_start, year_end)
File "/datapackage_pipelines_budgetkey/pipelines/people/company_appointments/maya/scrape-notification-list.py", line 75, in _collect_date_range
res = _maya_api_call(date_from, date_to, current_page)
File "/datapackage_pipelines_budgetkey/pipelines/people/company_appointments/maya/scrape-notification-list.py", line 57, in _maya_api_call
raise Exception("Failed to Call Maya API for date_from:{} date_to:{} page_num:{}".format(date_from, date_to, page_num)) from e
Exception: Failed to Call Maya API for date_from:2006-01-01 date_to:2007-01-01 page_num:0
Pipeline ID: people/company_appointments/maya/maya-notification-list
id, date, company, notification_type, s3_object_name, url
Dependency unsuccessful :Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed
Pipeline ID: people/company_appointments/maya/maya-notification-updates
Name, job and company description.
Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list From ./people/company_appointments/maya/maya-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates'] From ./people/company_appointments/maya/maya-list :['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]
Pipeline ID: people/company_appointments/maya/maya-resignations
רשימת כל המינויים והחברות מאתר כלכליסט
Pipeline ID: people/company_appointments/media/calcalist/calcalist
Name, job and company description.
stream_remote_resources
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 63, in load
bytes = _RemoteStream(source, self.__http_session, self.__http_timeout).open()
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 121, in open
self.seek(0)
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 143, in seek
self.__response.raise_for_status()
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 1021, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://www.themarker.com/st/inter/DB/tm/2015/minuyim30.xlsx
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 244, in <module>
rows = stream_reader(resource, url, ignore_missing or url == "", limit_rows,
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 199, in stream_reader
schema, headers, columns, stream, close = get_opener(url, _resource)()
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 175, in opener
_stream.open()
File "/usr/local/lib/python3.9/site-packages/tabulator/stream.py", line 425, in open
self.__parser.open(source, encoding=self.__encoding)
File "/usr/local/lib/python3.9/site-packages/tabulator/parsers/xlsx.py", line 81, in open
source_bytes = self.__loader.load(source, mode="b", encoding=encoding)
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 72, in load
raise exceptions.HTTPError(str(exception))
tabulator.exceptions.HTTPError: 404 Client Error: Not Found for url: https://www.themarker.com/st/inter/DB/tm/2015/minuyim30.xlsx
Pipeline ID: people/company_appointments/media/themarker/themarker
dump_to_sql
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :psycopg2.errors
ERROR :.
ERROR :InternalError_
ERROR ::
ERROR :could not open relation with OID 497924577
ERROR :The above exception was the direct cause of the following exception:
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/dump_to_sql.py", line 19, in <module>
ERROR :spew_flow(flow(ctx.parameters), ctx)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 181, in __exit__
ERROR :spew(self.datapackage, self.resource_iterator, stats=self.stats)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 68, in spew
ERROR :for res in resources_iterator:
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 68, in <genexpr>
ERROR :res_iter = (it if isinstance(it, ResourceWrapper) else ResourceWrapper(res, it)
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/dumper_base.py", line 82, in process_resources
ERROR :ret = self.process_resource(
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/to_sql.py", line 108, in process_resource
ERROR :storage.delete('')
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 183, in delete
ERROR :self.__reflect()
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 278, in __reflect
ERROR :self.__metadata.reflect(only=only, bind=self.__engine)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 5752, in reflect
ERROR :_reflect_info = insp._get_reflection_info(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 2018, in _get_reflection_info
ERROR :check_constraints=run(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1994, in run
ERROR :res = meth(filter_names=_fn, **kw)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1457, in get_multi_check_constraints
ERROR :self.dialect.get_multi_check_constraints(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/dialects/postgresql/base.py", line 4677, in get_multi_check_constraints
ERROR :result = connection.execute(query, params)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1412, in execute
ERROR :return meth(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 483, in _execute_on_connection
ERROR :return connection._execute_clauseelement(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1635, in _execute_clauseelement
ERROR :ret = self._execute_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1844, in _execute_context
ERROR :return self._exec_single_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1984, in _exec_single_context
ERROR :self._handle_dbapi_exception(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 2339, in _handle_dbapi_exception
ERROR :raise sqlalchemy_exception.with_traceback(exc_info[2]) from e
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :sqlalchemy.exc
ERROR :.
ERROR :InternalError
ERROR ::
ERROR :(psycopg2.errors.InternalError_) could not open relation with OID 497924577
[SQL: SELECT pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname, CASE WHEN (pg_catalog.pg_constraint.oid IS NOT NULL) THEN pg_catalog.pg_get_constraintdef(pg_catalog.pg_constraint.oid, %(pg_get_constraintdef_1)s) END AS anon_1, pg_catalog.pg_description.description
FROM pg_catalog.pg_class LEFT OUTER JOIN pg_catalog.pg_constraint ON pg_catalog.pg_class.oid = pg_catalog.pg_constraint.conrelid AND pg_catalog.pg_constraint.contype = %(contype_1)s LEFT OUTER JOIN pg_catalog.pg_description ON pg_catalog.pg_description.objoid = pg_catalog.pg_constraint.oid JOIN pg_catalog.pg_namespace ON pg_catalog.pg_namespace.oid = pg_catalog.pg_class.relnamespace
WHERE pg_catalog.pg_class.relkind = ANY (ARRAY[%(param_1)s, %(param_2)s, %(param_3)s]) AND pg_catalog.pg_table_is_visible(pg_catalog.pg_class.oid) AND pg_catalog.pg_namespace.nspname != %(nspname_1)s ORDER BY pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname]
[parameters: {'pg_get_constraintdef_1': True, 'contype_1': 'c', 'param_1': 'r', 'param_2': 'p', 'param_3': 'f', 'nspname_1': 'pg_catalog'}]
(Background on this error at: https://sqlalche.me/e/20/2j85)
Pipeline ID: people/company_appointments/state_owned/correlate
Pipeline ID: people/company_appointments/state_owned/parse_pdf_resources
columns: date, url
Pipeline ID: people/company_appointments/state_owned/reports
Pipeline ID: people/company_appointments/state_owned/state-owned-nominations
Pipeline ID: people/municipal_appointments/municipal_representitives
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 79, in _process
self.datapackage = self.process_datapackage(self.datapackage)
File "/usr/local/lib/python3.9/site-packages/dataflows/processors/set_type.py", line 71, in process_datapackage
assert added, 'Failed to find field {} in schema'.format(self.name)
AssertionError: Failed to find field re.compile('^party_letter$') in schema
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 15, in <module>
spew_flow(flow, ctx)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/utilities/flow_utils.py", line 46, in spew_flow
datastream = flow.datastream()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/flow.py", line 19, in datastream
return self._chain(ds)._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
[Previous line repeated 4 more times]
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 86, in _process
self.raise_exception(exception)
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 96, in raise_exception
raise error from cause
dataflows.base.exceptions.ProcessorError: Errored in processor set_type in position #3: Failed to find field re.compile('^party_letter$') in schema
Pipeline ID: people/political_candidates/election_candidates21
Pipeline ID: people/political_donations/political_donations
Pipeline ID: people/procurement/procurement-individuals
Invalid dependency :Cannot run until dependency passes validation: ./procurement/calls_for_bids/jobiz-calls-for-bids From ./procurement/calls_for_bids/jobiz-calls-for-bids :['Invalid dependency', 'Cannot run until dependency passes validation: ./procurement/calls_for_bids/jobiz-all'] From ./procurement/calls_for_bids/jobiz-calls-for-bids :['From ./procurement/calls_for_bids/jobiz-all', ['Dependency unsuccessful', 'Cannot run until dependency "./procurement/calls_for_bids/jobiz-scraper" is successfully executed']]
Pipeline ID: procurement/calls_for_bids/calls-for-bids-all
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 79, in _process
self.datapackage = self.process_datapackage(self.datapackage)
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 112, in process_datapackage
raise self.exc
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 60, in describe
sample = list(itertools.islice(self.iterable, self.SAMPLE_SIZE))
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 89, in handle_iterable
for x in self.iterable:
File "/datapackage_pipelines_budgetkey/pipelines/procurement/calls_for_bids/m_tmicha.py", line 47, in m_tmicha_scraper
assert total > 0
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 15, in <module>
spew_flow(flow, ctx)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/utilities/flow_utils.py", line 46, in spew_flow
datastream = flow.datastream()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/flow.py", line 19, in datastream
return self._chain(ds)._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 86, in _process
self.raise_exception(exception)
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 96, in raise_exception
raise error from cause
dataflows.base.exceptions.ProcessorError: Errored in processor iterable_loader in position #1:
Pipeline ID: procurement/calls_for_bids/calls-for-bids-moh
Pipeline ID: procurement/calls_for_bids/gov-publications
Pipeline ID: procurement/calls_for_bids/gov-publications-ministry-of-immigration
Pipeline ID: procurement/calls_for_bids/gov-publications2
Dependency unsuccessful :Cannot run until dependency "./procurement/calls_for_bids/jobiz-scraper" is successfully executed
Pipeline ID: procurement/calls_for_bids/jobiz-all
Invalid dependency :Cannot run until dependency passes validation: ./procurement/calls_for_bids/jobiz-all From ./procurement/calls_for_bids/jobiz-all :['Dependency unsuccessful', 'Cannot run until dependency "./procurement/calls_for_bids/jobiz-scraper" is successfully executed']
Pipeline ID: procurement/calls_for_bids/jobiz-calls-for-bids
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 174, in _new_conn
conn = connection.create_connection(
File "/usr/local/lib/python3.9/site-packages/urllib3/util/connection.py", line 72, in create_connection
for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
File "/usr/local/lib/python3.9/socket.py", line 954, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno -5] No address associated with hostname
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 714, in urlopen
httplib_response = self._make_request(
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 403, in _make_request
self._validate_conn(conn)
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 1053, in _validate_conn
conn.connect()
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 363, in connect
self.sock = conn = self._new_conn()
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 186, in _new_conn
raise NewConnectionError(
urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f5a65bd0760>: Failed to establish a new connection: [Errno -5] No address associated with hostname
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 486, in send
resp = conn.urlopen(
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 798, in urlopen
retries = retries.increment(
File "/usr/local/lib/python3.9/site-packages/urllib3/util/retry.py", line 592, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='jobiz.gov.il', port=443): Max retries exceeded with url: /ajax/results/%D7%94%D7%95%D7%93%D7%A2%D7%94%20%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA/0?ie=0&typeie=%D7%94%D7%95%D7%93%D7%A2%D7%94+%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA&search=Array (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5a65bd0760>: Failed to establish a new connection: [Errno -5] No address associated with hostname'))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 79, in _process
self.datapackage = self.process_datapackage(self.datapackage)
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 112, in process_datapackage
raise self.exc
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 60, in describe
sample = list(itertools.islice(self.iterable, self.SAMPLE_SIZE))
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 89, in handle_iterable
for x in self.iterable:
File "/datapackage_pipelines_budgetkey/pipelines/procurement/calls_for_bids/jobiz.py", line 19, in fetch_results
content = requests.get(URL.format(index)).json()
File "/usr/local/lib/python3.9/site-packages/requests/api.py", line 73, in get
return request("get", url, params=params, **kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/api.py", line 59, in request
return session.request(method=method, url=url, **kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 589, in request
resp = self.send(prep, **send_kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 703, in send
r = adapter.send(request, **kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 519, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='jobiz.gov.il', port=443): Max retries exceeded with url: /ajax/results/%D7%94%D7%95%D7%93%D7%A2%D7%94%20%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA/0?ie=0&typeie=%D7%94%D7%95%D7%93%D7%A2%D7%94+%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA&search=Array (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5a65bd0760>: Failed to establish a new connection: [Errno -5] No address associated with hostname'))
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 15, in <module>
spew_flow(flow, ctx)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/utilities/flow_utils.py", line 46, in spew_flow
datastream = flow.datastream()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/flow.py", line 19, in datastream
return self._chain(ds)._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
[Previous line repeated 5 more times]
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 86, in _process
self.raise_exception(exception)
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 96, in raise_exception
raise error from cause
dataflows.base.exceptions.ProcessorError: Errored in processor iterable_loader in position #1: HTTPSConnectionPool(host='jobiz.gov.il', port=443): Max retries exceeded with url: /ajax/results/%D7%94%D7%95%D7%93%D7%A2%D7%94%20%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA/0?ie=0&typeie=%D7%94%D7%95%D7%93%D7%A2%D7%94+%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA&search=Array (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5a65bd0760>: Failed to establish a new connection: [Errno -5] No address associated with hostname'))
Pipeline ID: procurement/calls_for_bids/jobiz-scraper
Pipeline ID: procurement/municipal/fetcher
אנו סוכמים מידע מתוך הדוחות הרבעוניים לתוך טבלה המכילה את המידע העדכני ביותר.
Pipeline ID: procurement/spending/latest-contract-spending
Pipeline ID: procurement/spending/quarterly-contract-spending-report-uris
כל משרדי הממשלה ויחידות הסמך (שמחוברות למערכת מרכבה) צריכים לפרסם אחת לרבעון (תוך 45 יום מסוף הרבעון) דו״ח התקשרויות לפי פורמט מוגדר היטב הממומש כדו״ח מובנה במערכת המרכבה. כל המשרדים מדווחים עצמאית והדו״חות עולים לאתר היחידה לחופש המידע, משם אנו אוספים אותם. התוצר של השרשר הזה הוא פירוט כל הדיווחים של כל המשרדים.
Pipeline ID: procurement/spending/quarterly-contract-spending-reports
Pipeline ID: procurement/spending/quarterly-contract-spending-reports-data
Pipeline ID: procurement/spending/spending-by-entity
Pipeline ID: procurement/tenders/all
dump.to_sql
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :psycopg2.errors
ERROR :.
ERROR :InternalError_
ERROR ::
ERROR :could not open relation with OID 497924577
ERROR :The above exception was the direct cause of the following exception:
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/dump/to_sql.py", line 15, in <module>
ERROR :spew_flow(flow(ctx.parameters), ctx)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 181, in __exit__
ERROR :spew(self.datapackage, self.resource_iterator, stats=self.stats)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 68, in spew
ERROR :for res in resources_iterator:
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 68, in <genexpr>
ERROR :res_iter = (it if isinstance(it, ResourceWrapper) else ResourceWrapper(res, it)
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/dumper_base.py", line 82, in process_resources
ERROR :ret = self.process_resource(
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/to_sql.py", line 108, in process_resource
ERROR :storage.delete('')
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 183, in delete
ERROR :self.__reflect()
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 278, in __reflect
ERROR :self.__metadata.reflect(only=only, bind=self.__engine)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 5752, in reflect
ERROR :_reflect_info = insp._get_reflection_info(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 2018, in _get_reflection_info
ERROR :check_constraints=run(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1994, in run
ERROR :res = meth(filter_names=_fn, **kw)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1457, in get_multi_check_constraints
ERROR :self.dialect.get_multi_check_constraints(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/dialects/postgresql/base.py", line 4677, in get_multi_check_constraints
ERROR :result = connection.execute(query, params)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1412, in execute
ERROR :return meth(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 483, in _execute_on_connection
ERROR :return connection._execute_clauseelement(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1635, in _execute_clauseelement
ERROR :ret = self._execute_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1844, in _execute_context
ERROR :return self._exec_single_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1984, in _exec_single_context
ERROR :self._handle_dbapi_exception(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 2339, in _handle_dbapi_exception
ERROR :raise sqlalchemy_exception.with_traceback(exc_info[2]) from e
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :sqlalchemy.exc
ERROR :.
ERROR :InternalError
ERROR ::
ERROR :(psycopg2.errors.InternalError_) could not open relation with OID 497924577
[SQL: SELECT pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname, CASE WHEN (pg_catalog.pg_constraint.oid IS NOT NULL) THEN pg_catalog.pg_get_constraintdef(pg_catalog.pg_constraint.oid, %(pg_get_constraintdef_1)s) END AS anon_1, pg_catalog.pg_description.description
FROM pg_catalog.pg_class LEFT OUTER JOIN pg_catalog.pg_constraint ON pg_catalog.pg_class.oid = pg_catalog.pg_constraint.conrelid AND pg_catalog.pg_constraint.contype = %(contype_1)s LEFT OUTER JOIN pg_catalog.pg_description ON pg_catalog.pg_description.objoid = pg_catalog.pg_constraint.oid JOIN pg_catalog.pg_namespace ON pg_catalog.pg_namespace.oid = pg_catalog.pg_class.relnamespace
WHERE pg_catalog.pg_class.relkind = ANY (ARRAY[%(param_1)s, %(param_2)s, %(param_3)s]) AND pg_catalog.pg_table_is_visible(pg_catalog.pg_class.oid) AND pg_catalog.pg_namespace.nspname != %(nspname_1)s ORDER BY pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname]
[parameters: {'pg_get_constraintdef_1': True, 'contype_1': 'c', 'param_1': 'r', 'param_2': 'p', 'param_3': 'f', 'nspname_1': 'pg_catalog'}]
(Background on this error at: https://sqlalche.me/e/20/2j85)
Pipeline ID: procurement/tenders/processed
Pipeline ID: reports/all
Pipeline ID: reports/ngos/ngo-activity-report
Pipeline ID: reports/ngos/ngo-district-report
Pipeline ID: simpledb/process_simple_dbs
Pipeline ID: support_programs/all-support-programs
Dependency unsuccessful :Cannot run until dependency "./supports/supports-2025" is successfully executed
Pipeline ID: supports/all
Pipeline ID: supports/by-payment-year
dump.to_sql
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :psycopg2.errors
ERROR :.
ERROR :InternalError_
ERROR ::
ERROR :could not open relation with OID 497993584
ERROR :The above exception was the direct cause of the following exception:
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/dump/to_sql.py", line 15, in <module>
ERROR :spew_flow(flow(ctx.parameters), ctx)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 181, in __exit__
ERROR :spew(self.datapackage, self.resource_iterator, stats=self.stats)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 68, in spew
ERROR :for res in resources_iterator:
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 68, in <genexpr>
ERROR :res_iter = (it if isinstance(it, ResourceWrapper) else ResourceWrapper(res, it)
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/dumper_base.py", line 82, in process_resources
ERROR :ret = self.process_resource(
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/to_sql.py", line 108, in process_resource
ERROR :storage.delete('')
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 183, in delete
ERROR :self.__reflect()
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 278, in __reflect
ERROR :self.__metadata.reflect(only=only, bind=self.__engine)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 5752, in reflect
ERROR :_reflect_info = insp._get_reflection_info(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 2018, in _get_reflection_info
ERROR :check_constraints=run(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1994, in run
ERROR :res = meth(filter_names=_fn, **kw)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1457, in get_multi_check_constraints
ERROR :self.dialect.get_multi_check_constraints(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/dialects/postgresql/base.py", line 4677, in get_multi_check_constraints
ERROR :result = connection.execute(query, params)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1412, in execute
ERROR :return meth(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 483, in _execute_on_connection
ERROR :return connection._execute_clauseelement(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1635, in _execute_clauseelement
ERROR :ret = self._execute_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1844, in _execute_context
ERROR :return self._exec_single_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1984, in _exec_single_context
ERROR :self._handle_dbapi_exception(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 2339, in _handle_dbapi_exception
ERROR :raise sqlalchemy_exception.with_traceback(exc_info[2]) from e
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :sqlalchemy.exc
ERROR :.
ERROR :InternalError
ERROR ::
ERROR :(psycopg2.errors.InternalError_) could not open relation with OID 497993584
[SQL: SELECT pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname, CASE WHEN (pg_catalog.pg_constraint.oid IS NOT NULL) THEN pg_catalog.pg_get_constraintdef(pg_catalog.pg_constraint.oid, %(pg_get_constraintdef_1)s) END AS anon_1, pg_catalog.pg_description.description
FROM pg_catalog.pg_class LEFT OUTER JOIN pg_catalog.pg_constraint ON pg_catalog.pg_class.oid = pg_catalog.pg_constraint.conrelid AND pg_catalog.pg_constraint.contype = %(contype_1)s LEFT OUTER JOIN pg_catalog.pg_description ON pg_catalog.pg_description.objoid = pg_catalog.pg_constraint.oid JOIN pg_catalog.pg_namespace ON pg_catalog.pg_namespace.oid = pg_catalog.pg_class.relnamespace
WHERE pg_catalog.pg_class.relkind = ANY (ARRAY[%(param_1)s, %(param_2)s, %(param_3)s]) AND pg_catalog.pg_table_is_visible(pg_catalog.pg_class.oid) AND pg_catalog.pg_namespace.nspname != %(nspname_1)s ORDER BY pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname]
[parameters: {'pg_get_constraintdef_1': True, 'contype_1': 'c', 'param_1': 'r', 'param_2': 'p', 'param_3': 'f', 'nspname_1': 'pg_catalog'}]
(Background on this error at: https://sqlalche.me/e/20/2j85)
Pipeline ID: supports/by-request-year
sample
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/processors/sample.py", line 11, in <module>
res_name = parameters.get('resource', datapackage['resources'][0]['name'])
IndexError: list index out of range
Pipeline ID: supports/criteria/support-criteria
Pipeline ID: supports/criteria/support-criteria-doj
Pipeline ID: supports/supports-2004
Pipeline ID: supports/supports-2005
Pipeline ID: supports/supports-2006
Pipeline ID: supports/supports-2007
Pipeline ID: supports/supports-2008
Pipeline ID: supports/supports-2009
Pipeline ID: supports/supports-2010
Pipeline ID: supports/supports-2011
Pipeline ID: supports/supports-2012
Pipeline ID: supports/supports-2013
Pipeline ID: supports/supports-2014
Pipeline ID: supports/supports-2015
Pipeline ID: supports/supports-2016
Pipeline ID: supports/supports-2017
Pipeline ID: supports/supports-2018
Pipeline ID: supports/supports-2019
Pipeline ID: supports/supports-2020
Pipeline ID: supports/supports-2021
Pipeline ID: supports/supports-2022
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 180, in flow
DF.load(wrapper(year), format='csv',
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 21, in wrapper
return scraper(gcd, year)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 159, in scraper
get_results_for_column(driver, rects[i])
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 96, in get_results_for_column
column.click()
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/remote/webelement.py", line 89, in click
self._execute(Command.CLICK_ELEMENT)
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/remote/webelement.py", line 773, in _execute
return self._parent.execute(command, params)
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/remote/webdriver.py", line 430, in execute
self.error_handler.check_response(response)
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/remote/errorhandler.py", line 247, in check_response
raise exception_class(message, screen, stacktrace)
selenium.common.exceptions.ElementClickInterceptedException: Message: element click intercepted: Element <rect width="19.408288043478258" height="50" fill="#0d6e84" stroke="#ffffff" stroke-width="1"></rect> is not clickable at point (394, 289). Other element would receive the click: <path d="M-4,0 A4,4 0 1,0 4,0 A4,4 0 1,0 -4,0z" fill="#ff8000" stroke="#ffffff" stroke-width="1"></path>
(Session info: chrome=88.0.4324.96)
Stacktrace:
#0 0x5ad524986199 <unknown>
WARNING :FAILED to teardown google-chrome
Pipeline ID: supports/supports-2023
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 180, in flow
DF.load(wrapper(year), format='csv',
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 21, in wrapper
return scraper(gcd, year)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 142, in scraper
chart = get_chart(driver)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 34, in get_chart
chart = WebDriverWait(driver, 60).until(
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/support/wait.py", line 87, in until
raise TimeoutException(message, screen, stacktrace)
selenium.common.exceptions.TimeoutException: Message:
Stacktrace:
#0 0x56ea53123199 <unknown>
WARNING :FAILED to teardown google-chrome
Pipeline ID: supports/supports-2024
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 180, in flow
DF.load(wrapper(year), format='csv',
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 21, in wrapper
return scraper(gcd, year)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 142, in scraper
chart = get_chart(driver)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 34, in get_chart
chart = WebDriverWait(driver, 60).until(
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/support/wait.py", line 87, in until
raise TimeoutException(message, screen, stacktrace)
selenium.common.exceptions.TimeoutException: Message:
Stacktrace:
#0 0x63ebe378b199 <unknown>
WARNING :FAILED to teardown google-chrome
Pipeline ID: supports/supports-2025
Pipeline ID: supports/supports-by-entity
Pipeline ID: supports/with-entities
Pipeline ID: supports/with-keys
Pipeline ID: units/social_services/social_services
Invalid dependency :Cannot run until dependency passes validation: ./procurement/calls_for_bids/calls-for-bids-all From ./procurement/calls_for_bids/calls-for-bids-all :['Invalid dependency', 'Cannot run until dependency passes validation: ./procurement/calls_for_bids/jobiz-calls-for-bids'] From ./procurement/calls_for_bids/calls-for-bids-all :['From ./procurement/calls_for_bids/jobiz-calls-for-bids', ['Invalid dependency', 'Cannot run until dependency passes validation: ./procurement/calls_for_bids/jobiz-all']] From ./procurement/calls_for_bids/calls-for-bids-all :['From ./procurement/calls_for_bids/jobiz-calls-for-bids', ['From ./procurement/calls_for_bids/jobiz-all', ['Dependency unsuccessful', 'Cannot run until dependency "./procurement/calls_for_bids/jobiz-scraper" is successfully executed']]]
Pipeline ID: budgetkey/elasticsearch/index_calls_for_bids
Invalid dependency :Cannot run until dependency passes validation: ./people/aggregated From ./people/aggregated :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/all'] From ./people/aggregated :['From ./people/all', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/all-appointments']] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/media/themarker/themarker" is successfully executed']]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-nominations']]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']]]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]]]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]]]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-resignations']]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']]]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]]]] From ./people/aggregated :['From ./people/all', ['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]]]] From ./people/aggregated :['From ./people/all', ['Dependency unsuccessful', 'Cannot run until dependency "./people/political_candidates/election_candidates21" is successfully executed']]
Pipeline ID: budgetkey/elasticsearch/index_people
Dependency unsuccessful :Cannot run until dependency "./supports/criteria/support-criteria" is successfully executed
Pipeline ID: budgetkey/elasticsearch/index_support_criteria
Dependency unsuccessful :Cannot run until dependency "./budgetkey/elasticsearch/index_gov_decisions" is successfully executed
Pipeline ID: budgetkey/elasticsearch/sitemaps
Dependency unsuccessful :Cannot run until dependency "./entities/safety-orders/safety-orders" is successfully executed
Pipeline ID: entities/safety-orders/safety-orders-by-entity
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-bonds-and-options-holdings-changes
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-company-director-appointments
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-company-officer-list
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-company-stakeholder-list
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-company-vip-appointments
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates From ./maya/maya-notification-parse-updates :['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']
Pipeline ID: maya/maya-complete-notification-list
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-holdings-changes
Dependency unsuccessful :Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed
Pipeline ID: maya/maya-notification-parse-updates
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-reported-academic-degrees
Invalid dependency :Cannot run until dependency passes validation: ./maya/maya-complete-notification-list From ./maya/maya-complete-notification-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./maya/maya-notification-parse-updates'] From ./maya/maya-complete-notification-list :['From ./maya/maya-notification-parse-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./maya/scrape-maya-notification-list" is successfully executed']]
Pipeline ID: maya/maya-reported-work-record
Dependency unsuccessful :Cannot run until dependency "./maya/scrape-maya-tase-companies" is successfully executed
Pipeline ID: maya/maya-tase-companies-current-management
Invalid dependency :Cannot run until dependency passes validation: ./people/all From ./people/all :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/all-appointments'] From ./people/all :['From ./people/company_appointments/all-appointments', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/media/themarker/themarker" is successfully executed']] From ./people/all :['From ./people/company_appointments/all-appointments', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-nominations']] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']]] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]]] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]]] From ./people/all :['From ./people/company_appointments/all-appointments', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-resignations']] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']]] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]]] From ./people/all :['From ./people/company_appointments/all-appointments', ['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]]] From ./people/all :['Dependency unsuccessful', 'Cannot run until dependency "./people/political_candidates/election_candidates21" is successfully executed']
Pipeline ID: people/aggregated
Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/all-appointments From ./people/company_appointments/all-appointments :['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/media/themarker/themarker" is successfully executed'] From ./people/company_appointments/all-appointments :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-nominations'] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-nominations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-nominations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]] From ./people/company_appointments/all-appointments :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-resignations'] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-resignations', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list']] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']]] From ./people/company_appointments/all-appointments :['From ./people/company_appointments/maya/maya-resignations', ['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]] Dependency unsuccessful :Cannot run until dependency "./people/political_candidates/election_candidates21" is successfully executed
Pipeline ID: people/all
Dependency unsuccessful :Cannot run until dependency "./people/company_appointments/media/themarker/themarker" is successfully executed Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-nominations From ./people/company_appointments/maya/maya-nominations :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list'] From ./people/company_appointments/maya/maya-nominations :['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']] From ./people/company_appointments/maya/maya-nominations :['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]] Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-resignations From ./people/company_appointments/maya/maya-resignations :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list'] From ./people/company_appointments/maya/maya-resignations :['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']] From ./people/company_appointments/maya/maya-resignations :['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]
Pipeline ID: people/company_appointments/all-appointments
Dependency unsuccessful :Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed
Pipeline ID: people/company_appointments/maya/collect-maya-forms-debug-stats
Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-nominations From ./people/company_appointments/maya/maya-nominations :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list'] From ./people/company_appointments/maya/maya-nominations :['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']] From ./people/company_appointments/maya/maya-nominations :['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]] Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-resignations From ./people/company_appointments/maya/maya-resignations :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list'] From ./people/company_appointments/maya/maya-resignations :['From ./people/company_appointments/maya/maya-list', ['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates']] From ./people/company_appointments/maya/maya-resignations :['From ./people/company_appointments/maya/maya-list', ['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]]
Pipeline ID: people/company_appointments/maya/correlate
Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates From ./people/company_appointments/maya/maya-notification-updates :['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']
Pipeline ID: people/company_appointments/maya/maya-list
Name, job and company description.
Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list From ./people/company_appointments/maya/maya-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates'] From ./people/company_appointments/maya/maya-list :['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]
Pipeline ID: people/company_appointments/maya/maya-nominations
id, date, company, notification_type, s3_object_name, url
Dependency unsuccessful :Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed
Pipeline ID: people/company_appointments/maya/maya-notification-updates
Name, job and company description.
Invalid dependency :Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-list From ./people/company_appointments/maya/maya-list :['Invalid dependency', 'Cannot run until dependency passes validation: ./people/company_appointments/maya/maya-notification-updates'] From ./people/company_appointments/maya/maya-list :['From ./people/company_appointments/maya/maya-notification-updates', ['Dependency unsuccessful', 'Cannot run until dependency "./people/company_appointments/maya/maya-notification-list" is successfully executed']]
Pipeline ID: people/company_appointments/maya/maya-resignations
Invalid dependency :Cannot run until dependency passes validation: ./procurement/calls_for_bids/jobiz-calls-for-bids From ./procurement/calls_for_bids/jobiz-calls-for-bids :['Invalid dependency', 'Cannot run until dependency passes validation: ./procurement/calls_for_bids/jobiz-all'] From ./procurement/calls_for_bids/jobiz-calls-for-bids :['From ./procurement/calls_for_bids/jobiz-all', ['Dependency unsuccessful', 'Cannot run until dependency "./procurement/calls_for_bids/jobiz-scraper" is successfully executed']]
Pipeline ID: procurement/calls_for_bids/calls-for-bids-all
Dependency unsuccessful :Cannot run until dependency "./procurement/calls_for_bids/jobiz-scraper" is successfully executed
Pipeline ID: procurement/calls_for_bids/jobiz-all
Invalid dependency :Cannot run until dependency passes validation: ./procurement/calls_for_bids/jobiz-all From ./procurement/calls_for_bids/jobiz-all :['Dependency unsuccessful', 'Cannot run until dependency "./procurement/calls_for_bids/jobiz-scraper" is successfully executed']
Pipeline ID: procurement/calls_for_bids/jobiz-calls-for-bids
Dependency unsuccessful :Cannot run until dependency "./supports/supports-2025" is successfully executed
Pipeline ID: supports/all
Pipeline ID: budgetkey/elasticsearch/index_entities
Pipeline ID: entities/entities
scraper
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/pyquery/pyquery.py", line 77, in fromstring
result = getattr(etree, meth)(context)
File "src/lxml/etree.pyx", line 3428, in lxml.etree.fromstring
File "src/lxml/parser.pxi", line 2066, in lxml.etree._parseMemoryDocument
File "src/lxml/parser.pxi", line 1921, in lxml.etree._parseDoc
File "src/lxml/parser.pxi", line 1937, in lxml.etree._parseDoc_unicode
File "src/lxml/parser.pxi", line 1136, in lxml.etree._BaseParser._parseUnicodeDoc
File "src/lxml/parser.pxi", line 647, in lxml.etree._ParserContext._handleParseResultDoc
File "src/lxml/parser.pxi", line 765, in lxml.etree._handleParseResult
File "src/lxml/parser.pxi", line 689, in lxml.etree._raiseParseError
File "<string>", line 1
lxml.etree.XMLSyntaxError: Document is empty, line 1, column 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/pipelines/bank_of_israel/scraper.py", line 124, in <module>
series_sum = get_all_series()
File "/datapackage_pipelines_budgetkey/pipelines/bank_of_israel/scraper.py", line 23, in get_all_series
series = get_series(sid)
File "/datapackage_pipelines_budgetkey/pipelines/bank_of_israel/scraper.py", line 36, in get_series
resp = pq(data.decode(response.encoding or 'utf8'))
File "/usr/local/lib/python3.9/site-packages/pyquery/pyquery.py", line 230, in __init__
elements = fromstring(context, self.parser)
File "/usr/local/lib/python3.9/site-packages/pyquery/pyquery.py", line 81, in fromstring
result = getattr(lxml.html, meth)(context)
File "/usr/local/lib/python3.9/site-packages/lxml/html/__init__.py", line 854, in fromstring
doc = document_fromstring(html, parser=parser, base_url=base_url, **kw)
File "/usr/local/lib/python3.9/site-packages/lxml/html/__init__.py", line 742, in document_fromstring
raise etree.ParserError(
lxml.etree.ParserError: Document is empty
Pipeline ID: bank_of_israel/bank_of_israel
dump_to_es
ERROR :DUMP TO ES ERROR ApiError(413, 'None')
ERROR :TB
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/processors/dump_to_es.py", line 146, in <module>
spew_flow(flow(ctx.parameters), ctx)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 181, in __exit__
spew(self.datapackage, self.resource_iterator, stats=self.stats)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 75, in spew
for rec in res:
File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/dumper_base.py", line 69, in row_counter
for row in iterator:
File "/usr/local/lib/python3.9/site-packages/tableschema_elasticsearch/storage.py", line 226, in write
for result, row in iter:
File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 453, in streaming_bulk
for data, (ok, info) in zip(
File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 359, in _process_bulk_chunk
yield from gen
File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 293, in _process_bulk_chunk_error
raise error
File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 343, in _process_bulk_chunk
resp = client.bulk(*args, operations=bulk_actions, **kwargs) # type: ignore[arg-type]
File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/utils.py", line 452, in wrapped
return api(*args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/__init__.py", line 828, in bulk
return self.perform_request( # type: ignore[return-value]
File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/_base.py", line 271, in perform_request
response = self._perform_request(
File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/_base.py", line 351, in _perform_request
raise HTTP_EXCEPTIONS.get(meta.status, ApiError)(
elasticsearch.ApiError: ApiError(413, 'None')
ERROR :Traceback (most recent call last):
ERROR :File "/datapackage_pipelines_budgetkey/processors/dump_to_es.py", line 146, in <module>
ERROR :spew_flow(flow(ctx.parameters), ctx)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 181, in __exit__
ERROR :spew(self.datapackage, self.resource_iterator, stats=self.stats)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 75, in spew
ERROR :for rec in res:
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/dumper_base.py", line 69, in row_counter
ERROR :for row in iterator:
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_elasticsearch/storage.py", line 226, in write
ERROR :for result, row in iter:
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 453, in streaming_bulk
ERROR :for data, (ok, info) in zip(
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 359, in _process_bulk_chunk
ERROR :yield from gen
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 293, in _process_bulk_chunk_error
ERROR :raise error
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/helpers/actions.py", line 343, in _process_bulk_chunk
ERROR :resp = client.bulk(*args, operations=bulk_actions, **kwargs) # type: ignore[arg-type]
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/utils.py", line 452, in wrapped
ERROR :return api(*args, **kwargs)
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/__init__.py", line 828, in bulk
ERROR :return self.perform_request( # type: ignore[return-value]
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/_base.py", line 271, in perform_request
ERROR :response = self._perform_request(
ERROR :File "/usr/local/lib/python3.9/site-packages/elasticsearch/_sync/client/_base.py", line 351, in _perform_request
ERROR :raise HTTP_EXCEPTIONS.get(meta.status, ApiError)(
ERROR :elasticsearch
ERROR :.
ERROR :ApiError
ERROR ::
ERROR :ApiError(413, 'None')
Pipeline ID: budgetkey/elasticsearch/index_gov_decisions
get_transactions
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/pipelines/donations/get_transactions.py", line 150, in <module>
spew(dp, process_resources(res_iter))
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 75, in spew
for rec in res:
File "/datapackage_pipelines_budgetkey/pipelines/donations/get_transactions.py", line 109, in get_transactions
for resp in self.get_for_candidate(cid):
File "/datapackage_pipelines_budgetkey/pipelines/donations/get_transactions.py", line 95, in get_for_candidate
resp = self.get_for_range(cid, range_start, range_end)
File "/datapackage_pipelines_budgetkey/pipelines/donations/get_transactions.py", line 70, in get_for_range
assert len(resp) == 6
AssertionError
Pipeline ID: donations/transactions
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/education/tochniyot/scraper.py", line 103, in flow
return Flow(scrape(),
File "/datapackage_pipelines_budgetkey/pipelines/education/tochniyot/scraper.py", line 65, in scrape
education_programs_count = get_education_programs_count()
File "/datapackage_pipelines_budgetkey/pipelines/education/tochniyot/scraper.py", line 56, in get_education_programs_count
res = send_tochniyot_request()
File "/datapackage_pipelines_budgetkey/pipelines/education/tochniyot/scraper.py", line 80, in send_tochniyot_request
resp.raise_for_status()
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 1021, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: https://apps.education.gov.il/TyhNet/ClientWs/TochnitCh.asmx/IturTochnitChByMeafyenim
Pipeline ID: education/tochniyot/tochniyot
scraper
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/pipelines/entities/moital/scraper.py", line 79, in <module>
spew(datapackage, [scrape()])
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 75, in spew
for rec in res:
File "/datapackage_pipelines_budgetkey/pipelines/entities/moital/scraper.py", line 32, in scrape
desired_capabilities=DesiredCapabilities.PHANTOMJS)
AttributeError: type object 'DesiredCapabilities' has no attribute 'PHANTOMJS'
Pipeline ID: entities/moital/moital_service_providers
stream_remote_resources
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 174, in _new_conn
conn = connection.create_connection(
File "/usr/local/lib/python3.9/site-packages/urllib3/util/connection.py", line 72, in create_connection
for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
File "/usr/local/lib/python3.9/socket.py", line 954, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno -2] Name or service not known
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 714, in urlopen
httplib_response = self._make_request(
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 415, in _make_request
conn.request(method, url, **httplib_request_kw)
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 244, in request
super(HTTPConnection, self).request(method, url, body=body, headers=headers)
File "/usr/local/lib/python3.9/http/client.py", line 1285, in request
self._send_request(method, url, body, headers, encode_chunked)
File "/usr/local/lib/python3.9/http/client.py", line 1331, in _send_request
self.endheaders(body, encode_chunked=encode_chunked)
File "/usr/local/lib/python3.9/http/client.py", line 1280, in endheaders
self._send_output(message_body, encode_chunked=encode_chunked)
File "/usr/local/lib/python3.9/http/client.py", line 1040, in _send_output
self.send(msg)
File "/usr/local/lib/python3.9/http/client.py", line 980, in send
self.connect()
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 205, in connect
conn = self._new_conn()
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 186, in _new_conn
raise NewConnectionError(
urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPConnection object at 0x7f5cec45c580>: Failed to establish a new connection: [Errno -2] Name or service not known
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 486, in send
resp = conn.urlopen(
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 798, in urlopen
retries = retries.increment(
File "/usr/local/lib/python3.9/site-packages/urllib3/util/retry.py", line 592, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPConnectionPool(host='employment.molsa.gov.il', port=80): Max retries exceeded with url: /Employment/SafetyAndHealth/Enforcement/Documents/%D7%A6%D7%95%D7%95%D7%99%20%D7%91%D7%98%D7%99%D7%97%D7%95%D7%AA.xlsx (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f5cec45c580>: Failed to establish a new connection: [Errno -2] Name or service not known'))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 63, in load
bytes = _RemoteStream(source, self.__http_session, self.__http_timeout).open()
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 121, in open
self.seek(0)
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 142, in seek
self.__response = self.__session.get(self.__source, stream=True, timeout=self.__timeout)
File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 602, in get
return self.request("GET", url, **kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 589, in request
resp = self.send(prep, **send_kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 703, in send
r = adapter.send(request, **kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 519, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPConnectionPool(host='employment.molsa.gov.il', port=80): Max retries exceeded with url: /Employment/SafetyAndHealth/Enforcement/Documents/%D7%A6%D7%95%D7%95%D7%99%20%D7%91%D7%98%D7%99%D7%97%D7%95%D7%AA.xlsx (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f5cec45c580>: Failed to establish a new connection: [Errno -2] Name or service not known'))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 244, in <module>
rows = stream_reader(resource, url, ignore_missing or url == "", limit_rows,
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 199, in stream_reader
schema, headers, columns, stream, close = get_opener(url, _resource)()
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 175, in opener
_stream.open()
File "/usr/local/lib/python3.9/site-packages/tabulator/stream.py", line 425, in open
self.__parser.open(source, encoding=self.__encoding)
File "/usr/local/lib/python3.9/site-packages/tabulator/parsers/xlsx.py", line 81, in open
source_bytes = self.__loader.load(source, mode="b", encoding=encoding)
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 72, in load
raise exceptions.HTTPError(str(exception))
tabulator.exceptions.HTTPError: HTTPConnectionPool(host='employment.molsa.gov.il', port=80): Max retries exceeded with url: /Employment/SafetyAndHealth/Enforcement/Documents/%D7%A6%D7%95%D7%95%D7%99%20%D7%91%D7%98%D7%99%D7%97%D7%95%D7%AA.xlsx (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f5cec45c580>: Failed to establish a new connection: [Errno -2] Name or service not known'))
Pipeline ID: entities/safety-orders/safety-orders
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/knesset/knesset_legal_advisor.py", line 44, in flow
document = gcl.download(url, use_curl=True, outfile=filename)
File "/datapackage_pipelines_budgetkey/common/google_chrome.py", line 162, in download
assert False, 'Failed to download file, %r' % downloads
AssertionError: Failed to download file, ['']
DEBUG :[chan 16] Max packet in: 32768 bytes
DEBUG :[chan 16] Max packet out: 32768 bytes
DEBUG :Secsh channel 16 opened.
DEBUG :[chan 16] Sesch channel 16 request ok
DEBUG :[chan 16] EOF received (16)
DEBUG :[chan 16] EOF sent (16)
DEBUG :Dropping user packet because connection is dead.
DEBUG :Dropping user packet because connection is dead.
Pipeline ID: knesset/knesset_legal_advisor
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/knesset/knesset_legal_advisor_letters.py", line 44, in flow
document = gcl.download(url, use_curl=True, outfile=filename)
File "/datapackage_pipelines_budgetkey/common/google_chrome.py", line 162, in download
assert False, 'Failed to download file, %r' % downloads
AssertionError: Failed to download file, ['']
DEBUG :[chan 16] Max packet in: 32768 bytes
DEBUG :[chan 16] Max packet out: 32768 bytes
DEBUG :Secsh channel 16 opened.
DEBUG :[chan 16] Sesch channel 16 request ok
DEBUG :[chan 16] EOF received (16)
DEBUG :[chan 16] EOF sent (16)
DEBUG :Dropping user packet because connection is dead.
Pipeline ID: knesset/knesset_legal_advisor_letters
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 79, in _process
self.datapackage = self.process_datapackage(self.datapackage)
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 112, in process_datapackage
raise self.exc
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 60, in describe
sample = list(itertools.islice(self.iterable, self.SAMPLE_SIZE))
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 89, in handle_iterable
for x in self.iterable:
File "/datapackage_pipelines_budgetkey/pipelines/maya/scrape_maya_notification_list.py", line 153, in scrape_maya_notification_list
first_date, last_date = get_existing_records()
File "/datapackage_pipelines_budgetkey/pipelines/maya/scrape_maya_notification_list.py", line 31, in get_existing_records
rs = con.execute(text("SELECT min(date), max(date) FROM maya_notifications where parser_version=:v"), v=PARSER_VERSION)
TypeError: execute() got an unexpected keyword argument 'v'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 15, in <module>
spew_flow(flow, ctx)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/utilities/flow_utils.py", line 46, in spew_flow
datastream = flow.datastream()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/flow.py", line 19, in datastream
return self._chain(ds)._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 86, in _process
self.raise_exception(exception)
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 96, in raise_exception
raise error from cause
dataflows.base.exceptions.ProcessorError: Errored in processor iterable_loader in position #9: execute() got an unexpected keyword argument 'v'
Pipeline ID: maya/scrape-maya-notification-list
dump_to_path
ERROR :Data Package validation error: Descriptor validation error: [] is too short at "schema/fields" in descriptor and at "properties/schema/properties/fields/minItems" in profile
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/dump_to_path.py", line 22, in <module>
with ingest() as ctx:
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 186, in ingest
params, datapackage, resource_iterator = _ingest(debug=debug)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 44, in _ingest
datapackage, resource_iterator, dependency_dp = process_input(sys.stdin, validate, debug)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/input_processor.py", line 88, in process_input
datapackage.validate(dp_to_validate)
File "/usr/local/lib/python3.9/site-packages/datapackage/validate.py", line 28, in validate
Package(descriptor, strict=True)
File "/usr/local/lib/python3.9/site-packages/datapackage/package.py", line 126, in __init__
self.__build()
File "/usr/local/lib/python3.9/site-packages/datapackage/package.py", line 534, in __build
updated_resource = Resource(descriptor,
File "/usr/local/lib/python3.9/site-packages/datapackage/resource.py", line 90, in __init__
self.__build()
File "/usr/local/lib/python3.9/site-packages/datapackage/resource.py", line 648, in __build
raise exception
File "/usr/local/lib/python3.9/site-packages/datapackage/resource.py", line 643, in __build
self.__profile.validate(self.__current_descriptor)
File "/usr/local/lib/python3.9/site-packages/datapackage/profile.py", line 89, in validate
raise exceptions.ValidationError(message, errors=errors)
datapackage.exceptions.ValidationError: There are 1 validation errors (see exception.errors)
Pipeline ID: maya/scrape-maya-tase-companies
scrape-notification-list
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 971, in json
return complexjson.loads(self.text, **kwargs)
File "/usr/local/lib/python3.9/json/__init__.py", line 346, in loads
return _default_decoder.decode(s)
File "/usr/local/lib/python3.9/json/decoder.py", line 337, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/usr/local/lib/python3.9/json/decoder.py", line 355, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/pipelines/people/company_appointments/maya/scrape-notification-list.py", line 55, in _maya_api_call
return res.json()
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 975, in json
raise RequestsJSONDecodeError(e.msg, e.doc, e.pos)
requests.exceptions.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/pipelines/people/company_appointments/maya/scrape-notification-list.py", line 130, in <module>
spew(datapackage, [collect()])
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 75, in spew
for rec in res:
File "/datapackage_pipelines_budgetkey/pipelines/people/company_appointments/maya/scrape-notification-list.py", line 113, in collect
yield from _collect_date_range(year_start, year_end)
File "/datapackage_pipelines_budgetkey/pipelines/people/company_appointments/maya/scrape-notification-list.py", line 75, in _collect_date_range
res = _maya_api_call(date_from, date_to, current_page)
File "/datapackage_pipelines_budgetkey/pipelines/people/company_appointments/maya/scrape-notification-list.py", line 57, in _maya_api_call
raise Exception("Failed to Call Maya API for date_from:{} date_to:{} page_num:{}".format(date_from, date_to, page_num)) from e
Exception: Failed to Call Maya API for date_from:2006-01-01 date_to:2007-01-01 page_num:0
Pipeline ID: people/company_appointments/maya/maya-notification-list
Name, job and company description.
stream_remote_resources
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 63, in load
bytes = _RemoteStream(source, self.__http_session, self.__http_timeout).open()
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 121, in open
self.seek(0)
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 143, in seek
self.__response.raise_for_status()
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 1021, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://www.themarker.com/st/inter/DB/tm/2015/minuyim30.xlsx
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 244, in <module>
rows = stream_reader(resource, url, ignore_missing or url == "", limit_rows,
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 199, in stream_reader
schema, headers, columns, stream, close = get_opener(url, _resource)()
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/stream_remote_resources.py", line 175, in opener
_stream.open()
File "/usr/local/lib/python3.9/site-packages/tabulator/stream.py", line 425, in open
self.__parser.open(source, encoding=self.__encoding)
File "/usr/local/lib/python3.9/site-packages/tabulator/parsers/xlsx.py", line 81, in open
source_bytes = self.__loader.load(source, mode="b", encoding=encoding)
File "/usr/local/lib/python3.9/site-packages/tabulator/loaders/remote.py", line 72, in load
raise exceptions.HTTPError(str(exception))
tabulator.exceptions.HTTPError: 404 Client Error: Not Found for url: https://www.themarker.com/st/inter/DB/tm/2015/minuyim30.xlsx
Pipeline ID: people/company_appointments/media/themarker/themarker
dump_to_sql
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :psycopg2.errors
ERROR :.
ERROR :InternalError_
ERROR ::
ERROR :could not open relation with OID 497924577
ERROR :The above exception was the direct cause of the following exception:
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/dump_to_sql.py", line 19, in <module>
ERROR :spew_flow(flow(ctx.parameters), ctx)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 181, in __exit__
ERROR :spew(self.datapackage, self.resource_iterator, stats=self.stats)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 68, in spew
ERROR :for res in resources_iterator:
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 68, in <genexpr>
ERROR :res_iter = (it if isinstance(it, ResourceWrapper) else ResourceWrapper(res, it)
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/dumper_base.py", line 82, in process_resources
ERROR :ret = self.process_resource(
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/to_sql.py", line 108, in process_resource
ERROR :storage.delete('')
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 183, in delete
ERROR :self.__reflect()
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 278, in __reflect
ERROR :self.__metadata.reflect(only=only, bind=self.__engine)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 5752, in reflect
ERROR :_reflect_info = insp._get_reflection_info(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 2018, in _get_reflection_info
ERROR :check_constraints=run(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1994, in run
ERROR :res = meth(filter_names=_fn, **kw)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1457, in get_multi_check_constraints
ERROR :self.dialect.get_multi_check_constraints(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/dialects/postgresql/base.py", line 4677, in get_multi_check_constraints
ERROR :result = connection.execute(query, params)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1412, in execute
ERROR :return meth(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 483, in _execute_on_connection
ERROR :return connection._execute_clauseelement(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1635, in _execute_clauseelement
ERROR :ret = self._execute_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1844, in _execute_context
ERROR :return self._exec_single_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1984, in _exec_single_context
ERROR :self._handle_dbapi_exception(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 2339, in _handle_dbapi_exception
ERROR :raise sqlalchemy_exception.with_traceback(exc_info[2]) from e
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :sqlalchemy.exc
ERROR :.
ERROR :InternalError
ERROR ::
ERROR :(psycopg2.errors.InternalError_) could not open relation with OID 497924577
[SQL: SELECT pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname, CASE WHEN (pg_catalog.pg_constraint.oid IS NOT NULL) THEN pg_catalog.pg_get_constraintdef(pg_catalog.pg_constraint.oid, %(pg_get_constraintdef_1)s) END AS anon_1, pg_catalog.pg_description.description
FROM pg_catalog.pg_class LEFT OUTER JOIN pg_catalog.pg_constraint ON pg_catalog.pg_class.oid = pg_catalog.pg_constraint.conrelid AND pg_catalog.pg_constraint.contype = %(contype_1)s LEFT OUTER JOIN pg_catalog.pg_description ON pg_catalog.pg_description.objoid = pg_catalog.pg_constraint.oid JOIN pg_catalog.pg_namespace ON pg_catalog.pg_namespace.oid = pg_catalog.pg_class.relnamespace
WHERE pg_catalog.pg_class.relkind = ANY (ARRAY[%(param_1)s, %(param_2)s, %(param_3)s]) AND pg_catalog.pg_table_is_visible(pg_catalog.pg_class.oid) AND pg_catalog.pg_namespace.nspname != %(nspname_1)s ORDER BY pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname]
[parameters: {'pg_get_constraintdef_1': True, 'contype_1': 'c', 'param_1': 'r', 'param_2': 'p', 'param_3': 'f', 'nspname_1': 'pg_catalog'}]
(Background on this error at: https://sqlalche.me/e/20/2j85)
Pipeline ID: people/company_appointments/state_owned/correlate
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 79, in _process
self.datapackage = self.process_datapackage(self.datapackage)
File "/usr/local/lib/python3.9/site-packages/dataflows/processors/set_type.py", line 71, in process_datapackage
assert added, 'Failed to find field {} in schema'.format(self.name)
AssertionError: Failed to find field re.compile('^party_letter$') in schema
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 15, in <module>
spew_flow(flow, ctx)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/utilities/flow_utils.py", line 46, in spew_flow
datastream = flow.datastream()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/flow.py", line 19, in datastream
return self._chain(ds)._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
[Previous line repeated 4 more times]
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 86, in _process
self.raise_exception(exception)
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 96, in raise_exception
raise error from cause
dataflows.base.exceptions.ProcessorError: Errored in processor set_type in position #3: Failed to find field re.compile('^party_letter$') in schema
Pipeline ID: people/political_candidates/election_candidates21
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 79, in _process
self.datapackage = self.process_datapackage(self.datapackage)
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 112, in process_datapackage
raise self.exc
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 60, in describe
sample = list(itertools.islice(self.iterable, self.SAMPLE_SIZE))
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 89, in handle_iterable
for x in self.iterable:
File "/datapackage_pipelines_budgetkey/pipelines/procurement/calls_for_bids/m_tmicha.py", line 47, in m_tmicha_scraper
assert total > 0
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 15, in <module>
spew_flow(flow, ctx)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/utilities/flow_utils.py", line 46, in spew_flow
datastream = flow.datastream()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/flow.py", line 19, in datastream
return self._chain(ds)._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 86, in _process
self.raise_exception(exception)
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 96, in raise_exception
raise error from cause
dataflows.base.exceptions.ProcessorError: Errored in processor iterable_loader in position #1:
Pipeline ID: procurement/calls_for_bids/calls-for-bids-moh
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 174, in _new_conn
conn = connection.create_connection(
File "/usr/local/lib/python3.9/site-packages/urllib3/util/connection.py", line 72, in create_connection
for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
File "/usr/local/lib/python3.9/socket.py", line 954, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno -5] No address associated with hostname
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 714, in urlopen
httplib_response = self._make_request(
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 403, in _make_request
self._validate_conn(conn)
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 1053, in _validate_conn
conn.connect()
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 363, in connect
self.sock = conn = self._new_conn()
File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 186, in _new_conn
raise NewConnectionError(
urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f5a65bd0760>: Failed to establish a new connection: [Errno -5] No address associated with hostname
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 486, in send
resp = conn.urlopen(
File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 798, in urlopen
retries = retries.increment(
File "/usr/local/lib/python3.9/site-packages/urllib3/util/retry.py", line 592, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='jobiz.gov.il', port=443): Max retries exceeded with url: /ajax/results/%D7%94%D7%95%D7%93%D7%A2%D7%94%20%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA/0?ie=0&typeie=%D7%94%D7%95%D7%93%D7%A2%D7%94+%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA&search=Array (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5a65bd0760>: Failed to establish a new connection: [Errno -5] No address associated with hostname'))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 79, in _process
self.datapackage = self.process_datapackage(self.datapackage)
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 112, in process_datapackage
raise self.exc
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 60, in describe
sample = list(itertools.islice(self.iterable, self.SAMPLE_SIZE))
File "/usr/local/lib/python3.9/site-packages/dataflows/helpers/iterable_loader.py", line 89, in handle_iterable
for x in self.iterable:
File "/datapackage_pipelines_budgetkey/pipelines/procurement/calls_for_bids/jobiz.py", line 19, in fetch_results
content = requests.get(URL.format(index)).json()
File "/usr/local/lib/python3.9/site-packages/requests/api.py", line 73, in get
return request("get", url, params=params, **kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/api.py", line 59, in request
return session.request(method=method, url=url, **kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 589, in request
resp = self.send(prep, **send_kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 703, in send
r = adapter.send(request, **kwargs)
File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 519, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='jobiz.gov.il', port=443): Max retries exceeded with url: /ajax/results/%D7%94%D7%95%D7%93%D7%A2%D7%94%20%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA/0?ie=0&typeie=%D7%94%D7%95%D7%93%D7%A2%D7%94+%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA&search=Array (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5a65bd0760>: Failed to establish a new connection: [Errno -5] No address associated with hostname'))
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 15, in <module>
spew_flow(flow, ctx)
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/utilities/flow_utils.py", line 46, in spew_flow
datastream = flow.datastream()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/flow.py", line 19, in datastream
return self._chain(ds)._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 75, in _process
datastream = self.source._process()
[Previous line repeated 5 more times]
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 86, in _process
self.raise_exception(exception)
File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 96, in raise_exception
raise error from cause
dataflows.base.exceptions.ProcessorError: Errored in processor iterable_loader in position #1: HTTPSConnectionPool(host='jobiz.gov.il', port=443): Max retries exceeded with url: /ajax/results/%D7%94%D7%95%D7%93%D7%A2%D7%94%20%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA/0?ie=0&typeie=%D7%94%D7%95%D7%93%D7%A2%D7%94+%D7%A6%D7%99%D7%91%D7%95%D7%A8%D7%99%D7%AA&search=Array (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5a65bd0760>: Failed to establish a new connection: [Errno -5] No address associated with hostname'))
Pipeline ID: procurement/calls_for_bids/jobiz-scraper
dump.to_sql
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :psycopg2.errors
ERROR :.
ERROR :InternalError_
ERROR ::
ERROR :could not open relation with OID 497924577
ERROR :The above exception was the direct cause of the following exception:
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/dump/to_sql.py", line 15, in <module>
ERROR :spew_flow(flow(ctx.parameters), ctx)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 181, in __exit__
ERROR :spew(self.datapackage, self.resource_iterator, stats=self.stats)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 68, in spew
ERROR :for res in resources_iterator:
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 68, in <genexpr>
ERROR :res_iter = (it if isinstance(it, ResourceWrapper) else ResourceWrapper(res, it)
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/dumper_base.py", line 82, in process_resources
ERROR :ret = self.process_resource(
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/to_sql.py", line 108, in process_resource
ERROR :storage.delete('')
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 183, in delete
ERROR :self.__reflect()
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 278, in __reflect
ERROR :self.__metadata.reflect(only=only, bind=self.__engine)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 5752, in reflect
ERROR :_reflect_info = insp._get_reflection_info(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 2018, in _get_reflection_info
ERROR :check_constraints=run(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1994, in run
ERROR :res = meth(filter_names=_fn, **kw)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1457, in get_multi_check_constraints
ERROR :self.dialect.get_multi_check_constraints(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/dialects/postgresql/base.py", line 4677, in get_multi_check_constraints
ERROR :result = connection.execute(query, params)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1412, in execute
ERROR :return meth(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 483, in _execute_on_connection
ERROR :return connection._execute_clauseelement(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1635, in _execute_clauseelement
ERROR :ret = self._execute_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1844, in _execute_context
ERROR :return self._exec_single_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1984, in _exec_single_context
ERROR :self._handle_dbapi_exception(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 2339, in _handle_dbapi_exception
ERROR :raise sqlalchemy_exception.with_traceback(exc_info[2]) from e
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :sqlalchemy.exc
ERROR :.
ERROR :InternalError
ERROR ::
ERROR :(psycopg2.errors.InternalError_) could not open relation with OID 497924577
[SQL: SELECT pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname, CASE WHEN (pg_catalog.pg_constraint.oid IS NOT NULL) THEN pg_catalog.pg_get_constraintdef(pg_catalog.pg_constraint.oid, %(pg_get_constraintdef_1)s) END AS anon_1, pg_catalog.pg_description.description
FROM pg_catalog.pg_class LEFT OUTER JOIN pg_catalog.pg_constraint ON pg_catalog.pg_class.oid = pg_catalog.pg_constraint.conrelid AND pg_catalog.pg_constraint.contype = %(contype_1)s LEFT OUTER JOIN pg_catalog.pg_description ON pg_catalog.pg_description.objoid = pg_catalog.pg_constraint.oid JOIN pg_catalog.pg_namespace ON pg_catalog.pg_namespace.oid = pg_catalog.pg_class.relnamespace
WHERE pg_catalog.pg_class.relkind = ANY (ARRAY[%(param_1)s, %(param_2)s, %(param_3)s]) AND pg_catalog.pg_table_is_visible(pg_catalog.pg_class.oid) AND pg_catalog.pg_namespace.nspname != %(nspname_1)s ORDER BY pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname]
[parameters: {'pg_get_constraintdef_1': True, 'contype_1': 'c', 'param_1': 'r', 'param_2': 'p', 'param_3': 'f', 'nspname_1': 'pg_catalog'}]
(Background on this error at: https://sqlalche.me/e/20/2j85)
Pipeline ID: procurement/tenders/processed
dump.to_sql
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :psycopg2.errors
ERROR :.
ERROR :InternalError_
ERROR ::
ERROR :could not open relation with OID 497993584
ERROR :The above exception was the direct cause of the following exception:
ERROR :Traceback (most recent call last):
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/dump/to_sql.py", line 15, in <module>
ERROR :spew_flow(flow(ctx.parameters), ctx)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 181, in __exit__
ERROR :spew(self.datapackage, self.resource_iterator, stats=self.stats)
ERROR :File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/wrapper/wrapper.py", line 68, in spew
ERROR :for res in resources_iterator:
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/base/datastream_processor.py", line 68, in <genexpr>
ERROR :res_iter = (it if isinstance(it, ResourceWrapper) else ResourceWrapper(res, it)
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/dumper_base.py", line 82, in process_resources
ERROR :ret = self.process_resource(
ERROR :File "/usr/local/lib/python3.9/site-packages/dataflows/processors/dumpers/to_sql.py", line 108, in process_resource
ERROR :storage.delete('')
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 183, in delete
ERROR :self.__reflect()
ERROR :File "/usr/local/lib/python3.9/site-packages/tableschema_sql/storage.py", line 278, in __reflect
ERROR :self.__metadata.reflect(only=only, bind=self.__engine)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 5752, in reflect
ERROR :_reflect_info = insp._get_reflection_info(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 2018, in _get_reflection_info
ERROR :check_constraints=run(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1994, in run
ERROR :res = meth(filter_names=_fn, **kw)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 1457, in get_multi_check_constraints
ERROR :self.dialect.get_multi_check_constraints(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/dialects/postgresql/base.py", line 4677, in get_multi_check_constraints
ERROR :result = connection.execute(query, params)
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1412, in execute
ERROR :return meth(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 483, in _execute_on_connection
ERROR :return connection._execute_clauseelement(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1635, in _execute_clauseelement
ERROR :ret = self._execute_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1844, in _execute_context
ERROR :return self._exec_single_context(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1984, in _exec_single_context
ERROR :self._handle_dbapi_exception(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 2339, in _handle_dbapi_exception
ERROR :raise sqlalchemy_exception.with_traceback(exc_info[2]) from e
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1965, in _exec_single_context
ERROR :self.dialect.do_execute(
ERROR :File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 921, in do_execute
ERROR :cursor.execute(statement, parameters)
ERROR :sqlalchemy.exc
ERROR :.
ERROR :InternalError
ERROR ::
ERROR :(psycopg2.errors.InternalError_) could not open relation with OID 497993584
[SQL: SELECT pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname, CASE WHEN (pg_catalog.pg_constraint.oid IS NOT NULL) THEN pg_catalog.pg_get_constraintdef(pg_catalog.pg_constraint.oid, %(pg_get_constraintdef_1)s) END AS anon_1, pg_catalog.pg_description.description
FROM pg_catalog.pg_class LEFT OUTER JOIN pg_catalog.pg_constraint ON pg_catalog.pg_class.oid = pg_catalog.pg_constraint.conrelid AND pg_catalog.pg_constraint.contype = %(contype_1)s LEFT OUTER JOIN pg_catalog.pg_description ON pg_catalog.pg_description.objoid = pg_catalog.pg_constraint.oid JOIN pg_catalog.pg_namespace ON pg_catalog.pg_namespace.oid = pg_catalog.pg_class.relnamespace
WHERE pg_catalog.pg_class.relkind = ANY (ARRAY[%(param_1)s, %(param_2)s, %(param_3)s]) AND pg_catalog.pg_table_is_visible(pg_catalog.pg_class.oid) AND pg_catalog.pg_namespace.nspname != %(nspname_1)s ORDER BY pg_catalog.pg_class.relname, pg_catalog.pg_constraint.conname]
[parameters: {'pg_get_constraintdef_1': True, 'contype_1': 'c', 'param_1': 'r', 'param_2': 'p', 'param_3': 'f', 'nspname_1': 'pg_catalog'}]
(Background on this error at: https://sqlalche.me/e/20/2j85)
Pipeline ID: supports/by-request-year
sample
Traceback (most recent call last):
File "/datapackage_pipelines_budgetkey/processors/sample.py", line 11, in <module>
res_name = parameters.get('resource', datapackage['resources'][0]['name'])
IndexError: list index out of range
Pipeline ID: supports/criteria/support-criteria
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 180, in flow
DF.load(wrapper(year), format='csv',
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 21, in wrapper
return scraper(gcd, year)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 159, in scraper
get_results_for_column(driver, rects[i])
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 96, in get_results_for_column
column.click()
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/remote/webelement.py", line 89, in click
self._execute(Command.CLICK_ELEMENT)
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/remote/webelement.py", line 773, in _execute
return self._parent.execute(command, params)
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/remote/webdriver.py", line 430, in execute
self.error_handler.check_response(response)
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/remote/errorhandler.py", line 247, in check_response
raise exception_class(message, screen, stacktrace)
selenium.common.exceptions.ElementClickInterceptedException: Message: element click intercepted: Element <rect width="19.408288043478258" height="50" fill="#0d6e84" stroke="#ffffff" stroke-width="1"></rect> is not clickable at point (394, 289). Other element would receive the click: <path d="M-4,0 A4,4 0 1,0 4,0 A4,4 0 1,0 -4,0z" fill="#ff8000" stroke="#ffffff" stroke-width="1"></path>
(Session info: chrome=88.0.4324.96)
Stacktrace:
#0 0x5ad524986199 <unknown>
WARNING :FAILED to teardown google-chrome
Pipeline ID: supports/supports-2023
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 180, in flow
DF.load(wrapper(year), format='csv',
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 21, in wrapper
return scraper(gcd, year)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 142, in scraper
chart = get_chart(driver)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 34, in get_chart
chart = WebDriverWait(driver, 60).until(
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/support/wait.py", line 87, in until
raise TimeoutException(message, screen, stacktrace)
selenium.common.exceptions.TimeoutException: Message:
Stacktrace:
#0 0x56ea53123199 <unknown>
WARNING :FAILED to teardown google-chrome
Pipeline ID: supports/supports-2024
flow
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/datapackage_pipelines/specs/../lib/flow.py", line 13, in <module>
flow = flow_module.flow(parameters, datapackage, resources, ctx.stats)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 180, in flow
DF.load(wrapper(year), format='csv',
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 21, in wrapper
return scraper(gcd, year)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 142, in scraper
chart = get_chart(driver)
File "/datapackage_pipelines_budgetkey/pipelines/supports/scraper.py", line 34, in get_chart
chart = WebDriverWait(driver, 60).until(
File "/usr/local/lib/python3.9/site-packages/selenium/webdriver/support/wait.py", line 87, in until
raise TimeoutException(message, screen, stacktrace)
selenium.common.exceptions.TimeoutException: Message:
Stacktrace:
#0 0x63ebe378b199 <unknown>
WARNING :FAILED to teardown google-chrome
Pipeline ID: supports/supports-2025
Pipeline ID: activities/social_services/social_service_reports
Pipeline ID: activities/social_services/social_service_suppliers
Pipeline ID: activities/social_services/social_service_tenders
Pipeline ID: activities/social_services/social_services
Pipeline ID: budget/municipal/fetcher
Pipeline ID: budget/national/changes/explanations/all
Pipeline ID: budget/national/changes/original/committee-zipfile
Pipeline ID: budget/national/changes/original/current-year-fixes
Pipeline ID: budget/national/changes/original/national-budget-changes
Pipeline ID: budget/national/changes/processed/national-budget-changes-aggregated
Pipeline ID: budget/national/changes/processed/transactions
We get the data the Ministry of Finance publishes in data.gov.il every year. It comes in XLS format, with one row per TAKANA and phase (original, approved, executed). In this pipeline we create from each triplet a single row that has all the data.
Pipeline ID: budget/national/original/national-budgets
This pipeline joins the different phases of the budget (allocated, revised and executed). In the original file there's a separate row for each of the phases. We like it better as a single row with all phase info. Another thing this pipeline does is to rename the column titles - to more friendly English names. Also, it create rows for all hierarchies - where upper hierarchies (2, 4 & 6 digits) are plain aggregations of the 8-digit items they contain.
Pipeline ID: budget/national/processed/aggregated-yearly
Pipeline ID: budget/national/processed/category-explanations
This pipeline joins budget items that span across years.
Pipeline ID: budget/national/processed/connected-items-explained
This pipeline joins budget items that span across years.
Pipeline ID: budget/national/processed/connected-national-budgets
Pipeline ID: budget/national/processed/just-the-total
Pipeline ID: budget/national/processed/roof-names
This pipeline joins the budget data to itself so that each item has a list of its immediate children.
Pipeline ID: budget/national/processed/with-extras
Pipeline ID: budgetkey/analysis/spending/publisher_entity_analysis
Pipeline ID: budgetkey/analysis/spending/publisher_foa_analysis
Pipeline ID: budgetkey/analysis/spending/united
Pipeline ID: budgetkey/budget-functional-aggregates
Pipeline ID: budgetkey/documents
Pipeline ID: budgetkey/elasticsearch/index_activities
Pipeline ID: budgetkey/elasticsearch/index_budget
Pipeline ID: budgetkey/elasticsearch/index_contract_spending
Pipeline ID: budgetkey/elasticsearch/index_muni_budgets
Pipeline ID: budgetkey/elasticsearch/index_muni_tenders
Pipeline ID: budgetkey/elasticsearch/index_national_budget_changes
Pipeline ID: budgetkey/elasticsearch/index_reports
Pipeline ID: budgetkey/elasticsearch/index_support_programs
Pipeline ID: budgetkey/elasticsearch/index_supports
Pipeline ID: budgetkey/elasticsearch/index_tenders
Pipeline ID: budgetkey/elasticsearch/index_units
Pipeline ID: budgetkey/emails/emails
Pipeline ID: donations/candidates
Pipeline ID: donations/parties
Pipeline ID: entities/all
Pipeline ID: entities/associations/guidestar/guidestar
Pipeline ID: entities/associations/guidestar/guidestar-scraper
רשימת כלל העמותות הפעילות, מתעדכן שבועית.
Pipeline ID: entities/associations/registrar/registry
Pipeline ID: entities/companies/registrar/details
רשימת כלל החברות הפעילות, מתעדכן שבועית.
Pipeline ID: entities/companies/registrar/scraper
רשימת אגודות שיתופיות, מתוך ממשק נתונים של מאגר של משרד הכלכלה / האגף לאיגוד שיתופי, מתעדכן שבועית
Pipeline ID: entities/cooperatives/cooperatives
Pipeline ID: entities/endowments/endowments
Pipeline ID: entities/fingerprints
Pipeline ID: entities/foi_orgs/foi_orgs
רשימת אגודות עותומניות, מתוך תשובה לבקשת חופש מידע
Pipeline ID: entities/ottoman/ottoman-association-registry
Pipeline ID: entities/partnerships/partnerships
Pipeline ID: entities/special/registry
Pipeline ID: ezvonot/ezvonot_accounts
Pipeline ID: ezvonot/ezvonot_cases
Pipeline ID: ezvonot/ezvonot_crontab
Pipeline ID: ezvonot/ezvonot_decisions
Pipeline ID: ezvonot/ezvonot_decisions_join
Pipeline ID: ezvonot/ezvonot_deposits
Pipeline ID: ezvonot/ezvonot_purposes
Pipeline ID: facilities/all/facilities_all
Pipeline ID: facilities/education/facilities_education
Pipeline ID: facilities/labor/facilities_labor
Pipeline ID: facilities/welfare/facilities_welfare
Pipeline ID: government_decisions/government_decisions
Pipeline ID: government_decisions/scraper
Pipeline ID: knesset/ethics_committee_decisions
Pipeline ID: knesset/knesset_committee_decisions
Pipeline ID: lamas/municipal-data
Pipeline ID: lamas/municipality-to-city
Pipeline ID: people/association_founders/association-founders
רשימת כל המינויים והחברות מאתר כלכליסט
Pipeline ID: people/company_appointments/media/calcalist/calcalist
Pipeline ID: people/company_appointments/state_owned/parse_pdf_resources
columns: date, url
Pipeline ID: people/company_appointments/state_owned/reports
Pipeline ID: people/company_appointments/state_owned/state-owned-nominations
Pipeline ID: people/municipal_appointments/municipal_representitives
Pipeline ID: people/political_donations/political_donations
Pipeline ID: people/procurement/procurement-individuals
Pipeline ID: procurement/calls_for_bids/gov-publications
Pipeline ID: procurement/calls_for_bids/gov-publications-ministry-of-immigration
Pipeline ID: procurement/calls_for_bids/gov-publications2
Pipeline ID: procurement/municipal/fetcher
אנו סוכמים מידע מתוך הדוחות הרבעוניים לתוך טבלה המכילה את המידע העדכני ביותר.
Pipeline ID: procurement/spending/latest-contract-spending
Pipeline ID: procurement/spending/quarterly-contract-spending-report-uris
כל משרדי הממשלה ויחידות הסמך (שמחוברות למערכת מרכבה) צריכים לפרסם אחת לרבעון (תוך 45 יום מסוף הרבעון) דו״ח התקשרויות לפי פורמט מוגדר היטב הממומש כדו״ח מובנה במערכת המרכבה. כל המשרדים מדווחים עצמאית והדו״חות עולים לאתר היחידה לחופש המידע, משם אנו אוספים אותם. התוצר של השרשר הזה הוא פירוט כל הדיווחים של כל המשרדים.
Pipeline ID: procurement/spending/quarterly-contract-spending-reports
Pipeline ID: procurement/spending/quarterly-contract-spending-reports-data
Pipeline ID: procurement/spending/spending-by-entity
Pipeline ID: procurement/tenders/all
Pipeline ID: reports/all
Pipeline ID: reports/ngos/ngo-activity-report
Pipeline ID: reports/ngos/ngo-district-report
Pipeline ID: simpledb/process_simple_dbs
Pipeline ID: support_programs/all-support-programs
Pipeline ID: supports/by-payment-year
Pipeline ID: supports/criteria/support-criteria-doj
Pipeline ID: supports/supports-2004
Pipeline ID: supports/supports-2005
Pipeline ID: supports/supports-2006
Pipeline ID: supports/supports-2007
Pipeline ID: supports/supports-2008
Pipeline ID: supports/supports-2009
Pipeline ID: supports/supports-2010
Pipeline ID: supports/supports-2011
Pipeline ID: supports/supports-2012
Pipeline ID: supports/supports-2013
Pipeline ID: supports/supports-2014
Pipeline ID: supports/supports-2015
Pipeline ID: supports/supports-2016
Pipeline ID: supports/supports-2017
Pipeline ID: supports/supports-2018
Pipeline ID: supports/supports-2019
Pipeline ID: supports/supports-2020
Pipeline ID: supports/supports-2021
Pipeline ID: supports/supports-2022
Pipeline ID: supports/supports-by-entity
Pipeline ID: supports/with-entities
Pipeline ID: supports/with-keys
Pipeline ID: units/social_services/social_services