From 7cf66cafc4ac58806196a901c84be6b059ac0269 Mon Sep 17 00:00:00 2001 From: symroe Date: Fri, 13 Dec 2024 08:05:47 +0000 Subject: [PATCH] deploy: 4275d48baa6ca6adf19ca9bbede1ff52f720fc5d --- api/failing.json | 2 +- index.html | 2862 ++++++------ logbooks/ABC/index.html | 100 +- logbooks/ABD/index.html | 102 +- logbooks/ABE/index.html | 102 +- logbooks/ADU/index.html | 102 +- logbooks/AGB/index.html | 332 +- logbooks/AGY/index.html | 102 +- logbooks/AMB/index.html | 366 +- logbooks/AND/index.html | 100 +- logbooks/ANN/index.html | 104 +- logbooks/ANS/index.html | 250 +- logbooks/ARU/index.html | 106 +- logbooks/ASF/index.html | 104 +- logbooks/ASH/index.html | 102 +- logbooks/BAB/index.html | 106 +- logbooks/BAI/index.html | 102 +- logbooks/BAN/index.html | 106 +- logbooks/BAS/index.html | 106 +- logbooks/BBD/index.html | 106 +- logbooks/BDF/index.html | 104 +- logbooks/BDG/index.html | 200 +- logbooks/BEN/index.html | 106 +- logbooks/BEX/index.html | 102 +- logbooks/BFS/index.html | 106 +- logbooks/BGE/index.html | 106 +- logbooks/BGW/index.html | 102 +- logbooks/BIR/index.html | 9194 +++++++++++++++++++-------------------- logbooks/BLA/index.html | 102 +- logbooks/BNE/index.html | 106 +- logbooks/BNH/index.html | 106 +- logbooks/BNS/index.html | 106 +- logbooks/BOL/index.html | 198 +- logbooks/BOS/index.html | 102 +- logbooks/BOT/index.html | 102 +- logbooks/BPC/index.html | 106 +- logbooks/BPL/index.html | 102 +- logbooks/BRA/index.html | 494 +-- logbooks/BRC/index.html | 102 +- logbooks/BRD/index.html | 106 +- logbooks/BRE/index.html | 104 +- logbooks/BRM/index.html | 102 +- logbooks/BRT/index.html | 200 +- logbooks/BRW/index.html | 102 +- logbooks/BRX/index.html | 102 +- logbooks/BRY/index.html | 106 +- logbooks/BST/index.html | 106 +- logbooks/BUN/index.html | 102 +- logbooks/BUR/index.html | 106 +- logbooks/CAB/index.html | 102 +- logbooks/CAM/index.html | 594 +-- logbooks/CAN/index.html | 114 +- logbooks/CAS/index.html | 206 +- logbooks/CAT/index.html | 102 +- logbooks/CAY/index.html | 106 +- logbooks/CGN/index.html | 102 +- logbooks/CHA/index.html | 106 +- logbooks/CHE/index.html | 106 +- logbooks/CHI/index.html | 102 +- logbooks/CHO/index.html | 102 +- logbooks/CHR/index.html | 104 +- logbooks/CHS/index.html | 102 +- logbooks/CHT/index.html | 102 +- logbooks/CHW/index.html | 106 +- logbooks/CLD/index.html | 106 +- logbooks/CLK/index.html | 206 +- logbooks/CMD/index.html | 106 +- logbooks/CMN/index.html | 106 +- logbooks/COL/index.html | 504 +-- logbooks/CON/index.html | 106 +- logbooks/COT/index.html | 102 +- logbooks/COV/index.html | 106 +- logbooks/CRF/index.html | 106 +- logbooks/CRW/index.html | 102 +- logbooks/CRY/index.html | 106 +- logbooks/CWY/index.html | 106 +- logbooks/DAC/index.html | 105 +- logbooks/DAL/index.html | 106 +- logbooks/DAR/index.html | 102 +- logbooks/DBY/index.html | 106 +- logbooks/DEB/index.html | 102 +- logbooks/DEN/index.html | 104 +- logbooks/DER/index.html | 514 +-- logbooks/DEV/index.html | 106 +- logbooks/DGY/index.html | 102 +- logbooks/DNC/index.html | 106 +- logbooks/DND/index.html | 102 +- logbooks/DOV/index.html | 102 +- logbooks/DRS/index.html | 102 +- logbooks/DST/index.html | 106 +- logbooks/DUD/index.html | 530 +-- logbooks/DUR/index.html | 110 +- logbooks/EAL/index.html | 106 +- logbooks/EAS/index.html | 102 +- logbooks/EAT/index.html | 102 +- logbooks/EAY/index.html | 118 +- logbooks/EDE/index.html | 106 +- logbooks/EDH/index.html | 106 +- logbooks/EDU/index.html | 138 +- logbooks/EHA/index.html | 102 +- logbooks/EHE/index.html | 106 +- logbooks/ELI/index.html | 106 +- logbooks/ELM/index.html | 104 +- logbooks/ELN/index.html | 230 +- logbooks/ELS/index.html | 180 +- logbooks/ENF/index.html | 106 +- logbooks/EPP/index.html | 106 +- logbooks/EPS/index.html | 102 +- logbooks/ERE/index.html | 104 +- logbooks/ERW/index.html | 204 +- logbooks/ESS/index.html | 3458 +++++++-------- logbooks/ESX/index.html | 106 +- logbooks/EXE/index.html | 102 +- logbooks/FAL/index.html | 332 +- logbooks/FAR/index.html | 102 +- logbooks/FEN/index.html | 102 +- logbooks/FIF/index.html | 3334 +++++++------- logbooks/FLN/index.html | 101 +- logbooks/FOE/index.html | 102 +- logbooks/FYL/index.html | 506 +-- logbooks/GAT/index.html | 106 +- logbooks/GED/index.html | 102 +- logbooks/GLG/index.html | 138 +- logbooks/GLO/index.html | 102 +- logbooks/GLS/index.html | 200 +- logbooks/GOS/index.html | 102 +- logbooks/GRA/index.html | 102 +- logbooks/GRE/index.html | 130 +- logbooks/GRT/index.html | 104 +- logbooks/GRY/index.html | 414 +- logbooks/GWN/index.html | 106 +- logbooks/HAA/index.html | 102 +- logbooks/HAL/index.html | 106 +- logbooks/HAM/index.html | 106 +- logbooks/HAO/index.html | 372 +- logbooks/HAR/index.html | 159 +- logbooks/HAS/index.html | 102 +- logbooks/HAT/index.html | 102 +- logbooks/HAV/index.html | 106 +- logbooks/HCK/index.html | 106 +- logbooks/HEF/index.html | 106 +- logbooks/HER/index.html | 102 +- logbooks/HIG/index.html | 108 +- logbooks/HIL/index.html | 106 +- logbooks/HIN/index.html | 102 +- logbooks/HLD/index.html | 432 +- logbooks/HMF/index.html | 106 +- logbooks/HNS/index.html | 106 +- logbooks/HOR/index.html | 104 +- logbooks/HPL/index.html | 278 +- logbooks/HRT/index.html | 106 +- logbooks/HRW/index.html | 106 +- logbooks/HRY/index.html | 106 +- logbooks/HUN/index.html | 106 +- logbooks/HYN/index.html | 102 +- logbooks/IOS/index.html | 102 +- logbooks/IPS/index.html | 104 +- logbooks/ISL/index.html | 106 +- logbooks/IVC/index.html | 188 +- logbooks/KEC/index.html | 106 +- logbooks/KEN/index.html | 106 +- logbooks/KHL/index.html | 204 +- logbooks/KIN/index.html | 106 +- logbooks/KIR/index.html | 106 +- logbooks/KTT/index.html | 104 +- logbooks/KWL/index.html | 102 +- logbooks/LAC/index.html | 106 +- logbooks/LAN/index.html | 106 +- logbooks/LBH/index.html | 106 +- logbooks/LCE/index.html | 106 +- logbooks/LDS/index.html | 108 +- logbooks/LEC/index.html | 106 +- logbooks/LEE/index.html | 102 +- logbooks/LEW/index.html | 106 +- logbooks/LIC/index.html | 102 +- logbooks/LIF/index.html | 104 +- logbooks/LIN/index.html | 106 +- logbooks/LIV/index.html | 106 +- logbooks/LUT/index.html | 486 +-- logbooks/MAI/index.html | 104 +- logbooks/MAL/index.html | 102 +- logbooks/MAN/index.html | 108 +- logbooks/MAV/index.html | 102 +- logbooks/MDB/index.html | 104 +- logbooks/MDE/index.html | 102 +- logbooks/MDW/index.html | 101 +- logbooks/MEL/index.html | 102 +- logbooks/MIK/index.html | 106 +- logbooks/MLN/index.html | 244 +- logbooks/MOL/index.html | 336 +- logbooks/MON/index.html | 102 +- logbooks/MRT/index.html | 106 +- logbooks/MRY/index.html | 310 +- logbooks/MSS/index.html | 104 +- logbooks/MSU/index.html | 106 +- logbooks/MTY/index.html | 102 +- logbooks/MUL/index.html | 420 +- logbooks/NAY/index.html | 366 +- logbooks/NBL/index.html | 106 +- logbooks/NDE/index.html | 102 +- logbooks/NEA/index.html | 102 +- logbooks/NEC/index.html | 102 +- logbooks/NED/index.html | 106 +- logbooks/NEL/index.html | 354 +- logbooks/NET/index.html | 106 +- logbooks/NEW/index.html | 104 +- logbooks/NFK/index.html | 3844 ++++++++-------- logbooks/NGM/index.html | 106 +- logbooks/NHE/index.html | 106 +- logbooks/NKE/index.html | 102 +- logbooks/NLK/index.html | 3570 +++++++-------- logbooks/NLN/index.html | 102 +- logbooks/NNO/index.html | 102 +- logbooks/NOW/index.html | 412 +- logbooks/NSM/index.html | 106 +- logbooks/NTL/index.html | 123 +- logbooks/NTY/index.html | 106 +- logbooks/NUN/index.html | 328 +- logbooks/NWL/index.html | 102 +- logbooks/NWM/index.html | 106 +- logbooks/NWP/index.html | 106 +- logbooks/OAD/index.html | 102 +- logbooks/OLD/index.html | 106 +- logbooks/ORK/index.html | 322 +- logbooks/OXF/index.html | 106 +- logbooks/OXO/index.html | 104 +- logbooks/PEM/index.html | 106 +- logbooks/PEN/index.html | 242 +- logbooks/PKN/index.html | 422 +- logbooks/PLY/index.html | 106 +- logbooks/POR/index.html | 102 +- logbooks/POW/index.html | 106 +- logbooks/PRE/index.html | 104 +- logbooks/PTE/index.html | 106 +- logbooks/RCC/index.html | 106 +- logbooks/RCH/index.html | 106 +- logbooks/RCT/index.html | 106 +- logbooks/RDB/index.html | 106 +- logbooks/RDG/index.html | 104 +- logbooks/RED/index.html | 102 +- logbooks/REI/index.html | 102 +- logbooks/RFW/index.html | 446 +- logbooks/RIB/index.html | 102 +- logbooks/RIC/index.html | 106 +- logbooks/ROC/index.html | 414 +- logbooks/ROS/index.html | 264 +- logbooks/ROT/index.html | 106 +- logbooks/RUG/index.html | 350 +- logbooks/RUH/index.html | 102 +- logbooks/RUN/index.html | 102 +- logbooks/RUS/index.html | 102 +- logbooks/RUT/index.html | 102 +- logbooks/SAL/index.html | 106 +- logbooks/SAW/index.html | 106 +- logbooks/SAY/index.html | 242 +- logbooks/SCA/index.html | 102 +- logbooks/SCB/index.html | 102 +- logbooks/SDE/index.html | 390 +- logbooks/SEV/index.html | 106 +- logbooks/SFK/index.html | 200 +- logbooks/SFT/index.html | 106 +- logbooks/SGC/index.html | 106 +- logbooks/SHA/index.html | 106 +- logbooks/SHF/index.html | 106 +- logbooks/SHN/index.html | 104 +- logbooks/SHO/index.html | 102 +- logbooks/SHR/index.html | 106 +- logbooks/SKE/index.html | 106 +- logbooks/SKP/index.html | 106 +- logbooks/SLF/index.html | 106 +- logbooks/SLG/index.html | 102 +- logbooks/SLK/index.html | 618 +-- logbooks/SND/index.html | 3506 +++++++-------- logbooks/SNO/index.html | 102 +- logbooks/SOL/index.html | 106 +- logbooks/SOS/index.html | 106 +- logbooks/SOX/index.html | 102 +- logbooks/SPE/index.html | 102 +- logbooks/SRI/index.html | 106 +- logbooks/SRY/index.html | 106 +- logbooks/SST/index.html | 436 +- logbooks/STE/index.html | 102 +- logbooks/STF/index.html | 108 +- logbooks/STG/index.html | 240 +- logbooks/STH/index.html | 99 +- logbooks/STN/index.html | 102 +- logbooks/STO/index.html | 106 +- logbooks/STR/index.html | 102 +- logbooks/STS/index.html | 106 +- logbooks/STV/index.html | 102 +- logbooks/SUR/index.html | 102 +- logbooks/SWA/index.html | 106 +- logbooks/SWD/index.html | 206 +- logbooks/SWK/index.html | 106 +- logbooks/SWL/index.html | 102 +- logbooks/TAM/index.html | 106 +- logbooks/TAN/index.html | 102 +- logbooks/TAW/index.html | 102 +- logbooks/TEI/index.html | 104 +- logbooks/TEN/index.html | 104 +- logbooks/TES/index.html | 102 +- logbooks/TEW/index.html | 102 +- logbooks/TFW/index.html | 106 +- logbooks/THA/index.html | 106 +- logbooks/THE/index.html | 112 +- logbooks/THR/index.html | 104 +- logbooks/TOB/index.html | 102 +- logbooks/TOF/index.html | 102 +- logbooks/TON/index.html | 102 +- logbooks/TOR/index.html | 102 +- logbooks/TRF/index.html | 101 +- logbooks/TUN/index.html | 102 +- logbooks/TWH/index.html | 102 +- logbooks/UTT/index.html | 102 +- logbooks/VAL/index.html | 102 +- logbooks/VGL/index.html | 222 +- logbooks/WAE/index.html | 106 +- logbooks/WAR/index.html | 106 +- logbooks/WAT/index.html | 102 +- logbooks/WAW/index.html | 450 +- logbooks/WBK/index.html | 102 +- logbooks/WDE/index.html | 106 +- logbooks/WDU/index.html | 274 +- logbooks/WEA/index.html | 102 +- logbooks/WEW/index.html | 104 +- logbooks/WFT/index.html | 106 +- logbooks/WGN/index.html | 106 +- logbooks/WIL/index.html | 108 +- logbooks/WIN/index.html | 206 +- logbooks/WKF/index.html | 106 +- logbooks/WLA/index.html | 102 +- logbooks/WLI/index.html | 102 +- logbooks/WLL/index.html | 584 +-- logbooks/WLN/index.html | 317 +- logbooks/WLV/index.html | 106 +- logbooks/WND/index.html | 106 +- logbooks/WNM/index.html | 102 +- logbooks/WOC/index.html | 102 +- logbooks/WOI/index.html | 102 +- logbooks/WOK/index.html | 106 +- logbooks/WOR/index.html | 106 +- logbooks/WOT/index.html | 102 +- logbooks/WOX/index.html | 104 +- logbooks/WRL/index.html | 106 +- logbooks/WRT/index.html | 238 +- logbooks/WRX/index.html | 106 +- logbooks/WSK/index.html | 106 +- logbooks/WSM/index.html | 106 +- logbooks/WSX/index.html | 106 +- logbooks/WYC/index.html | 102 +- logbooks/WYE/index.html | 102 +- logbooks/WYR/index.html | 106 +- logbooks/YOR/index.html | 103 +- logbooks/ZET/index.html | 228 +- 354 files changed, 39391 insertions(+), 39480 deletions(-) diff --git a/api/failing.json b/api/failing.json index 165c185a03..8e702020f9 100644 --- a/api/failing.json +++ b/api/failing.json @@ -1 +1 @@ -[{"council_id":"BDG","missing":false,"latest_run":{"status_code":1,"log_text":"[08:35:54] Fetching Scraper for: BDG handlers.py:23\n Begin attempting to scrape: BDG handlers.py:27\n Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n[08:35:55] ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n ...data deleted. base.py:264\n Scraping from base.py:49\n https://modgov.lbbd.gov.uk/internet/mgWebService.asmx/GetC \n ouncillorsByWard \n[08:36:11] [Errno 110] Connection timed out handlers.py:36\n Finished attempting to scrape: BDG base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 122, in _connect\n stream = self._network_backend.connect_tcp(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 205, in connect_tcp\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectTimeout: [Errno 110] Connection timed out\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectTimeout: [Errno 110] Connection timed out\n","start":"2024-12-10 08:35:54.067786","end":"2024-12-10 08:36:11.752839","duration":17}},{"council_id":"BOL","missing":false,"latest_run":{"status_code":1,"log_text":"[09:30:52] Fetching Scraper for: BOL handlers.py:23\n[09:30:53] Begin attempting to scrape: BOL handlers.py:27\n[09:30:54] Deleting existing data... base.py:257\n[09:30:55] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[09:30:56] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.democracy.bolton.gov.uk/cmis5/People.aspx \n [Errno -2] Name or service not known handlers.py:36\n Finished attempting to scrape: BOL base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 122, in _connect\n stream = self._network_backend.connect_tcp(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 205, in connect_tcp\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectError: [Errno -2] Name or service not known\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 272, in get_councillors\n req = self.get(\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectError: [Errno -2] Name or service not known\n","start":"2024-12-10 09:30:52.977579","end":"2024-12-10 09:30:56.463623","duration":3}},{"council_id":"BRT","missing":false,"latest_run":{"status_code":1,"log_text":"[09:12:34] Fetching Scraper for: BRT handlers.py:23\n Begin attempting to scrape: BRT handlers.py:27\n[09:12:35] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[09:12:36] ...data deleted. base.py:264\n Scraping from base.py:49\n http://democracy.broxtowe.gov.uk/mgWebService.asmx/GetCoun \n cillorsByWard \n[09:12:52] [Errno 110] Connection timed out handlers.py:36\n Finished attempting to scrape: BRT base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 122, in _connect\n stream = self._network_backend.connect_tcp(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 205, in connect_tcp\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectTimeout: [Errno 110] Connection timed out\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectTimeout: [Errno 110] Connection timed out\n","start":"2024-12-10 09:12:34.726905","end":"2024-12-10 09:12:52.487138","duration":17}},{"council_id":"CAN","missing":false,"latest_run":{"status_code":1,"log_text":"[10:01:07] Fetching Scraper for: CAN handlers.py:23\n Begin attempting to scrape: CAN handlers.py:27\n[10:01:08] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[10:01:09] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.cannockchasedc.gov.uk/council/about-council/yo \n ur-councillors \n[10:01:11] list index out of range handlers.py:36\n Finished attempting to scrape: CAN base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 161, in get_councillors\n container = self.get_list_container()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 158, in get_list_container\n return selected[0]\nIndexError: list index out of range\n","start":"2024-12-10 10:01:07.666772","end":"2024-12-10 10:01:11.933570","duration":4}},{"council_id":"CAS","missing":false,"latest_run":{"status_code":1,"log_text":"[10:06:38] Fetching Scraper for: CAS handlers.py:23\n Begin attempting to scrape: CAS handlers.py:27\n Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n[10:06:39] ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n ...data deleted. base.py:264\n Scraping from https://www.castlepoint.gov.uk/councillors base.py:49\n[10:06:41] Scraping from base.py:49\n https://www.castlepoint.gov.ukhttps://castlepoint.cmis.uk. \n com/castlepoint/Committees/CurrentCommittees.aspx?a=1 \n [Errno -2] Name or service not known handlers.py:36\n Finished attempting to scrape: CAS base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 122, in _connect\n stream = self._network_backend.connect_tcp(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 205, in connect_tcp\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectError: [Errno -2] Name or service not known\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"scrapers/CAS-castle-point/councillors.py\", line 15, in get_single_councillor\n soup = self.get_page(url)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 141, in get_page\n page = self.get(url, extra_headers=self.extra_headers).text\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectError: [Errno -2] Name or service not known\n","start":"2024-12-10 10:06:38.026653","end":"2024-12-10 10:06:41.478880","duration":3}},{"council_id":"EAY","missing":false,"latest_run":{"status_code":1,"log_text":"[10:35:39] Fetching Scraper for: EAY handlers.py:23\n Begin attempting to scrape: EAY handlers.py:27\n Deleting existing data... base.py:257\n[10:35:40] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[10:35:41] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.east-ayrshire.gov.uk/CouncilAndGovernment/Abou \n t-the-Council/Councillors-and-Provost/YourCouncillor.aspx \n[10:35:42] Scraping from base.py:49\n https://www.east-ayrshire.gov.uk/CouncilAndGovernment/Abou \n t-the-Council/Councillors-and-Provost/YourCouncillor.aspx? \n 9 \n[10:35:43] 'NoneType' object has no attribute 'find_parent' handlers.py:36\n Finished attempting to scrape: EAY base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"scrapers/EAY-east-ayrshire/councillors.py\", line 22, in get_single_councillor\n .find_parent(\"div\")\nAttributeError: 'NoneType' object has no attribute 'find_parent'\n","start":"2024-12-10 10:35:39.425303","end":"2024-12-10 10:35:43.340709","duration":3}},{"council_id":"EDU","missing":false,"latest_run":{"status_code":1,"log_text":"[10:13:02] Fetching Scraper for: EDU handlers.py:23\n Begin attempting to scrape: EDU handlers.py:27\n Deleting existing data... base.py:257\n[10:13:03] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[10:13:04] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.eastdunbarton.gov.uk/residents/council-democra \n cy/committees-and-councillors/councillors-2017 \n Client error '404 Not Found' for url handlers.py:36\n 'https://www.eastdunbarton.gov.uk/residents/council-de \n mocracy/committees-and-councillors/councillors-2017' \n For more information check: \n https://developer.mozilla.org/en-US/docs/Web/HTTP/Stat \n us/404 \n Finished attempting to scrape: EDU base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 161, in get_councillors\n container = self.get_list_container()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 152, in get_list_container\n self.base_url_soup = self.get_page(self.base_url)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 141, in get_page\n page = self.get(url, extra_headers=self.extra_headers).text\n File \"/var/task/lgsf/scrapers/base.py\", line 56, in get\n response.raise_for_status()\n File \"/opt/python/httpx/_models.py\", line 761, in raise_for_status\n raise HTTPStatusError(message, request=request, response=self)\nhttpx.HTTPStatusError: Client error '404 Not Found' for url 'https://www.eastdunbarton.gov.uk/residents/council-democracy/committees-and-councillors/councillors-2017'\nFor more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/404\n","start":"2024-12-10 10:13:02.264604","end":"2024-12-10 10:13:04.549612","duration":2}},{"council_id":"ELS","missing":false,"latest_run":{"status_code":1,"log_text":"[08:21:33] Fetching Scraper for: ELS handlers.py:23\n Begin attempting to scrape: ELS handlers.py:27\n[08:21:34] Deleting existing data... base.py:257\n[08:21:35] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[08:21:36] ...data deleted. base.py:264\n Scraping from base.py:49\n https://cne-siar.gov.uk/home/your-council/council-members/ \n Client error '404 Not Found' for url handlers.py:36\n 'https://www.cne-siar.gov.uk/home/your-council/council \n -members/' \n For more information check: \n https://developer.mozilla.org/en-US/docs/Web/HTTP/Stat \n us/404 \n[08:21:37] Finished attempting to scrape: ELS base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 161, in get_councillors\n container = self.get_list_container()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 152, in get_list_container\n self.base_url_soup = self.get_page(self.base_url)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 141, in get_page\n page = self.get(url, extra_headers=self.extra_headers).text\n File \"/var/task/lgsf/scrapers/base.py\", line 56, in get\n response.raise_for_status()\n File \"/opt/python/httpx/_models.py\", line 761, in raise_for_status\n raise HTTPStatusError(message, request=request, response=self)\nhttpx.HTTPStatusError: Client error '404 Not Found' for url 'https://www.cne-siar.gov.uk/home/your-council/council-members/'\nFor more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/404\n","start":"2024-12-10 08:21:33.702569","end":"2024-12-10 08:21:37.263122","duration":3}},{"council_id":"FYL","missing":false,"latest_run":{"status_code":1,"log_text":"[08:22:07] Fetching Scraper for: FYL handlers.py:23\n Begin attempting to scrape: FYL handlers.py:27\n[08:22:08] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n Getting all files in Councillors/json... base.py:209\n[08:22:09] ...found 17 files in Councillors/json base.py:225\n Getting all files in Councillors/raw... base.py:209\n ...found 17 files in Councillors/raw base.py:225\n ...found 35 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 35 files base.py:236\n[08:22:10] ...data deleted. base.py:264\n Scraping from base.py:49\n https://fylde.cmis.uk.com/fylde/CouncillorsandMP.aspx \n[08:22:11] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/48/ScreenMode/Alphabetical/D \n efault.aspx \n[08:22:14] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/48/ScreenMode/Alphabetical/D \n efault.aspx \n[08:22:15] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/148/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:18] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/148/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:20] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/84/ScreenMode/Alphabetical/D \n efault.aspx \n[08:22:24] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/84/ScreenMode/Alphabetical/D \n efault.aspx \n[08:22:25] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/211/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:27] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/211/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:29] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/207/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:31] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/207/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:32] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/52/ScreenMode/Alphabetical/D \n efault.aspx \n[08:22:36] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/52/ScreenMode/Alphabetical/D \n efault.aspx \n[08:22:37] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/208/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:40] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/208/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:41] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/43/ScreenMode/Alphabetical/D \n efault.aspx \n[08:22:44] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/43/ScreenMode/Alphabetical/D \n efault.aspx \n[08:22:48] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/172/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:51] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/172/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:52] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/213/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:55] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/213/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:56] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/209/ScreenMode/Alphabetical/ \n Default.aspx \n[08:22:59] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/209/ScreenMode/Alphabetical/ \n Default.aspx \n[08:23:00] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/15/ScreenMode/Alphabetical/D \n efault.aspx \n[08:23:03] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/15/ScreenMode/Alphabetical/D \n efault.aspx \n[08:23:04] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/206/ScreenMode/Alphabetical/ \n Default.aspx \n[08:23:07] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/206/ScreenMode/Alphabetical/ \n Default.aspx \n[08:23:10] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/174/ScreenMode/Alphabetical/ \n Default.aspx \n[08:23:12] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/174/ScreenMode/Alphabetical/ \n Default.aspx \n[08:23:14] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/14/ScreenMode/Alphabetical/D \n efault.aspx \n[08:23:16] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/14/ScreenMode/Alphabetical/D \n efault.aspx \n[08:23:18] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/186/ScreenMode/Alphabetical/ \n Default.aspx \n[08:23:21] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/186/ScreenMode/Alphabetical/ \n Default.aspx \n[08:23:22] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/176/ScreenMode/Alphabetical/ \n Default.aspx \n[08:23:25] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/176/ScreenMode/Alphabetical/ \n Default.aspx \n[08:23:27] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/222/ScreenMode/Alphabetical/ \n Default.aspx \n[08:23:29] 'NoneType' object has no attribute 'parent' handlers.py:36\n Committing batch 1 consisting of 34 files base.py:297\n[08:23:30] Finished attempting to scrape: FYL base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 300, in get_single_councillor\n party = self.get_party_name(list_page_html)\n File \"scrapers/FYL-fylde/councillors.py\", line 17, in get_party_name\n .parent.get_text(strip=True)\nAttributeError: 'NoneType' object has no attribute 'parent'\n","start":"2024-12-10 08:22:07.776916","end":"2024-12-10 08:23:30.523275","duration":82}},{"council_id":"GLG","missing":false,"latest_run":{"status_code":1,"log_text":"[10:12:45] Fetching Scraper for: GLG handlers.py:23\n Begin attempting to scrape: GLG handlers.py:27\n Deleting existing data... base.py:257\n[10:12:46] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[10:12:47] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.glasgow.gov.uk/councillorsandcommittees/allMem \n bers.asp?sort=0&page=0&rec=100 \n Client error '404 Not Found' for url handlers.py:36\n 'https://www.glasgow.gov.uk/councillorsandcommittees/a \n llMembers.asp?sort=0&page=0&rec=100' \n For more information check: \n https://developer.mozilla.org/en-US/docs/Web/HTTP/Stat \n us/404 \n Finished attempting to scrape: GLG base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 161, in get_councillors\n container = self.get_list_container()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 152, in get_list_container\n self.base_url_soup = self.get_page(self.base_url)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 141, in get_page\n page = self.get(url, extra_headers=self.extra_headers).text\n File \"/var/task/lgsf/scrapers/base.py\", line 56, in get\n response.raise_for_status()\n File \"/opt/python/httpx/_models.py\", line 761, in raise_for_status\n raise HTTPStatusError(message, request=request, response=self)\nhttpx.HTTPStatusError: Client error '404 Not Found' for url 'https://www.glasgow.gov.uk/councillorsandcommittees/allMembers.asp?sort=0&page=0&rec=100'\nFor more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/404\n","start":"2024-12-10 10:12:45.348241","end":"2024-12-10 10:12:47.462082","duration":2}},{"council_id":"GLS","missing":false,"latest_run":{"status_code":1,"log_text":"[09:15:13] Fetching Scraper for: GLS handlers.py:23\n Begin attempting to scrape: GLS handlers.py:27\n[09:15:14] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[09:15:15] ...data deleted. base.py:264\n Scraping from base.py:49\n http://glostext.gloucestershire.gov.uk//mgWebService.asmx/ \n GetCouncillorsByWard \n[09:15:31] [Errno 110] Connection timed out handlers.py:36\n Finished attempting to scrape: GLS base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 122, in _connect\n stream = self._network_backend.connect_tcp(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 205, in connect_tcp\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectTimeout: [Errno 110] Connection timed out\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectTimeout: [Errno 110] Connection timed out\n","start":"2024-12-10 09:15:13.781015","end":"2024-12-10 09:15:31.545267","duration":17}},{"council_id":"GRE","missing":false,"latest_run":{"status_code":1,"log_text":"[08:28:19] Fetching Scraper for: GRE handlers.py:23\n Begin attempting to scrape: GRE handlers.py:27\n Deleting existing data... base.py:257\n[08:28:20] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[08:28:21] ...data deleted. base.py:264\n Scraping from base.py:49\n https://committees.royalgreenwich.gov.uk/Councillors/tabid \n /63/ScreenMode/Alphabetical/Default.aspx \n[08:28:26] Client error '404 Not Found' for url handlers.py:36\n 'https://committees.royalgreenwich.gov.uk/mgError.aspx \n ' \n For more information check: \n https://developer.mozilla.org/en-US/docs/Web/HTTP/Stat \n us/404 \n Finished attempting to scrape: GRE base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 272, in get_councillors\n req = self.get(\n File \"/var/task/lgsf/scrapers/base.py\", line 56, in get\n response.raise_for_status()\n File \"/opt/python/httpx/_models.py\", line 761, in raise_for_status\n raise HTTPStatusError(message, request=request, response=self)\nhttpx.HTTPStatusError: Client error '404 Not Found' for url 'https://committees.royalgreenwich.gov.uk/mgError.aspx'\nFor more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/404\n","start":"2024-12-10 08:28:19.498630","end":"2024-12-10 08:28:26.236491","duration":6}},{"council_id":"KHL","missing":false,"latest_run":{"status_code":1,"log_text":"[08:48:42] Fetching Scraper for: KHL handlers.py:23\n Begin attempting to scrape: KHL handlers.py:27\n Deleting existing data... base.py:257\n[08:48:43] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[08:48:44] ...data deleted. base.py:264\n Scraping from base.py:49\n https://cmis.hullcc.gov.uk/cmis/CouncillorsandSeniorOffice \n rs/CouncillorsandSeniorOfficers.aspx \n [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify handlers.py:36\n failed: unable to get local issuer certificate \n (_ssl.c:1007) \n Finished attempting to scrape: KHL base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 154, in _connect\n stream = stream.start_tls(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 152, in start_tls\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 272, in get_councillors\n req = self.get(\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n","start":"2024-12-10 08:48:42.395772","end":"2024-12-10 08:48:44.507341","duration":2}},{"council_id":"SHE","missing":false,"latest_run":{"status_code":null,"log_text":"[11:28:20] Fetching Scraper for: SHE handlers.py:22\n Begin attempting to scrape: SHE handlers.py:25\n Deleting existing data... base.py:234\n Getting all files in SHE... base.py:186\n[11:28:21] Getting all files in SHE/json... base.py:186\n ...found 30 files in SHE/json base.py:202\n Getting all files in SHE/raw... base.py:186\n ...found 30 files in SHE/raw base.py:202\n ...found 61 files in SHE base.py:202\n Deleting batch no. 1 consisting of 61 files base.py:211\n[11:28:32] An error occurred (ThrottlingException) when calling handlers.py:34\n the CreateCommit operation (reached max retries: 4): \n Rate exceeded \n Finished attempting to scrape: SHE base.py:319\n","errors":"An error occurred (ThrottlingException) when calling the CreateCommit operation (reached max retries: 4): Rate exceeded","start":"2022-04-04 11:28:20.509898","end":"2022-04-04 11:28:32.871624","duration":12}},{"council_id":"STG","missing":false,"latest_run":{"status_code":1,"log_text":"[09:30:17] Fetching Scraper for: STG handlers.py:23\n Begin attempting to scrape: STG handlers.py:27\n Deleting existing data... base.py:257\n[09:30:18] Getting all files in Councillors... base.py:209\n Getting all files in Councillors/json... base.py:209\n ...found 19 files in Councillors/json base.py:225\n Getting all files in Councillors/raw... base.py:209\n ...found 19 files in Councillors/raw base.py:225\n ...found 39 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 39 files base.py:236\n[09:30:19] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/politic \n ians-and-elections/councillors/ \n[09:30:21] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/elaine-watterson/ \n[09:30:24] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/gene-maxwell/ \n[09:30:25] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/martin-earl/ \n[09:30:26] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/gerry-mcgarvey/ \n[09:30:27] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/paul-henke/ \n[09:30:29] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/rosemary-fraser/ \n[09:30:30] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/alasdair-tollemache/ \n[09:30:31] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/david-wilson/ \n[09:30:33] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/robin-kleinman/ \n[09:30:34] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/thomas-heald/ \n[09:30:35] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/danny-gibson/ \n[09:30:36] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/jim-thomson/ \n[09:30:38] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/rachel-nunn/ \n[09:30:39] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/susan-mcgill/ \n[09:30:40] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/jen-preston/ \n[09:30:41] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/neil-benny/ \n[09:30:43] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/scott-farmer/ \n[09:30:44] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/bryan-flannagan/ \n[09:30:45] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/gerry-mclaughlan/ \n[09:30:46] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/vacant/ \n[09:30:48] 'NoneType' object is not subscriptable handlers.py:36\n Committing batch 1 consisting of 38 files base.py:297\n[09:30:49] Finished attempting to scrape: STG base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"scrapers/STG-stirling/councillors.py\", line 42, in get_single_councillor\n councillor.email = soup.select_one(\"a[href^=mailto]\")[\nTypeError: 'NoneType' object is not subscriptable\n","start":"2024-12-10 09:30:17.320960","end":"2024-12-10 09:30:49.446170","duration":32}},{"council_id":"SWD","missing":false,"latest_run":{"status_code":1,"log_text":"[09:51:17] Fetching Scraper for: SWD handlers.py:23\n Begin attempting to scrape: SWD handlers.py:27\n Deleting existing data... base.py:257\n[09:51:18] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[09:51:19] ...data deleted. base.py:264\n Scraping from base.py:49\n https://democracy.winchester.gov.uk/mgWebService.asmx/GetC \n ouncillorsByWard \n [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify handlers.py:36\n failed: unable to get local issuer certificate \n (_ssl.c:1007) \n Finished attempting to scrape: SWD base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 154, in _connect\n stream = stream.start_tls(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 152, in start_tls\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n","start":"2024-12-10 09:51:17.435927","end":"2024-12-10 09:51:19.505916","duration":2}},{"council_id":"THE","missing":false,"latest_run":{"status_code":1,"log_text":"[08:29:34] Fetching Scraper for: THE handlers.py:23\n Begin attempting to scrape: THE handlers.py:27\n[08:29:35] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[08:29:36] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.threerivers.gov.uk/listing/councillors \n[08:29:38] 'NoneType' object has no attribute 'findNext' handlers.py:36\n[08:29:39] Finished attempting to scrape: THE base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 161, in get_councillors\n container = self.get_list_container()\n File \"scrapers/THE-three-rivers/councillors.py\", line 15, in get_list_container\n return soup.find(\"h3\", text=\"District Councillor\").findNext(\"ul\")\nAttributeError: 'NoneType' object has no attribute 'findNext'\n","start":"2024-12-10 08:29:34.632986","end":"2024-12-10 08:29:39.124264","duration":4}},{"council_id":"TRF","missing":false,"latest_run":{"status_code":1,"log_text":"[09:37:53] Fetching Scraper for: TRF handlers.py:23\n Begin attempting to scrape: TRF handlers.py:27\n Deleting existing data... base.py:257\n[09:37:54] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n ...data deleted. base.py:264\n Scraping from base.py:49\n https://democratic.trafford.gov.uk/mgWebService.asmx/GetCo \n uncillorsByWard \n[09:38:25] The read operation timed out handlers.py:36\n Finished attempting to scrape: TRF base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 101, in handle_request\n return self._connection.handle_request(request)\n File \"/opt/python/httpcore/_sync/http11.py\", line 143, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/http11.py\", line 113, in handle_request\n ) = self._receive_response_headers(**kwargs)\n File \"/opt/python/httpcore/_sync/http11.py\", line 186, in _receive_response_headers\n event = self._receive_event(timeout=timeout)\n File \"/opt/python/httpcore/_sync/http11.py\", line 224, in _receive_event\n data = self._network_stream.read(\n File \"/opt/python/httpcore/_backends/sync.py\", line 124, in read\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ReadTimeout: The read operation timed out\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ReadTimeout: The read operation timed out\n","start":"2024-12-10 09:37:53.135949","end":"2024-12-10 09:38:25.489303","duration":32}},{"council_id":"VGL","missing":false,"latest_run":{"status_code":1,"log_text":"[08:45:54] Fetching Scraper for: VGL handlers.py:23\n Begin attempting to scrape: VGL handlers.py:27\n Deleting existing data... base.py:257\n[08:45:55] Getting all files in Councillors... base.py:209\n Getting all files in Councillors/json... base.py:209\n ...found 16 files in Councillors/json base.py:225\n Getting all files in Councillors/raw... base.py:209\n ...found 16 files in Councillors/raw base.py:225\n ...found 33 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 33 files base.py:236\n[08:45:56] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Councillors.aspx \n[08:45:58] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Asbrey-Anne.aspx \n[08:46:00] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Aviet-Julie.aspx \n[08:46:01] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Ball-Gareth.aspx \n[08:46:03] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Birch-Rhiannon.aspx \n[08:46:05] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Brooks-Bronwen.aspx \n[08:46:06] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Bruce-Gillian.aspx \n[08:46:08] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Buckley-Ian.aspx \n[08:46:11] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Burnett-Lis.aspx \n[08:46:12] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Campbell-Samantha.aspx \n[08:46:14] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Carroll-George.aspx \n[08:46:16] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Cave-Christine.aspx \n[08:46:17] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Champion-Charles.aspx \n[08:46:19] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Charles-Janice.aspx \n[08:46:21] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Collins-Amelia.aspx \n[08:46:22] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Cowpe-Marianne.aspx \n[08:46:24] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Drake-Pamela.aspx \n[08:46:26] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Driscoll-Vincent.aspx \n[08:46:28] 'NoneType' object has no attribute 'get_text' handlers.py:36\n Committing batch 1 consisting of 32 files base.py:297\n[08:46:29] Finished attempting to scrape: VGL base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"scrapers/VGL-the-vale-of-glamorgan/councillors.py\", line 36, in get_single_councillor\n email_text = soup.select_one(\"#S4_EmailPlaceholder\").get_text(\nAttributeError: 'NoneType' object has no attribute 'get_text'\n","start":"2024-12-10 08:45:54.249358","end":"2024-12-10 08:46:29.544848","duration":35}},{"council_id":"WIN","missing":false,"latest_run":{"status_code":1,"log_text":"[09:57:00] Fetching Scraper for: WIN handlers.py:23\n Begin attempting to scrape: WIN handlers.py:27\n[09:57:01] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[09:57:02] ...data deleted. base.py:264\n Scraping from base.py:49\n https://democracy.winchester.gov.uk/mgWebService.asmx/GetC \n ouncillorsByWard \n [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify handlers.py:36\n failed: unable to get local issuer certificate \n (_ssl.c:1007) \n Finished attempting to scrape: WIN base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 154, in _connect\n stream = stream.start_tls(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 152, in start_tls\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n","start":"2024-12-10 09:57:00.766384","end":"2024-12-10 09:57:02.968276","duration":2}},{"council_id":"WLN","missing":false,"latest_run":{"status_code":1,"log_text":"[10:19:51] Fetching Scraper for: WLN handlers.py:23\n Begin attempting to scrape: WLN handlers.py:27\n[10:19:52] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n[10:19:53] Getting all files in Councillors/json... base.py:209\n ...found 25 files in Councillors/json base.py:225\n Getting all files in Councillors/raw... base.py:209\n ...found 25 files in Councillors/raw base.py:225\n ...found 51 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 51 files base.py:236\n[10:19:54] ...data deleted. base.py:264\n Scraping from https://westlothian.gov.uk/councillors base.py:49\n[10:19:57] Scraping from base.py:49\n https://westlothian.gov.uk/article/33888/Linlithgow \n[10:19:58] Scraping from base.py:49\n https://westlothian.gov.uk/article/33889/Broxburn-Uphall-a \n nd-Winchburgh \n[10:19:59] Scraping from base.py:49\n https://westlothian.gov.uk/article/33890/Livingston-North \n Scraping from base.py:49\n https://westlothian.gov.uk/article/33891/Livingston-South \n[10:20:00] Scraping from base.py:49\n https://westlothian.gov.uk/article/33893/East-Livingston-a \n nd-East-Calder \n[10:20:01] Scraping from base.py:49\n https://westlothian.gov.uk/article/33892/Fauldhouse-and-th \n e-Breich-Valley \n[10:20:02] Scraping from base.py:49\n https://westlothian.gov.uk/article/33894/Whitburn-and-Blac \n kburn \n[10:20:03] Scraping from base.py:49\n https://westlothian.gov.uk/article/33895/Bathgate \n[10:20:04] Scraping from base.py:49\n https://westlothian.gov.uk/article/33896/Armadale-and-Blac \n kridge \n[10:20:05] Scraping from base.py:49\n https://westlothian.gov.uk/article/33897/Councillor-Tom-Co \n nn \n[10:20:06] Scraping from base.py:49\n https://westlothian.gov.uk/article/33898/Councillor-Paulin \n e-Orr \n[10:20:07] Scraping from base.py:49\n https://westlothian.gov.uk/article/33899/Councillor-Sally- \n Pattle \n[10:20:09] Scraping from base.py:49\n https://westlothian.gov.uk/article/34305/Councillor-Tony-B \n oyle \n[10:20:10] Scraping from base.py:49\n https://westlothian.gov.uk/article/34302/Councillor-Diane- \n Calder \n[10:20:11] Scraping from base.py:49\n https://westlothian.gov.uk/article/34303/Councillor-Janet- \n Campbell \n[10:20:12] Scraping from base.py:49\n https://westlothian.gov.uk/article/34304/Councillor-Angela \n -Doran-Timson \n[10:20:13] Scraping from base.py:49\n https://westlothian.gov.uk/article/34307/Councillor-Alison \n -Adamson \n[10:20:15] Scraping from base.py:49\n https://westlothian.gov.uk/article/34306/Councillor-Robert \n -De-Bold \n[10:20:16] Scraping from base.py:49\n https://westlothian.gov.uk/article/34308/Councillor-Anne-M \n cMillan \n[10:20:17] Scraping from base.py:49\n https://westlothian.gov.uk/article/34309/Councillor-Andrew \n -Miller \n[10:20:18] Scraping from base.py:49\n https://westlothian.gov.uk/article/34311/Councillor-Lawren \n ce-Fitzpatrick \n[10:20:19] Scraping from base.py:49\n https://westlothian.gov.uk/article/34310/Councillor-Peter- \n Heggie \n[10:20:20] Scraping from base.py:49\n https://westlothian.gov.uk/article/34312/Councillor-Maria- \n MacAulay \n[10:20:21] Scraping from base.py:49\n https://westlothian.gov.uk/article/34313/Councillor-Moira- \n McKee-Shemilt \n[10:20:22] Scraping from base.py:49\n https://westlothian.gov.uk/article/34317/Councillor-Damian \n -Doran-Timson \n[10:20:23] Scraping from base.py:49\n https://westlothian.gov.uk/article/34315/Councillor-Carl-J \n ohn \n[10:20:24] Scraping from base.py:49\n https://westlothian.gov.uk/article/34316/Councillor-Danny- \n Logue \n[10:20:26] Scraping from base.py:49\n https://westlothian.gov.uk/article/34314/Councillor-Veroni \n ca-Smith \n[10:20:27] Scraping from base.py:49\n https://westlothian.gov.uk/article/34318/Councillor-Paulin \n e-Clark \n[10:20:28] Scraping from base.py:49\n https://westlothian.gov.uk/article/34319/Councillor-Craig- \n Meek \n[10:20:29] Scraping from base.py:49\n https://westlothian.gov.uk/article/34320/Councillor-Cathy- \n Muldoon \n[10:20:30] Scraping from base.py:49\n https://westlothian.gov.uk/article/34321/Councillor-Jim-Di \n ckson \n[10:20:31] Scraping from base.py:49\n https://westlothian.gov.uk/article/34322/Councillor-Mary-D \n ickson \n[10:20:32] Scraping from base.py:49\n https://westlothian.gov.uk/article/34323/Councillor-George \n -Paul \n[10:20:33] Scraping from base.py:49\n https://westlothian.gov.uk/article/34324/Councillor-David- \n Russell \n[10:20:34] list index out of range handlers.py:36\n Committing batch 1 consisting of 50 files base.py:297\n[10:20:36] Finished attempting to scrape: WLN base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"scrapers/WLN-west-lothian/councillors.py\", line 37, in get_single_councillor\n councillor.email = soup.select(\"a[href^=mailto]\")[0].get_text(\nIndexError: list index out of range\n","start":"2024-12-10 10:19:51.782538","end":"2024-12-10 10:20:36.185971","duration":44}}] +[{"council_id":"BDG","missing":false,"latest_run":{"status_code":1,"log_text":"[08:49:19] Fetching Scraper for: BDG handlers.py:23\n Begin attempting to scrape: BDG handlers.py:27\n Deleting existing data... base.py:257\n[08:49:20] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[08:49:21] ...data deleted. base.py:264\n Scraping from base.py:49\n https://modgov.lbbd.gov.uk/internet/mgWebService.asmx/GetC \n ouncillorsByWard \n[08:49:36] [Errno 110] Connection timed out handlers.py:36\n Finished attempting to scrape: BDG base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 122, in _connect\n stream = self._network_backend.connect_tcp(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 205, in connect_tcp\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectTimeout: [Errno 110] Connection timed out\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectTimeout: [Errno 110] Connection timed out\n","start":"2024-12-11 08:49:19.245593","end":"2024-12-11 08:49:36.971201","duration":17}},{"council_id":"BOL","missing":false,"latest_run":{"status_code":1,"log_text":"[09:00:29] Fetching Scraper for: BOL handlers.py:23\n Begin attempting to scrape: BOL handlers.py:27\n[09:00:30] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[09:00:31] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.democracy.bolton.gov.uk/cmis5/People.aspx \n [Errno -2] Name or service not known handlers.py:36\n Finished attempting to scrape: BOL base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 122, in _connect\n stream = self._network_backend.connect_tcp(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 205, in connect_tcp\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectError: [Errno -2] Name or service not known\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 272, in get_councillors\n req = self.get(\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectError: [Errno -2] Name or service not known\n","start":"2024-12-11 09:00:29.589295","end":"2024-12-11 09:00:31.707552","duration":2}},{"council_id":"BRT","missing":false,"latest_run":{"status_code":1,"log_text":"[10:42:51] Fetching Scraper for: BRT handlers.py:23\n Begin attempting to scrape: BRT handlers.py:27\n[10:42:52] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[10:42:53] ...data deleted. base.py:264\n Scraping from base.py:49\n http://democracy.broxtowe.gov.uk/mgWebService.asmx/GetCoun \n cillorsByWard \n[10:43:08] [Errno 110] Connection timed out handlers.py:36\n[10:43:09] Finished attempting to scrape: BRT base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 122, in _connect\n stream = self._network_backend.connect_tcp(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 205, in connect_tcp\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectTimeout: [Errno 110] Connection timed out\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectTimeout: [Errno 110] Connection timed out\n","start":"2024-12-11 10:42:51.820742","end":"2024-12-11 10:43:09.212030","duration":17}},{"council_id":"CAN","missing":false,"latest_run":{"status_code":1,"log_text":"[10:10:19] Fetching Scraper for: CAN handlers.py:23\n Begin attempting to scrape: CAN handlers.py:27\n Deleting existing data... base.py:257\n[10:10:20] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[10:10:21] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.cannockchasedc.gov.uk/council/about-council/yo \n ur-councillors \n[10:10:23] list index out of range handlers.py:36\n Finished attempting to scrape: CAN base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 161, in get_councillors\n container = self.get_list_container()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 158, in get_list_container\n return selected[0]\nIndexError: list index out of range\n","start":"2024-12-11 10:10:19.172081","end":"2024-12-11 10:10:23.583463","duration":4}},{"council_id":"CAS","missing":false,"latest_run":{"status_code":1,"log_text":"[10:39:07] Fetching Scraper for: CAS handlers.py:23\n Begin attempting to scrape: CAS handlers.py:27\n Deleting existing data... base.py:257\n[10:39:08] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n ...data deleted. base.py:264\n Scraping from https://www.castlepoint.gov.uk/councillors base.py:49\n[10:39:10] Scraping from base.py:49\n https://www.castlepoint.gov.ukhttps://castlepoint.cmis.uk. \n com/castlepoint/Committees/CurrentCommittees.aspx?a=1 \n [Errno -2] Name or service not known handlers.py:36\n Finished attempting to scrape: CAS base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 122, in _connect\n stream = self._network_backend.connect_tcp(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 205, in connect_tcp\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectError: [Errno -2] Name or service not known\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"scrapers/CAS-castle-point/councillors.py\", line 15, in get_single_councillor\n soup = self.get_page(url)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 141, in get_page\n page = self.get(url, extra_headers=self.extra_headers).text\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectError: [Errno -2] Name or service not known\n","start":"2024-12-11 10:39:07.303844","end":"2024-12-11 10:39:10.492887","duration":3}},{"council_id":"EAY","missing":false,"latest_run":{"status_code":1,"log_text":"[09:19:39] Fetching Scraper for: EAY handlers.py:23\n Begin attempting to scrape: EAY handlers.py:27\n Deleting existing data... base.py:257\n[09:19:40] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.east-ayrshire.gov.uk/CouncilAndGovernment/Abou \n t-the-Council/Councillors-and-Provost/YourCouncillor.aspx \n[09:19:42] Scraping from base.py:49\n https://www.east-ayrshire.gov.uk/CouncilAndGovernment/Abou \n t-the-Council/Councillors-and-Provost/YourCouncillor.aspx? \n 9 \n 'NoneType' object has no attribute 'find_parent' handlers.py:36\n[09:19:43] Finished attempting to scrape: EAY base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"scrapers/EAY-east-ayrshire/councillors.py\", line 22, in get_single_councillor\n .find_parent(\"div\")\nAttributeError: 'NoneType' object has no attribute 'find_parent'\n","start":"2024-12-11 09:19:39.068099","end":"2024-12-11 09:19:43.066380","duration":3}},{"council_id":"EDU","missing":false,"latest_run":{"status_code":1,"log_text":"[09:18:40] Fetching Scraper for: EDU handlers.py:23\n Begin attempting to scrape: EDU handlers.py:27\n[09:18:41] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n[09:18:42] ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.eastdunbarton.gov.uk/residents/council-democra \n cy/committees-and-councillors/councillors-2017 \n[09:18:43] Client error '404 Not Found' for url handlers.py:36\n 'https://www.eastdunbarton.gov.uk/residents/council-de \n mocracy/committees-and-councillors/councillors-2017' \n For more information check: \n https://developer.mozilla.org/en-US/docs/Web/HTTP/Stat \n us/404 \n Finished attempting to scrape: EDU base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 161, in get_councillors\n container = self.get_list_container()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 152, in get_list_container\n self.base_url_soup = self.get_page(self.base_url)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 141, in get_page\n page = self.get(url, extra_headers=self.extra_headers).text\n File \"/var/task/lgsf/scrapers/base.py\", line 56, in get\n response.raise_for_status()\n File \"/opt/python/httpx/_models.py\", line 761, in raise_for_status\n raise HTTPStatusError(message, request=request, response=self)\nhttpx.HTTPStatusError: Client error '404 Not Found' for url 'https://www.eastdunbarton.gov.uk/residents/council-democracy/committees-and-councillors/councillors-2017'\nFor more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/404\n","start":"2024-12-11 09:18:40.897609","end":"2024-12-11 09:18:43.247957","duration":2}},{"council_id":"ELS","missing":false,"latest_run":{"status_code":1,"log_text":"[08:19:07] Fetching Scraper for: ELS handlers.py:23\n Begin attempting to scrape: ELS handlers.py:27\n Deleting existing data... base.py:257\n[08:19:08] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[08:19:09] ...data deleted. base.py:264\n Scraping from base.py:49\n https://cne-siar.gov.uk/home/your-council/council-members/ \n Client error '404 Not Found' for url handlers.py:36\n 'https://www.cne-siar.gov.uk/home/your-council/council \n -members/' \n For more information check: \n https://developer.mozilla.org/en-US/docs/Web/HTTP/Stat \n us/404 \n[08:19:10] Finished attempting to scrape: ELS base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 161, in get_councillors\n container = self.get_list_container()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 152, in get_list_container\n self.base_url_soup = self.get_page(self.base_url)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 141, in get_page\n page = self.get(url, extra_headers=self.extra_headers).text\n File \"/var/task/lgsf/scrapers/base.py\", line 56, in get\n response.raise_for_status()\n File \"/opt/python/httpx/_models.py\", line 761, in raise_for_status\n raise HTTPStatusError(message, request=request, response=self)\nhttpx.HTTPStatusError: Client error '404 Not Found' for url 'https://www.cne-siar.gov.uk/home/your-council/council-members/'\nFor more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/404\n","start":"2024-12-11 08:19:07.535315","end":"2024-12-11 08:19:10.175918","duration":2}},{"council_id":"FYL","missing":false,"latest_run":{"status_code":1,"log_text":"[10:21:28] Fetching Scraper for: FYL handlers.py:23\n Begin attempting to scrape: FYL handlers.py:27\n Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n Getting all files in Councillors/json... base.py:209\n[10:21:29] ...found 17 files in Councillors/json base.py:225\n Getting all files in Councillors/raw... base.py:209\n ...found 17 files in Councillors/raw base.py:225\n ...found 35 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 35 files base.py:236\n[10:21:30] ...data deleted. base.py:264\n Scraping from base.py:49\n https://fylde.cmis.uk.com/fylde/CouncillorsandMP.aspx \n Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/48/ScreenMode/Alphabetical/D \n efault.aspx \n[10:21:34] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/48/ScreenMode/Alphabetical/D \n efault.aspx \n[10:21:36] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/148/ScreenMode/Alphabetical/ \n Default.aspx \n[10:21:38] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/148/ScreenMode/Alphabetical/ \n Default.aspx \n[10:21:40] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/84/ScreenMode/Alphabetical/D \n efault.aspx \n[10:21:42] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/84/ScreenMode/Alphabetical/D \n efault.aspx \n[10:21:43] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/211/ScreenMode/Alphabetical/ \n Default.aspx \n[10:21:46] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/211/ScreenMode/Alphabetical/ \n Default.aspx \n[10:21:47] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/207/ScreenMode/Alphabetical/ \n Default.aspx \n[10:21:49] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/207/ScreenMode/Alphabetical/ \n Default.aspx \n[10:21:50] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/52/ScreenMode/Alphabetical/D \n efault.aspx \n[10:21:55] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/52/ScreenMode/Alphabetical/D \n efault.aspx \n[10:21:56] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/208/ScreenMode/Alphabetical/ \n Default.aspx \n[10:21:59] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/208/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:00] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/43/ScreenMode/Alphabetical/D \n efault.aspx \n[10:22:02] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/43/ScreenMode/Alphabetical/D \n efault.aspx \n[10:22:04] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/172/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:06] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/172/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:08] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/213/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:10] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/213/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:11] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/209/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:13] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/209/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:17] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/15/ScreenMode/Alphabetical/D \n efault.aspx \n[10:22:20] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/15/ScreenMode/Alphabetical/D \n efault.aspx \n[10:22:22] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/206/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:24] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/206/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:25] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/174/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:28] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/174/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:29] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/14/ScreenMode/Alphabetical/D \n efault.aspx \n[10:22:31] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/14/ScreenMode/Alphabetical/D \n efault.aspx \n[10:22:33] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/186/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:35] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/186/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:37] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/176/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:41] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/176/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:43] Scraping from base.py:49\n http://fylde.cmis.uk.com/fylde/CouncillorsandMP/tabid/63/c \n tl/ViewCMIS_Person/mid/383/id/222/ScreenMode/Alphabetical/ \n Default.aspx \n[10:22:44] 'NoneType' object has no attribute 'parent' handlers.py:36\n Committing batch 1 consisting of 34 files base.py:297\n[10:22:46] Finished attempting to scrape: FYL base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 300, in get_single_councillor\n party = self.get_party_name(list_page_html)\n File \"scrapers/FYL-fylde/councillors.py\", line 17, in get_party_name\n .parent.get_text(strip=True)\nAttributeError: 'NoneType' object has no attribute 'parent'\n","start":"2024-12-11 10:21:28.007261","end":"2024-12-11 10:22:46.113689","duration":78}},{"council_id":"GLG","missing":false,"latest_run":{"status_code":1,"log_text":"[10:17:54] Fetching Scraper for: GLG handlers.py:23\n Begin attempting to scrape: GLG handlers.py:27\n[10:17:56] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[10:17:57] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.glasgow.gov.uk/councillorsandcommittees/allMem \n bers.asp?sort=0&page=0&rec=100 \n Client error '404 Not Found' for url handlers.py:36\n 'https://www.glasgow.gov.uk/councillorsandcommittees/a \n llMembers.asp?sort=0&page=0&rec=100' \n For more information check: \n https://developer.mozilla.org/en-US/docs/Web/HTTP/Stat \n us/404 \n Finished attempting to scrape: GLG base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 161, in get_councillors\n container = self.get_list_container()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 152, in get_list_container\n self.base_url_soup = self.get_page(self.base_url)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 141, in get_page\n page = self.get(url, extra_headers=self.extra_headers).text\n File \"/var/task/lgsf/scrapers/base.py\", line 56, in get\n response.raise_for_status()\n File \"/opt/python/httpx/_models.py\", line 761, in raise_for_status\n raise HTTPStatusError(message, request=request, response=self)\nhttpx.HTTPStatusError: Client error '404 Not Found' for url 'https://www.glasgow.gov.uk/councillorsandcommittees/allMembers.asp?sort=0&page=0&rec=100'\nFor more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/404\n","start":"2024-12-11 10:17:54.898205","end":"2024-12-11 10:17:57.874945","duration":2}},{"council_id":"GLS","missing":false,"latest_run":{"status_code":1,"log_text":"[09:39:41] Fetching Scraper for: GLS handlers.py:23\n Begin attempting to scrape: GLS handlers.py:27\n Deleting existing data... base.py:257\n[09:39:42] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[09:39:43] ...data deleted. base.py:264\n Scraping from base.py:49\n http://glostext.gloucestershire.gov.uk//mgWebService.asmx/ \n GetCouncillorsByWard \n[09:39:58] [Errno 110] Connection timed out handlers.py:36\n[09:39:59] Finished attempting to scrape: GLS base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 122, in _connect\n stream = self._network_backend.connect_tcp(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 205, in connect_tcp\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectTimeout: [Errno 110] Connection timed out\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectTimeout: [Errno 110] Connection timed out\n","start":"2024-12-11 09:39:41.263308","end":"2024-12-11 09:39:59.254893","duration":17}},{"council_id":"GRE","missing":false,"latest_run":{"status_code":1,"log_text":"[10:10:07] Fetching Scraper for: GRE handlers.py:23\n Begin attempting to scrape: GRE handlers.py:27\n Deleting existing data... base.py:257\n[10:10:08] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[10:10:09] ...data deleted. base.py:264\n Scraping from base.py:49\n https://committees.royalgreenwich.gov.uk/Councillors/tabid \n /63/ScreenMode/Alphabetical/Default.aspx \n[10:10:15] Client error '404 Not Found' for url handlers.py:36\n 'https://committees.royalgreenwich.gov.uk/mgError.aspx \n ' \n For more information check: \n https://developer.mozilla.org/en-US/docs/Web/HTTP/Stat \n us/404 \n Finished attempting to scrape: GRE base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 272, in get_councillors\n req = self.get(\n File \"/var/task/lgsf/scrapers/base.py\", line 56, in get\n response.raise_for_status()\n File \"/opt/python/httpx/_models.py\", line 761, in raise_for_status\n raise HTTPStatusError(message, request=request, response=self)\nhttpx.HTTPStatusError: Client error '404 Not Found' for url 'https://committees.royalgreenwich.gov.uk/mgError.aspx'\nFor more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/404\n","start":"2024-12-11 10:10:07.343251","end":"2024-12-11 10:10:15.915833","duration":8}},{"council_id":"HAR","missing":false,"latest_run":{"status_code":1,"log_text":"[09:55:33] Fetching Scraper for: HAR handlers.py:23\n Begin attempting to scrape: HAR handlers.py:27\n Deleting existing data... base.py:257\n[09:55:34] Getting all files in Councillors... base.py:209\n Getting all files in Councillors/json... base.py:209\n ...found 33 files in Councillors/json base.py:225\n Getting all files in Councillors/raw... base.py:209\n ...found 33 files in Councillors/raw base.py:225\n ...found 67 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 67 files base.py:236\n[09:55:35] ...data deleted. base.py:264\n Scraping from base.py:49\n http://moderngov.harlow.gov.uk/mgWebService.asmx/GetCounci \n llorsByWard \n[09:56:05] The read operation timed out handlers.py:36\n[09:56:06] Finished attempting to scrape: HAR base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 101, in handle_request\n return self._connection.handle_request(request)\n File \"/opt/python/httpcore/_sync/http11.py\", line 143, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/http11.py\", line 113, in handle_request\n ) = self._receive_response_headers(**kwargs)\n File \"/opt/python/httpcore/_sync/http11.py\", line 186, in _receive_response_headers\n event = self._receive_event(timeout=timeout)\n File \"/opt/python/httpcore/_sync/http11.py\", line 224, in _receive_event\n data = self._network_stream.read(\n File \"/opt/python/httpcore/_backends/sync.py\", line 124, in read\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ReadTimeout: The read operation timed out\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ReadTimeout: The read operation timed out\n","start":"2024-12-11 09:55:33.122866","end":"2024-12-11 09:56:06.020811","duration":32}},{"council_id":"KHL","missing":false,"latest_run":{"status_code":1,"log_text":"[08:48:02] Fetching Scraper for: KHL handlers.py:23\n Begin attempting to scrape: KHL handlers.py:27\n Deleting existing data... base.py:257\n[08:48:03] Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[08:48:04] ...data deleted. base.py:264\n Scraping from base.py:49\n https://cmis.hullcc.gov.uk/cmis/CouncillorsandSeniorOffice \n rs/CouncillorsandSeniorOfficers.aspx \n [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify handlers.py:36\n failed: unable to get local issuer certificate \n (_ssl.c:1007) \n Finished attempting to scrape: KHL base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 154, in _connect\n stream = stream.start_tls(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 152, in start_tls\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 272, in get_councillors\n req = self.get(\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n","start":"2024-12-11 08:48:02.254356","end":"2024-12-11 08:48:04.370885","duration":2}},{"council_id":"SHE","missing":false,"latest_run":{"status_code":null,"log_text":"[11:28:20] Fetching Scraper for: SHE handlers.py:22\n Begin attempting to scrape: SHE handlers.py:25\n Deleting existing data... base.py:234\n Getting all files in SHE... base.py:186\n[11:28:21] Getting all files in SHE/json... base.py:186\n ...found 30 files in SHE/json base.py:202\n Getting all files in SHE/raw... base.py:186\n ...found 30 files in SHE/raw base.py:202\n ...found 61 files in SHE base.py:202\n Deleting batch no. 1 consisting of 61 files base.py:211\n[11:28:32] An error occurred (ThrottlingException) when calling handlers.py:34\n the CreateCommit operation (reached max retries: 4): \n Rate exceeded \n Finished attempting to scrape: SHE base.py:319\n","errors":"An error occurred (ThrottlingException) when calling the CreateCommit operation (reached max retries: 4): Rate exceeded","start":"2022-04-04 11:28:20.509898","end":"2022-04-04 11:28:32.871624","duration":12}},{"council_id":"STG","missing":false,"latest_run":{"status_code":1,"log_text":"[10:18:00] Fetching Scraper for: STG handlers.py:23\n[10:18:01] Begin attempting to scrape: STG handlers.py:27\n Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n Getting all files in Councillors/json... base.py:209\n[10:18:02] ...found 19 files in Councillors/json base.py:225\n Getting all files in Councillors/raw... base.py:209\n ...found 19 files in Councillors/raw base.py:225\n ...found 39 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 39 files base.py:236\n[10:18:03] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/politic \n ians-and-elections/councillors/ \n[10:18:04] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/elaine-watterson/ \n[10:18:05] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/gene-maxwell/ \n[10:18:06] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/martin-earl/ \n[10:18:07] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/gerry-mcgarvey/ \n[10:18:08] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/paul-henke/ \n[10:18:09] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/rosemary-fraser/ \n[10:18:10] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/alasdair-tollemache/ \n[10:18:11] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/david-wilson/ \n[10:18:12] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/robin-kleinman/ \n[10:18:13] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/thomas-heald/ \n[10:18:14] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/danny-gibson/ \n[10:18:15] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/jim-thomson/ \n[10:18:16] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/rachel-nunn/ \n[10:18:17] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/susan-mcgill/ \n[10:18:18] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/jen-preston/ \n[10:18:20] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/neil-benny/ \n[10:18:21] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/scott-farmer/ \n[10:18:22] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/bryan-flannagan/ \n[10:18:24] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/gerry-mclaughlan/ \n[10:18:25] Scraping from base.py:49\n https://www.stirling.gov.uk/council-and-committees/council \n lors/your-councillors/willie-ferguson/ \n[10:18:26] 'NoneType' object is not subscriptable handlers.py:36\n Committing batch 1 consisting of 38 files base.py:297\n[10:18:27] Finished attempting to scrape: STG base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"scrapers/STG-stirling/councillors.py\", line 42, in get_single_councillor\n councillor.email = soup.select_one(\"a[href^=mailto]\")[\nTypeError: 'NoneType' object is not subscriptable\n","start":"2024-12-11 10:18:00.973730","end":"2024-12-11 10:18:27.500092","duration":26}},{"council_id":"SWD","missing":false,"latest_run":{"status_code":1,"log_text":"[09:19:01] Fetching Scraper for: SWD handlers.py:23\n Begin attempting to scrape: SWD handlers.py:27\n[09:19:02] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[09:19:03] ...data deleted. base.py:264\n Scraping from base.py:49\n https://democracy.winchester.gov.uk/mgWebService.asmx/GetC \n ouncillorsByWard \n [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify handlers.py:36\n failed: unable to get local issuer certificate \n (_ssl.c:1007) \n Finished attempting to scrape: SWD base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 154, in _connect\n stream = stream.start_tls(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 152, in start_tls\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n","start":"2024-12-11 09:19:01.563923","end":"2024-12-11 09:19:03.925436","duration":2}},{"council_id":"THE","missing":false,"latest_run":{"status_code":1,"log_text":"[10:39:00] Fetching Scraper for: THE handlers.py:23\n Begin attempting to scrape: THE handlers.py:27\n[10:39:01] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n[10:39:02] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.threerivers.gov.uk/listing/councillors \n[10:39:03] 'NoneType' object has no attribute 'findNext' handlers.py:36\n[10:39:04] Finished attempting to scrape: THE base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 55, in run\n for councillor_html in self.get_councillors():\n File \"/var/task/lgsf/councillors/scrapers.py\", line 161, in get_councillors\n container = self.get_list_container()\n File \"scrapers/THE-three-rivers/councillors.py\", line 15, in get_list_container\n return soup.find(\"h3\", text=\"District Councillor\").findNext(\"ul\")\nAttributeError: 'NoneType' object has no attribute 'findNext'\n","start":"2024-12-11 10:39:00.768188","end":"2024-12-11 10:39:04.167385","duration":3}},{"council_id":"VGL","missing":false,"latest_run":{"status_code":1,"log_text":"[09:21:51] Fetching Scraper for: VGL handlers.py:23\n Begin attempting to scrape: VGL handlers.py:27\n[09:21:52] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n[09:21:53] Getting all files in Councillors/json... base.py:209\n ...found 16 files in Councillors/json base.py:225\n Getting all files in Councillors/raw... base.py:209\n ...found 16 files in Councillors/raw base.py:225\n ...found 33 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 33 files base.py:236\n[09:21:54] ...data deleted. base.py:264\n Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Councillors.aspx \n[09:21:56] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Asbrey-Anne.aspx \n[09:21:58] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Aviet-Julie.aspx \n[09:22:00] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Ball-Gareth.aspx \n[09:22:01] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Birch-Rhiannon.aspx \n[09:22:04] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Brooks-Bronwen.aspx \n[09:22:06] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Bruce-Gillian.aspx \n[09:22:07] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Buckley-Ian.aspx \n[09:22:09] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Burnett-Lis.aspx \n[09:22:11] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Campbell-Samantha.aspx \n[09:22:13] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Carroll-George.aspx \n[09:22:14] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Cave-Christine.aspx \n[09:22:16] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Champion-Charles.aspx \n[09:22:17] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Charles-Janice.aspx \n[09:22:19] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Collins-Amelia.aspx \n[09:22:21] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Cowpe-Marianne.aspx \n[09:22:22] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Drake-Pamela.aspx \n[09:22:24] Scraping from base.py:49\n https://www.valeofglamorgan.gov.uk/en/our_council/Council- \n Structure/councillors/Driscoll-Vincent.aspx \n[09:22:26] 'NoneType' object has no attribute 'get_text' handlers.py:36\n Committing batch 1 consisting of 32 files base.py:297\n[09:22:27] Finished attempting to scrape: VGL base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"scrapers/VGL-the-vale-of-glamorgan/councillors.py\", line 36, in get_single_councillor\n email_text = soup.select_one(\"#S4_EmailPlaceholder\").get_text(\nAttributeError: 'NoneType' object has no attribute 'get_text'\n","start":"2024-12-11 09:21:51.864050","end":"2024-12-11 09:22:27.610187","duration":35}},{"council_id":"WIN","missing":false,"latest_run":{"status_code":1,"log_text":"[09:01:40] Fetching Scraper for: WIN handlers.py:23\n Begin attempting to scrape: WIN handlers.py:27\n[09:01:41] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n[09:01:42] ...found 1 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 1 files base.py:236\n ...data deleted. base.py:264\n Scraping from base.py:49\n https://democracy.winchester.gov.uk/mgWebService.asmx/GetC \n ouncillorsByWard \n [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify handlers.py:36\n failed: unable to get local issuer certificate \n (_ssl.c:1007) \n[09:01:43] Finished attempting to scrape: WIN base.py:345\n","errors":"Traceback (most recent call last):\n File \"/opt/python/httpx/_transports/default.py\", line 69, in map_httpcore_exceptions\n yield\n File \"/opt/python/httpx/_transports/default.py\", line 233, in handle_request\n resp = self._pool.handle_request(req)\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 216, in handle_request\n raise exc from None\n File \"/opt/python/httpcore/_sync/connection_pool.py\", line 196, in handle_request\n response = connection.handle_request(\n File \"/opt/python/httpcore/_sync/connection.py\", line 99, in handle_request\n raise exc\n File \"/opt/python/httpcore/_sync/connection.py\", line 76, in handle_request\n stream = self._connect(request)\n File \"/opt/python/httpcore/_sync/connection.py\", line 154, in _connect\n stream = stream.start_tls(**kwargs)\n File \"/opt/python/httpcore/_backends/sync.py\", line 152, in start_tls\n with map_exceptions(exc_map):\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpcore/_exceptions.py\", line 14, in map_exceptions\n raise to_exc(exc) from exc\nhttpcore.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 197, in run\n wards = self.get_councillors()\n File \"/var/task/lgsf/councillors/scrapers.py\", line 216, in get_councillors\n req = self.get(self.format_councillor_api_url())\n File \"/var/task/lgsf/scrapers/base.py\", line 55, in get\n response = self.http_client.get(url, headers=headers, timeout=30)\n File \"/opt/python/httpx/_client.py\", line 1054, in get\n return self.request(\n File \"/opt/python/httpx/_client.py\", line 827, in request\n return self.send(request, auth=auth, follow_redirects=follow_redirects)\n File \"/opt/python/httpx/_client.py\", line 914, in send\n response = self._send_handling_auth(\n File \"/opt/python/httpx/_client.py\", line 942, in _send_handling_auth\n response = self._send_handling_redirects(\n File \"/opt/python/httpx/_client.py\", line 979, in _send_handling_redirects\n response = self._send_single_request(request)\n File \"/opt/python/httpx/_client.py\", line 1015, in _send_single_request\n response = transport.handle_request(request)\n File \"/opt/python/httpx/_transports/default.py\", line 232, in handle_request\n with map_httpcore_exceptions():\n File \"/var/lang/lib/python3.10/contextlib.py\", line 153, in __exit__\n self.gen.throw(typ, value, traceback)\n File \"/opt/python/httpx/_transports/default.py\", line 86, in map_httpcore_exceptions\n raise mapped_exc(message) from exc\nhttpx.ConnectError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1007)\n","start":"2024-12-11 09:01:40.898322","end":"2024-12-11 09:01:43.069755","duration":2}},{"council_id":"WLN","missing":false,"latest_run":{"status_code":1,"log_text":"[08:19:49] Fetching Scraper for: WLN handlers.py:23\n Begin attempting to scrape: WLN handlers.py:27\n[08:19:50] Deleting existing data... base.py:257\n Getting all files in Councillors... base.py:209\n Getting all files in Councillors/json... base.py:209\n[08:19:51] ...found 25 files in Councillors/json base.py:225\n Getting all files in Councillors/raw... base.py:209\n ...found 25 files in Councillors/raw base.py:225\n ...found 51 files in Councillors base.py:225\n Deleting batch no. 1 consisting of 51 files base.py:236\n[08:19:52] ...data deleted. base.py:264\n Scraping from https://westlothian.gov.uk/councillors base.py:49\n[08:19:55] Scraping from base.py:49\n https://westlothian.gov.uk/article/33888/Linlithgow \n Scraping from base.py:49\n https://westlothian.gov.uk/article/33889/Broxburn-Uphall-a \n nd-Winchburgh \n[08:19:56] Scraping from base.py:49\n https://westlothian.gov.uk/article/33890/Livingston-North \n[08:19:57] Scraping from base.py:49\n https://westlothian.gov.uk/article/33891/Livingston-South \n[08:19:58] Scraping from base.py:49\n https://westlothian.gov.uk/article/33893/East-Livingston-a \n nd-East-Calder \n[08:19:59] Scraping from base.py:49\n https://westlothian.gov.uk/article/33892/Fauldhouse-and-th \n e-Breich-Valley \n[08:20:00] Scraping from base.py:49\n https://westlothian.gov.uk/article/33894/Whitburn-and-Blac \n kburn \n Scraping from base.py:49\n https://westlothian.gov.uk/article/33895/Bathgate \n[08:20:01] Scraping from base.py:49\n https://westlothian.gov.uk/article/33896/Armadale-and-Blac \n kridge \n[08:20:02] Scraping from base.py:49\n https://westlothian.gov.uk/article/33897/Councillor-Tom-Co \n nn \n[08:20:03] Scraping from base.py:49\n https://westlothian.gov.uk/article/33898/Councillor-Paulin \n e-Orr \n[08:20:04] Scraping from base.py:49\n https://westlothian.gov.uk/article/33899/Councillor-Sally- \n Pattle \n[08:20:05] Scraping from base.py:49\n https://westlothian.gov.uk/article/34305/Councillor-Tony-B \n oyle \n[08:20:06] Scraping from base.py:49\n https://westlothian.gov.uk/article/34302/Councillor-Diane- \n Calder \n[08:20:07] Scraping from base.py:49\n https://westlothian.gov.uk/article/34303/Councillor-Janet- \n Campbell \n[08:20:08] Scraping from base.py:49\n https://westlothian.gov.uk/article/34304/Councillor-Angela \n -Doran-Timson \n[08:20:09] Scraping from base.py:49\n https://westlothian.gov.uk/article/34307/Councillor-Alison \n -Adamson \n[08:20:10] Scraping from base.py:49\n https://westlothian.gov.uk/article/34306/Councillor-Robert \n -De-Bold \n[08:20:11] Scraping from base.py:49\n https://westlothian.gov.uk/article/34308/Councillor-Anne-M \n cMillan \n[08:20:13] Scraping from base.py:49\n https://westlothian.gov.uk/article/34309/Councillor-Andrew \n -Miller \n[08:20:15] Scraping from base.py:49\n https://westlothian.gov.uk/article/34311/Councillor-Lawren \n ce-Fitzpatrick \n[08:20:16] Scraping from base.py:49\n https://westlothian.gov.uk/article/34310/Councillor-Peter- \n Heggie \n[08:20:17] Scraping from base.py:49\n https://westlothian.gov.uk/article/34312/Councillor-Maria- \n MacAulay \n[08:20:18] Scraping from base.py:49\n https://westlothian.gov.uk/article/34313/Councillor-Moira- \n McKee-Shemilt \n[08:20:19] Scraping from base.py:49\n https://westlothian.gov.uk/article/34317/Councillor-Damian \n -Doran-Timson \n[08:20:20] Scraping from base.py:49\n https://westlothian.gov.uk/article/34315/Councillor-Carl-J \n ohn \n[08:20:21] Scraping from base.py:49\n https://westlothian.gov.uk/article/34316/Councillor-Danny- \n Logue \n[08:20:22] Scraping from base.py:49\n https://westlothian.gov.uk/article/34314/Councillor-Veroni \n ca-Smith \n[08:20:23] Scraping from base.py:49\n https://westlothian.gov.uk/article/34318/Councillor-Paulin \n e-Clark \n[08:20:24] Scraping from base.py:49\n https://westlothian.gov.uk/article/34319/Councillor-Craig- \n Meek \n[08:20:25] Scraping from base.py:49\n https://westlothian.gov.uk/article/34320/Councillor-Cathy- \n Muldoon \n[08:20:26] Scraping from base.py:49\n https://westlothian.gov.uk/article/34321/Councillor-Jim-Di \n ckson \n[08:20:27] Scraping from base.py:49\n https://westlothian.gov.uk/article/34322/Councillor-Mary-D \n ickson \n[08:20:28] Scraping from base.py:49\n https://westlothian.gov.uk/article/34323/Councillor-George \n -Paul \n[08:20:29] Scraping from base.py:49\n https://westlothian.gov.uk/article/34324/Councillor-David- \n Russell \n[08:20:30] list index out of range handlers.py:36\n Committing batch 1 consisting of 50 files base.py:297\n[08:20:31] Finished attempting to scrape: WLN base.py:345\n","errors":"Traceback (most recent call last):\n File \"/var/task/lgsf/aws_lambda/handlers.py\", line 32, in scraper_worker_handler\n scraper.run(run_log)\n File \"/var/task/lgsf/councillors/scrapers.py\", line 57, in run\n councillor = self.get_single_councillor(councillor_html)\n File \"scrapers/WLN-west-lothian/councillors.py\", line 37, in get_single_councillor\n councillor.email = soup.select(\"a[href^=mailto]\")[0].get_text(\nIndexError: list index out of range\n","start":"2024-12-11 08:19:49.843195","end":"2024-12-11 08:20:31.653670","duration":41}}] diff --git a/index.html b/index.html index 3275a63918..7bd47cc2c1 100644 --- a/index.html +++ b/index.html @@ -49,10 +49,6 @@

All log runs

- -
-
-
@@ -129,6 +125,10 @@

All log runs

+ +
+
+
@@ -145,10 +145,6 @@

All log runs

- -
-
-
@@ -225,6 +221,10 @@

All log runs

+ +
+
+
@@ -241,10 +241,6 @@

All log runs

- -
-
-
@@ -321,6 +317,10 @@

All log runs

+ +
+
+
@@ -337,10 +337,6 @@

All log runs

- -
-
-
@@ -417,6 +413,10 @@

All log runs

+ +
+
+
@@ -433,10 +433,6 @@

All log runs

- -
-
-
@@ -513,6 +509,10 @@

All log runs

+ +
+
+
@@ -529,10 +529,6 @@

All log runs

- -
-
-
@@ -609,6 +605,10 @@

All log runs

+ +
+
+
@@ -625,10 +625,6 @@

All log runs

- -
-
-
@@ -705,6 +701,10 @@

All log runs

+ +
+
+
@@ -721,10 +721,6 @@

All log runs

- -
-
-
@@ -801,6 +797,10 @@

All log runs

+ +
+
+
@@ -817,10 +817,6 @@

All log runs

- -
-
-
@@ -897,6 +893,10 @@

All log runs

+ +
+
+
@@ -913,10 +913,6 @@

All log runs

- -
-
-
@@ -993,6 +989,10 @@

All log runs

+ +
+
+
@@ -1009,10 +1009,6 @@

All log runs

- -
-
-
@@ -1089,6 +1085,10 @@

All log runs

+ +
+
+
@@ -1105,10 +1105,6 @@

All log runs

- -
-
-
@@ -1185,6 +1181,10 @@

All log runs

+ +
+
+
@@ -1201,10 +1201,6 @@

All log runs

- -
-
-
@@ -1281,6 +1277,10 @@

All log runs

+ +
+
+
@@ -1297,10 +1297,6 @@

All log runs

- -
-
-
@@ -1377,6 +1373,10 @@

All log runs

+ +
+
+
@@ -1393,10 +1393,6 @@

All log runs

- -
-
-
@@ -1473,6 +1469,10 @@

All log runs

+ +
+
+
@@ -1489,10 +1489,6 @@

All log runs

- -
-
-
@@ -1569,6 +1565,10 @@

All log runs

+ +
+
+
@@ -1585,10 +1585,6 @@

All log runs

- -
-
-
@@ -1665,6 +1661,10 @@

All log runs

+ +
+
+
@@ -1681,10 +1681,6 @@

All log runs

- -
-
-
@@ -1761,6 +1757,10 @@

All log runs

+ +
+
+
@@ -1777,10 +1777,6 @@

All log runs

- -
-
-
@@ -1857,6 +1853,10 @@

All log runs

+ +
+
+
@@ -1873,10 +1873,6 @@

All log runs

- -
-
-
@@ -1953,6 +1949,10 @@

All log runs

+ +
+
+
@@ -1969,10 +1969,6 @@

All log runs

- -
-
-
@@ -2049,6 +2045,10 @@

All log runs

+ +
+
+
@@ -2065,10 +2065,6 @@

All log runs

- -
-
-
@@ -2145,6 +2141,10 @@

All log runs

+ +
+
+
@@ -2161,10 +2161,6 @@

All log runs

- -
-
-
@@ -2241,6 +2237,10 @@

All log runs

+ +
+
+
@@ -2257,10 +2257,6 @@

All log runs

- -
-
-
@@ -2337,6 +2333,10 @@

All log runs

+ +
+
+
@@ -2353,10 +2353,6 @@

All log runs

- -
-
-
@@ -2433,6 +2429,10 @@

All log runs

+ +
+
+
@@ -2449,10 +2449,6 @@

All log runs

- -
-
-
@@ -2529,6 +2525,10 @@

All log runs

+ +
+
+
@@ -2545,10 +2545,6 @@

All log runs

- -
-
-
@@ -2625,6 +2621,10 @@

All log runs

+ +
+
+
@@ -2641,10 +2641,6 @@

All log runs

- -
-
-
@@ -2721,6 +2717,10 @@

All log runs

+ +
+
+
@@ -2737,10 +2737,6 @@

All log runs

- -
-
-
@@ -2817,6 +2813,10 @@

All log runs

+ +
+
+
@@ -2833,10 +2833,6 @@

All log runs

- -
-
-
@@ -2913,6 +2909,10 @@

All log runs

+ +
+
+
@@ -2929,10 +2929,6 @@

All log runs

- -
-
-
@@ -3009,6 +3005,10 @@

All log runs

+ +
+
+
@@ -3025,10 +3025,6 @@

All log runs

- -
-
-
@@ -3105,6 +3101,10 @@

All log runs

+ +
+
+
@@ -3121,10 +3121,6 @@

All log runs

- -
-
-
@@ -3201,6 +3197,10 @@

All log runs

+ +
+
+
@@ -3217,10 +3217,6 @@

All log runs

- -
-
-
@@ -3297,6 +3293,10 @@

All log runs

+ +
+
+
@@ -3313,10 +3313,6 @@

All log runs

- -
-
-
@@ -3393,6 +3389,10 @@

All log runs

+ +
+
+
@@ -3409,10 +3409,6 @@

All log runs

- -
-
-
@@ -3489,6 +3485,10 @@

All log runs

+ +
+
+
@@ -3505,10 +3505,6 @@

All log runs

- -
-
-
@@ -3585,6 +3581,10 @@

All log runs

+ +
+
+
@@ -3601,10 +3601,6 @@

All log runs

- -
-
-
@@ -3681,6 +3677,10 @@

All log runs

+ +
+
+
@@ -3697,10 +3697,6 @@

All log runs

- -
-
-
@@ -3777,6 +3773,10 @@

All log runs

+ +
+
+
@@ -3793,10 +3793,6 @@

All log runs

- -
-
-
@@ -3873,6 +3869,10 @@

All log runs

+ +
+
+
@@ -3889,10 +3889,6 @@

All log runs

- -
-
-
@@ -3969,6 +3965,10 @@

All log runs

+ +
+
+
@@ -3985,10 +3985,6 @@

All log runs

- -
-
-
@@ -4065,6 +4061,10 @@

All log runs

+ +
+
+
@@ -4081,10 +4081,6 @@

All log runs

- -
-
-
@@ -4161,6 +4157,10 @@

All log runs

+ +
+
+
@@ -4177,10 +4177,6 @@

All log runs

- -
-
-
@@ -4257,6 +4253,10 @@

All log runs

+ +
+
+
@@ -4273,10 +4273,6 @@

All log runs

- -
-
-
@@ -4353,6 +4349,10 @@

All log runs

+ +
+
+
@@ -4369,10 +4369,6 @@

All log runs

- -
-
-
@@ -4449,6 +4445,10 @@

All log runs

+ +
+
+
@@ -4465,10 +4465,6 @@

All log runs

- -
-
-
@@ -4545,6 +4541,10 @@

All log runs

+ +
+
+
@@ -4561,10 +4561,6 @@

All log runs

- -
-
-
@@ -4641,6 +4637,10 @@

All log runs

+ +
+
+
@@ -4657,10 +4657,6 @@

All log runs

- -
-
-
@@ -4737,6 +4733,10 @@

All log runs

+ +
+
+
@@ -4753,10 +4753,6 @@

All log runs

- -
-
-
@@ -4833,6 +4829,10 @@

All log runs

+ +
+
+
@@ -4849,10 +4849,6 @@

All log runs

- -
-
-
@@ -4929,6 +4925,10 @@

All log runs

+ +
+
+
@@ -4945,10 +4945,6 @@

All log runs

- -
-
-
@@ -5025,6 +5021,10 @@

All log runs

+ +
+
+
@@ -5041,10 +5041,6 @@

All log runs

- -
-
-
@@ -5121,6 +5117,10 @@

All log runs

+ +
+
+
@@ -5137,10 +5137,6 @@

All log runs

- -
-
-
@@ -5217,6 +5213,10 @@

All log runs

+ +
+
+
@@ -5233,10 +5233,6 @@

All log runs

- -
-
-
@@ -5313,6 +5309,10 @@

All log runs

+ +
+
+
@@ -5329,10 +5329,6 @@

All log runs

- -
-
-
@@ -5409,6 +5405,10 @@

All log runs

+ +
+
+
@@ -5425,10 +5425,6 @@

All log runs

- -
-
-
@@ -5505,6 +5501,10 @@

All log runs

+ +
+
+
@@ -5521,10 +5521,6 @@

All log runs

- -
-
-
@@ -5601,6 +5597,10 @@

All log runs

+ +
+
+
@@ -5617,10 +5617,6 @@

All log runs

- -
-
-
@@ -5697,6 +5693,10 @@

All log runs

+ +
+
+
@@ -5713,10 +5713,6 @@

All log runs

- -
-
-
@@ -5793,6 +5789,10 @@

All log runs

+ +
+
+
@@ -5809,10 +5809,6 @@

All log runs

- -
-
-
@@ -5889,6 +5885,10 @@

All log runs

+ +
+
+
@@ -5905,10 +5905,6 @@

All log runs

- -
-
-
@@ -5985,6 +5981,10 @@

All log runs

+ +
+
+
@@ -6001,10 +6001,6 @@

All log runs

- -
-
-
@@ -6081,6 +6077,10 @@

All log runs

+ +
+
+
@@ -6097,10 +6097,6 @@

All log runs

- -
-
-
@@ -6177,6 +6173,10 @@

All log runs

+ +
+
+
@@ -6193,10 +6193,6 @@

All log runs

- -
-
-
@@ -6273,6 +6269,10 @@

All log runs

+ +
+
+
@@ -6289,10 +6289,6 @@

All log runs

- -
-
-
@@ -6369,6 +6365,10 @@

All log runs

+ +
+
+
@@ -6385,10 +6385,6 @@

All log runs

- -
-
-
@@ -6465,6 +6461,10 @@

All log runs

+ +
+
+
@@ -6481,10 +6481,6 @@

All log runs

- -
-
-
@@ -6561,6 +6557,10 @@

All log runs

+ +
+
+
@@ -6577,10 +6577,6 @@

All log runs

- -
-
-
@@ -6657,6 +6653,10 @@

All log runs

+ +
+
+
@@ -6673,10 +6673,6 @@

All log runs

- -
-
-
@@ -6753,6 +6749,10 @@

All log runs

+ +
+
+
@@ -6769,10 +6769,6 @@

All log runs

- -
-
-
@@ -6849,6 +6845,10 @@

All log runs

+ +
+
+
@@ -6865,10 +6865,6 @@

All log runs

- -
-
-
@@ -6945,6 +6941,10 @@

All log runs

+ +
+
+
@@ -6961,10 +6961,6 @@

All log runs

- -
-
-
@@ -7041,6 +7037,10 @@

All log runs

+ +
+
+
@@ -7057,10 +7057,6 @@

All log runs

- -
-
-
@@ -7137,6 +7133,10 @@

All log runs

+ +
+
+
@@ -7153,10 +7153,6 @@

All log runs

- -
-
-
@@ -7233,6 +7229,10 @@

All log runs

+ +
+
+
@@ -7249,10 +7249,6 @@

All log runs

- -
-
-
@@ -7329,6 +7325,10 @@

All log runs

+ +
+
+
@@ -7345,10 +7345,6 @@

All log runs

- -
-
-
@@ -7425,6 +7421,10 @@

All log runs

+ +
+
+
@@ -7441,10 +7441,6 @@

All log runs

- -
-
-
@@ -7521,6 +7517,10 @@

All log runs

+ +
+
+
@@ -7537,10 +7537,6 @@

All log runs

- -
-
-
@@ -7617,6 +7613,10 @@

All log runs

+ +
+
+
@@ -7633,10 +7633,6 @@

All log runs

- -
-
-
@@ -7713,6 +7709,10 @@

All log runs

+ +
+
+
@@ -7729,10 +7729,6 @@

All log runs

- -
-
-
@@ -7809,6 +7805,10 @@

All log runs

+ +
+
+
@@ -7825,10 +7825,6 @@

All log runs

- -
-
-
@@ -7905,6 +7901,10 @@

All log runs

+ +
+
+
@@ -7921,10 +7921,6 @@

All log runs

- -
-
-
@@ -8001,6 +7997,10 @@

All log runs

+ +
+
+
@@ -8017,10 +8017,6 @@

All log runs

- -
-
-
@@ -8097,6 +8093,10 @@

All log runs

+ +
+
+
@@ -8113,10 +8113,6 @@

All log runs

- -
-
-
@@ -8193,6 +8189,10 @@

All log runs

+ +
+
+
@@ -8209,10 +8209,6 @@

All log runs

- -
-
-
@@ -8289,6 +8285,10 @@

All log runs

+ +
+
+
@@ -8305,10 +8305,6 @@

All log runs

- -
-
-
@@ -8385,6 +8381,10 @@

All log runs

+ +
+
+
@@ -8401,10 +8401,6 @@

All log runs

- -
-
-
@@ -8481,6 +8477,10 @@

All log runs

+ +
+
+
@@ -8497,10 +8497,6 @@

All log runs

- -
-
-
@@ -8577,6 +8573,10 @@

All log runs

+ +
+
+
@@ -8593,10 +8593,6 @@

All log runs

- -
-
-
@@ -8673,6 +8669,10 @@

All log runs

+ +
+
+
@@ -8689,10 +8689,6 @@

All log runs

- -
-
-
@@ -8769,6 +8765,10 @@

All log runs

+ +
+
+
@@ -8785,10 +8785,6 @@

All log runs

- -
-
-
@@ -8865,6 +8861,10 @@

All log runs

+ +
+
+
@@ -8881,10 +8881,6 @@

All log runs

- -
-
-
@@ -8961,6 +8957,10 @@

All log runs

+ +
+
+
@@ -8977,10 +8977,6 @@

All log runs

- -
-
-
@@ -9057,6 +9053,10 @@

All log runs

+ +
+
+
@@ -9073,10 +9073,6 @@

All log runs

- -
-
-
@@ -9153,6 +9149,10 @@

All log runs

+ +
+
+
@@ -9169,10 +9169,6 @@

All log runs

- -
-
-
@@ -9249,6 +9245,10 @@

All log runs

+ +
+
+
@@ -9265,10 +9265,6 @@

All log runs

- -
-
-
@@ -9345,6 +9341,10 @@

All log runs

+ +
+
+
@@ -9361,10 +9361,6 @@

All log runs

- -
-
-
@@ -9441,6 +9437,10 @@

All log runs

+ +
+
+
@@ -9457,10 +9457,6 @@

All log runs

- -
-
-
@@ -9537,6 +9533,10 @@

All log runs

+ +
+
+
@@ -9553,10 +9553,6 @@

All log runs

- -
-
-
@@ -9633,6 +9629,10 @@

All log runs

+ +
+
+
@@ -9649,10 +9649,6 @@

All log runs

- -
-
-
@@ -9729,6 +9725,10 @@

All log runs

+ +
+
+
@@ -9745,10 +9745,6 @@

All log runs

- -
-
-
@@ -9825,6 +9821,10 @@

All log runs

+ +
+
+
@@ -9841,10 +9841,6 @@

All log runs

- -
-
-
@@ -9921,6 +9917,10 @@

All log runs

+ +
+
+
@@ -9937,10 +9937,6 @@

All log runs

- -
-
-
@@ -10017,6 +10013,10 @@

All log runs

+ +
+
+
@@ -10033,10 +10033,6 @@

All log runs

- -
-
-
@@ -10113,6 +10109,10 @@

All log runs

+ +
+
+
@@ -10129,10 +10129,6 @@

All log runs

- -
-
-
@@ -10209,6 +10205,10 @@

All log runs

+ +
+
+
@@ -10225,10 +10225,6 @@

All log runs

- -
-
-
@@ -10305,6 +10301,10 @@

All log runs

+ +
+
+
@@ -10321,10 +10321,6 @@

All log runs

- -
-
-
@@ -10401,6 +10397,10 @@

All log runs

+ +
+
+
@@ -10417,10 +10417,6 @@

All log runs

- -
-
-
@@ -10497,6 +10493,10 @@

All log runs

+ +
+
+
@@ -10513,10 +10513,6 @@

All log runs

- -
-
-
@@ -10593,6 +10589,10 @@

All log runs

+ +
+
+
@@ -10609,10 +10609,6 @@

All log runs

- -
-
-
@@ -10689,6 +10685,10 @@

All log runs

+ +
+
+
@@ -10705,10 +10705,6 @@

All log runs

- -
-
-
@@ -10785,6 +10781,10 @@

All log runs

+ +
+
+
@@ -10801,10 +10801,6 @@

All log runs

- -
-
-
@@ -10881,6 +10877,10 @@

All log runs

+ +
+
+
@@ -10897,10 +10897,6 @@

All log runs

- -
-
-
@@ -10977,6 +10973,10 @@

All log runs

+ +
+
+
@@ -10993,10 +10993,6 @@

All log runs

- -
-
-
@@ -11073,6 +11069,10 @@

All log runs

+ +
+
+
@@ -11089,10 +11089,6 @@

All log runs

- -
-
-
@@ -11169,6 +11165,10 @@

All log runs

+ +
+
+
@@ -11185,10 +11185,6 @@

All log runs

- -
-
-
@@ -11265,6 +11261,10 @@

All log runs

+ +
+
+
@@ -11281,10 +11281,6 @@

All log runs

- -
-
-
@@ -11361,6 +11357,10 @@

All log runs

+ +
+
+
@@ -11377,10 +11377,6 @@

All log runs

- -
-
-
@@ -11457,6 +11453,10 @@

All log runs

+ +
+
+
@@ -11473,10 +11473,6 @@

All log runs

- -
-
-
@@ -11553,6 +11549,10 @@

All log runs

+ +
+
+
@@ -11569,10 +11569,6 @@

All log runs

- -
-
-
@@ -11649,6 +11645,10 @@

All log runs

+ +
+
+
@@ -11665,10 +11665,6 @@

All log runs

- -
-
-
@@ -11745,6 +11741,10 @@

All log runs

+ +
+
+
@@ -11761,10 +11761,6 @@

All log runs

- -
-
-
@@ -11841,6 +11837,10 @@

All log runs

+ +
+
+
@@ -11857,10 +11857,6 @@

All log runs

- -
-
-
@@ -11937,6 +11933,10 @@

All log runs

+ +
+
+
@@ -11953,10 +11953,6 @@

All log runs

- -
-
-
@@ -12033,6 +12029,10 @@

All log runs

+ +
+
+
@@ -12049,10 +12049,6 @@

All log runs

- -
-
-
@@ -12129,6 +12125,10 @@

All log runs

+ +
+
+
@@ -12145,10 +12145,6 @@

All log runs

- -
-
-
@@ -12225,6 +12221,10 @@

All log runs

+ +
+
+
@@ -12241,10 +12241,6 @@

All log runs

- -
-
-
@@ -12321,6 +12317,10 @@

All log runs

+ +
+
+
@@ -12337,10 +12337,6 @@

All log runs

- -
-
-
@@ -12417,6 +12413,10 @@

All log runs

+ +
+
+
@@ -12433,10 +12433,6 @@

All log runs

- -
-
-
@@ -12513,6 +12509,10 @@

All log runs

+ +
+
+
@@ -12529,10 +12529,6 @@

All log runs

- -
-
-
@@ -12609,6 +12605,10 @@

All log runs

+ +
+
+
@@ -12625,10 +12625,6 @@

All log runs

- -
-
-
@@ -12705,6 +12701,10 @@

All log runs

+ +
+
+
@@ -12721,10 +12721,6 @@

All log runs

- -
-
-
@@ -12801,26 +12797,26 @@

All log runs

+ +
+
+
- + HAR - OK + ERROR
- -
-
-
@@ -12897,6 +12893,10 @@

All log runs

+ +
+
+
@@ -12913,10 +12913,6 @@

All log runs

- -
-
-
@@ -12993,6 +12989,10 @@

All log runs

+ +
+
+
@@ -13009,10 +13009,6 @@

All log runs

- -
-
-
@@ -13089,6 +13085,10 @@

All log runs

+ +
+
+
@@ -13105,10 +13105,6 @@

All log runs

- -
-
-
@@ -13185,6 +13181,10 @@

All log runs

+ +
+
+
@@ -13201,10 +13201,6 @@

All log runs

- -
-
-
@@ -13281,6 +13277,10 @@

All log runs

+ +
+
+
@@ -13297,10 +13297,6 @@

All log runs

- -
-
-
@@ -13377,6 +13373,10 @@

All log runs

+ +
+
+
@@ -13393,10 +13393,6 @@

All log runs

- -
-
-
@@ -13473,6 +13469,10 @@

All log runs

+ +
+
+
@@ -13489,10 +13489,6 @@

All log runs

- -
-
-
@@ -13569,6 +13565,10 @@

All log runs

+ +
+
+
@@ -13585,10 +13585,6 @@

All log runs

- -
-
-
@@ -13665,6 +13661,10 @@

All log runs

+ +
+
+
@@ -13681,10 +13681,6 @@

All log runs

- -
-
-
@@ -13761,6 +13757,10 @@

All log runs

+ +
+
+
@@ -13777,10 +13777,6 @@

All log runs

- -
-
-
@@ -13857,6 +13853,10 @@

All log runs

+ +
+
+
@@ -13873,10 +13873,6 @@

All log runs

- -
-
-
@@ -13953,6 +13949,10 @@

All log runs

+ +
+
+
@@ -13969,10 +13969,6 @@

All log runs

- -
-
-
@@ -14049,6 +14045,10 @@

All log runs

+ +
+
+
@@ -14065,10 +14065,6 @@

All log runs

- -
-
-
@@ -14145,6 +14141,10 @@

All log runs

+ +
+
+
@@ -14161,10 +14161,6 @@

All log runs

- -
-
-
@@ -14241,6 +14237,10 @@

All log runs

+ +
+
+
@@ -14257,10 +14257,6 @@

All log runs

- -
-
-
@@ -14337,6 +14333,10 @@

All log runs

+ +
+
+
@@ -14353,10 +14353,6 @@

All log runs

- -
-
-
@@ -14433,6 +14429,10 @@

All log runs

+ +
+
+
@@ -14449,10 +14449,6 @@

All log runs

- -
-
-
@@ -14529,6 +14525,10 @@

All log runs

+ +
+
+
@@ -14545,10 +14545,6 @@

All log runs

- -
-
-
@@ -14625,6 +14621,10 @@

All log runs

+ +
+
+
@@ -14641,10 +14641,6 @@

All log runs

- -
-
-
@@ -14721,6 +14717,10 @@

All log runs

+ +
+
+
@@ -14737,10 +14737,6 @@

All log runs

- -
-
-
@@ -14817,6 +14813,10 @@

All log runs

+ +
+
+
@@ -14833,10 +14833,6 @@

All log runs

- -
-
-
@@ -14913,6 +14909,10 @@

All log runs

+ +
+
+
@@ -14929,10 +14929,6 @@

All log runs

- -
-
-
@@ -15009,6 +15005,10 @@

All log runs

+ +
+
+
@@ -15025,10 +15025,6 @@

All log runs

- -
-
-
@@ -15105,6 +15101,10 @@

All log runs

+ +
+
+
@@ -15121,10 +15121,6 @@

All log runs

- -
-
-
@@ -15201,6 +15197,10 @@

All log runs

+ +
+
+
@@ -15217,10 +15217,6 @@

All log runs

- -
-
-
@@ -15297,6 +15293,10 @@

All log runs

+ +
+
+
@@ -15313,10 +15313,6 @@

All log runs

- -
-
-
@@ -15393,6 +15389,10 @@

All log runs

+ +
+
+
@@ -15409,10 +15409,6 @@

All log runs

- -
-
-
@@ -15489,6 +15485,10 @@

All log runs

+ +
+
+
@@ -15505,10 +15505,6 @@

All log runs

- -
-
-
@@ -15585,6 +15581,10 @@

All log runs

+ +
+
+
@@ -15601,10 +15601,6 @@

All log runs

- -
-
-
@@ -15681,6 +15677,10 @@

All log runs

+ +
+
+
@@ -15697,10 +15697,6 @@

All log runs

- -
-
-
@@ -15777,6 +15773,10 @@

All log runs

+ +
+
+
@@ -15793,10 +15793,6 @@

All log runs

- -
-
-
@@ -15873,6 +15869,10 @@

All log runs

+ +
+
+
@@ -15889,10 +15889,6 @@

All log runs

- -
-
-
@@ -15969,6 +15965,10 @@

All log runs

+ +
+
+
@@ -15985,10 +15985,6 @@

All log runs

- -
-
-
@@ -16065,6 +16061,10 @@

All log runs

+ +
+
+
@@ -16081,10 +16081,6 @@

All log runs

- -
-
-
@@ -16161,6 +16157,10 @@

All log runs

+ +
+
+
@@ -16177,10 +16177,6 @@

All log runs

- -
-
-
@@ -16257,6 +16253,10 @@

All log runs

+ +
+
+
@@ -16273,10 +16273,6 @@

All log runs

- -
-
-
@@ -16353,6 +16349,10 @@

All log runs

+ +
+
+
@@ -16369,10 +16369,6 @@

All log runs

- -
-
-
@@ -16449,6 +16445,10 @@

All log runs

+ +
+
+
@@ -16465,10 +16465,6 @@

All log runs

- -
-
-
@@ -16545,6 +16541,10 @@

All log runs

+ +
+
+
@@ -16561,10 +16561,6 @@

All log runs

- -
-
-
@@ -16641,6 +16637,10 @@

All log runs

+ +
+
+
@@ -16657,10 +16657,6 @@

All log runs

- -
-
-
@@ -16737,6 +16733,10 @@

All log runs

+ +
+
+
@@ -16753,10 +16753,6 @@

All log runs

- -
-
-
@@ -16833,6 +16829,10 @@

All log runs

+ +
+
+
@@ -16849,10 +16849,6 @@

All log runs

- -
-
-
@@ -16929,6 +16925,10 @@

All log runs

+ +
+
+
@@ -16945,10 +16945,6 @@

All log runs

- -
-
-
@@ -17025,6 +17021,10 @@

All log runs

+ +
+
+
@@ -17041,10 +17041,6 @@

All log runs

- -
-
-
@@ -17121,6 +17117,10 @@

All log runs

+ +
+
+
@@ -17137,10 +17137,6 @@

All log runs

- -
-
-
@@ -17217,6 +17213,10 @@

All log runs

+ +
+
+
@@ -17233,10 +17233,6 @@

All log runs

- -
-
-
@@ -17313,6 +17309,10 @@

All log runs

+ +
+
+
@@ -17329,10 +17329,6 @@

All log runs

- -
-
-
@@ -17409,6 +17405,10 @@

All log runs

+ +
+
+
@@ -17425,10 +17425,6 @@

All log runs

- -
-
-
@@ -17505,6 +17501,10 @@

All log runs

+ +
+
+
@@ -17521,10 +17521,6 @@

All log runs

- -
-
-
@@ -17601,6 +17597,10 @@

All log runs

+ +
+
+
@@ -17617,10 +17617,6 @@

All log runs

- -
-
-
@@ -17697,6 +17693,10 @@

All log runs

+ +
+
+
@@ -17713,10 +17713,6 @@

All log runs

- -
-
-
@@ -17793,6 +17789,10 @@

All log runs

+ +
+
+
@@ -17809,10 +17809,6 @@

All log runs

- -
-
-
@@ -17889,6 +17885,10 @@

All log runs

+ +
+
+
@@ -17905,10 +17905,6 @@

All log runs

- -
-
-
@@ -17985,6 +17981,10 @@

All log runs

+ +
+
+
@@ -18001,10 +18001,6 @@

All log runs

- -
-
-
@@ -18081,6 +18077,10 @@

All log runs

+ +
+
+
@@ -18097,10 +18097,6 @@

All log runs

- -
-
-
@@ -18177,6 +18173,10 @@

All log runs

+ +
+
+
@@ -18193,10 +18193,6 @@

All log runs

- -
-
-
@@ -18273,6 +18269,10 @@

All log runs

+ +
+
+
@@ -18289,10 +18289,6 @@

All log runs

- -
-
-
@@ -18369,6 +18365,10 @@

All log runs

+ +
+
+
@@ -18385,10 +18385,6 @@

All log runs

- -
-
-
@@ -18465,6 +18461,10 @@

All log runs

+ +
+
+
@@ -18481,10 +18481,6 @@

All log runs

- -
-
-
@@ -18561,6 +18557,10 @@

All log runs

+ +
+
+
@@ -18577,10 +18577,6 @@

All log runs

- -
-
-
@@ -18657,6 +18653,10 @@

All log runs

+ +
+
+
@@ -18673,10 +18673,6 @@

All log runs

- -
-
-
@@ -18753,6 +18749,10 @@

All log runs

+ +
+
+
@@ -18769,10 +18769,6 @@

All log runs

- -
-
-
@@ -18849,6 +18845,10 @@

All log runs

+ +
+
+
@@ -18865,10 +18865,6 @@

All log runs

- -
-
-
@@ -18945,6 +18941,10 @@

All log runs

+ +
+
+
@@ -18961,10 +18961,6 @@

All log runs

- -
-
-
@@ -19041,6 +19037,10 @@

All log runs

+ +
+
+
@@ -19057,10 +19057,6 @@

All log runs

- -
-
-
@@ -19137,6 +19133,10 @@

All log runs

+ +
+
+
@@ -19153,10 +19153,6 @@

All log runs

- -
-
-
@@ -19233,6 +19229,10 @@

All log runs

+ +
+
+
@@ -19249,10 +19249,6 @@

All log runs

- -
-
-
@@ -19329,6 +19325,10 @@

All log runs

+ +
+
+
@@ -19345,10 +19345,6 @@

All log runs

- -
-
-
@@ -19425,6 +19421,10 @@

All log runs

+ +
+
+
@@ -19441,10 +19441,6 @@

All log runs

- -
-
-
@@ -19521,6 +19517,10 @@

All log runs

+ +
+
+
@@ -19537,10 +19537,6 @@

All log runs

- -
-
-
@@ -19617,6 +19613,10 @@

All log runs

+ +
+
+
@@ -19633,10 +19633,6 @@

All log runs

- -
-
-
@@ -19713,6 +19709,10 @@

All log runs

+ +
+
+
@@ -19729,10 +19729,6 @@

All log runs

- -
-
-
@@ -19809,6 +19805,10 @@

All log runs

+ +
+
+
@@ -19825,10 +19825,6 @@

All log runs

- -
-
-
@@ -19905,6 +19901,10 @@

All log runs

+ +
+
+
@@ -19921,10 +19921,6 @@

All log runs

- -
-
-
@@ -20001,6 +19997,10 @@

All log runs

+ +
+
+
@@ -20017,10 +20017,6 @@

All log runs

- -
-
-
@@ -20097,6 +20093,10 @@

All log runs

+ +
+
+
@@ -20113,10 +20113,6 @@

All log runs

- -
-
-
@@ -20193,6 +20189,10 @@

All log runs

+ +
+
+
@@ -20209,10 +20209,6 @@

All log runs

- -
-
-
@@ -20289,6 +20285,10 @@

All log runs

+ +
+
+
@@ -20305,10 +20305,6 @@

All log runs

- -
-
-
@@ -20385,6 +20381,10 @@

All log runs

+ +
+
+
@@ -20401,10 +20401,6 @@

All log runs

- -
-
-
@@ -20481,6 +20477,10 @@

All log runs

+ +
+
+
@@ -20497,10 +20497,6 @@

All log runs

- -
-
-
@@ -20577,6 +20573,10 @@

All log runs

+ +
+
+
@@ -20593,10 +20593,6 @@

All log runs

- -
-
-
@@ -20673,6 +20669,10 @@

All log runs

+ +
+
+
@@ -20689,10 +20689,6 @@

All log runs

- -
-
-
@@ -20769,6 +20765,10 @@

All log runs

+ +
+
+
@@ -20785,10 +20785,6 @@

All log runs

- -
-
-
@@ -20865,6 +20861,10 @@

All log runs

+ +
+
+
@@ -20881,10 +20881,6 @@

All log runs

- -
-
-
@@ -20961,6 +20957,10 @@

All log runs

+ +
+
+
@@ -20977,10 +20977,6 @@

All log runs

- -
-
-
@@ -21057,6 +21053,10 @@

All log runs

+ +
+
+
@@ -21073,10 +21073,6 @@

All log runs

- -
-
-
@@ -21153,6 +21149,10 @@

All log runs

+ +
+
+
@@ -21169,10 +21169,6 @@

All log runs

- -
-
-
@@ -21249,6 +21245,10 @@

All log runs

+ +
+
+
@@ -21265,83 +21265,83 @@

All log runs

- +
- +
- +
- +
- +
- -
+
+
- +
- -
+
+
- +
- +
- -
+
+
- +
- +
- +
- +
- +
- +
- +
- +
- +
@@ -21361,10 +21361,6 @@

All log runs

- -
-
-
@@ -21441,6 +21437,10 @@

All log runs

+ +
+
+
@@ -21457,10 +21457,6 @@

All log runs

- -
-
-
@@ -21537,6 +21533,10 @@

All log runs

+ +
+
+
@@ -21553,10 +21553,6 @@

All log runs

- -
-
-
@@ -21633,6 +21629,10 @@

All log runs

+ +
+
+
@@ -21649,10 +21649,6 @@

All log runs

- -
-
-
@@ -21729,6 +21725,10 @@

All log runs

+ +
+
+
@@ -21745,10 +21745,6 @@

All log runs

- -
-
-
@@ -21825,6 +21821,10 @@

All log runs

+ +
+
+
@@ -21841,10 +21841,6 @@

All log runs

- -
-
-
@@ -21921,6 +21917,10 @@

All log runs

+ +
+
+
@@ -21937,10 +21937,6 @@

All log runs

- -
-
-
@@ -22017,6 +22013,10 @@

All log runs

+ +
+
+
@@ -22033,10 +22033,6 @@

All log runs

- -
-
-
@@ -22113,6 +22109,10 @@

All log runs

+ +
+
+
@@ -22129,10 +22129,6 @@

All log runs

- -
-
-
@@ -22209,6 +22205,10 @@

All log runs

+ +
+
+
@@ -22225,10 +22225,6 @@

All log runs

- -
-
-
@@ -22305,6 +22301,10 @@

All log runs

+ +
+
+
@@ -22321,10 +22321,6 @@

All log runs

- -
-
-
@@ -22401,6 +22397,10 @@

All log runs

+ +
+
+
@@ -22417,10 +22417,6 @@

All log runs

- -
-
-
@@ -22497,6 +22493,10 @@

All log runs

+ +
+
+
@@ -22513,10 +22513,6 @@

All log runs

- -
-
-
@@ -22593,6 +22589,10 @@

All log runs

+ +
+
+
@@ -22609,10 +22609,6 @@

All log runs

- -
-
-
@@ -22689,6 +22685,10 @@

All log runs

+ +
+
+
@@ -22705,10 +22705,6 @@

All log runs

- -
-
-
@@ -22785,6 +22781,10 @@

All log runs

+ +
+
+
@@ -22801,10 +22801,6 @@

All log runs

- -
-
-
@@ -22881,6 +22877,10 @@

All log runs

+ +
+
+
@@ -22897,10 +22897,6 @@

All log runs

- -
-
-
@@ -22977,6 +22973,10 @@

All log runs

+ +
+
+
@@ -22993,10 +22993,6 @@

All log runs

- -
-
-
@@ -23073,6 +23069,10 @@

All log runs

+ +
+
+
@@ -23089,10 +23089,6 @@

All log runs

- -
-
-
@@ -23169,6 +23165,10 @@

All log runs

+ +
+
+
@@ -23185,10 +23185,6 @@

All log runs

- -
-
-
@@ -23265,6 +23261,10 @@

All log runs

+ +
+
+
@@ -23281,10 +23281,6 @@

All log runs

- -
-
-
@@ -23361,6 +23357,10 @@

All log runs

+ +
+
+
@@ -23377,10 +23377,6 @@

All log runs

- -
-
-
@@ -23457,6 +23453,10 @@

All log runs

+ +
+
+
@@ -23473,10 +23473,6 @@

All log runs

- -
-
-
@@ -23553,6 +23549,10 @@

All log runs

+ +
+
+
@@ -23569,10 +23569,6 @@

All log runs

- -
-
-
@@ -23649,6 +23645,10 @@

All log runs

+ +
+
+
@@ -23665,10 +23665,6 @@

All log runs

- -
-
-
@@ -23745,6 +23741,10 @@

All log runs

+ +
+
+
@@ -23761,10 +23761,6 @@

All log runs

- -
-
-
@@ -23841,6 +23837,10 @@

All log runs

+ +
+
+
@@ -23857,10 +23857,6 @@

All log runs

- -
-
-
@@ -23937,6 +23933,10 @@

All log runs

+ +
+
+
@@ -23953,10 +23953,6 @@

All log runs

- -
-
-
@@ -24033,6 +24029,10 @@

All log runs

+ +
+
+
@@ -24049,10 +24049,6 @@

All log runs

- -
-
-
@@ -24129,6 +24125,10 @@

All log runs

+ +
+
+
@@ -24145,10 +24145,6 @@

All log runs

- -
-
-
@@ -24225,6 +24221,10 @@

All log runs

+ +
+
+
@@ -24241,10 +24241,6 @@

All log runs

- -
-
-
@@ -24321,6 +24317,10 @@

All log runs

+ +
+
+
@@ -24337,10 +24337,6 @@

All log runs

- -
-
-
@@ -24417,6 +24413,10 @@

All log runs

+ +
+
+
@@ -24433,10 +24433,6 @@

All log runs

- -
-
-
@@ -24513,6 +24509,10 @@

All log runs

+ +
+
+
@@ -24529,10 +24529,6 @@

All log runs

- -
-
-
@@ -24609,6 +24605,10 @@

All log runs

+ +
+
+
@@ -24625,10 +24625,6 @@

All log runs

- -
-
-
@@ -24705,6 +24701,10 @@

All log runs

+ +
+
+
@@ -24721,10 +24721,6 @@

All log runs

- -
-
-
@@ -24801,6 +24797,10 @@

All log runs

+ +
+
+
@@ -24817,10 +24817,6 @@

All log runs

- -
-
-
@@ -24897,6 +24893,10 @@

All log runs

+ +
+
+
@@ -24913,10 +24913,6 @@

All log runs

- -
-
-
@@ -24993,6 +24989,10 @@

All log runs

+ +
+
+
@@ -25009,10 +25009,6 @@

All log runs

- -
-
-
@@ -25089,6 +25085,10 @@

All log runs

+ +
+
+
@@ -25201,10 +25201,6 @@

All log runs

- -
-
-
@@ -25281,6 +25277,10 @@

All log runs

+ +
+
+
@@ -25297,10 +25297,6 @@

All log runs

- -
-
-
@@ -25377,6 +25373,10 @@

All log runs

+ +
+
+
@@ -25393,10 +25393,6 @@

All log runs

- -
-
-
@@ -25473,6 +25469,10 @@

All log runs

+ +
+
+
@@ -25489,10 +25489,6 @@

All log runs

- -
-
-
@@ -25569,6 +25565,10 @@

All log runs

+ +
+
+
@@ -25585,10 +25585,6 @@

All log runs

- -
-
-
@@ -25665,6 +25661,10 @@

All log runs

+ +
+
+
@@ -25681,10 +25681,6 @@

All log runs

- -
-
-
@@ -25761,6 +25757,10 @@

All log runs

+ +
+
+
@@ -25777,10 +25777,6 @@

All log runs

- -
-
-
@@ -25857,6 +25853,10 @@

All log runs

+ +
+
+
@@ -25873,10 +25873,6 @@

All log runs

- -
-
-
@@ -25953,6 +25949,10 @@

All log runs

+ +
+
+
@@ -25969,10 +25969,6 @@

All log runs

- -
-
-
@@ -26049,6 +26045,10 @@

All log runs

+ +
+
+
@@ -26065,10 +26065,6 @@

All log runs

- -
-
-
@@ -26145,6 +26141,10 @@

All log runs

+ +
+
+
@@ -26161,10 +26161,6 @@

All log runs

- -
-
-
@@ -26241,6 +26237,10 @@

All log runs

+ +
+
+
@@ -26257,10 +26257,6 @@

All log runs

- -
-
-
@@ -26337,6 +26333,10 @@

All log runs

+ +
+
+
@@ -26353,10 +26353,6 @@

All log runs

- -
-
-
@@ -26433,6 +26429,10 @@

All log runs

+ +
+
+
@@ -26449,10 +26449,6 @@

All log runs

- -
-
-
@@ -26529,6 +26525,10 @@

All log runs

+ +
+
+
@@ -26545,10 +26545,6 @@

All log runs

- -
-
-
@@ -26625,6 +26621,10 @@

All log runs

+ +
+
+
@@ -26641,10 +26641,6 @@

All log runs

- -
-
-
@@ -26721,6 +26717,10 @@

All log runs

+ +
+
+
@@ -26737,10 +26737,6 @@

All log runs

- -
-
-
@@ -26817,6 +26813,10 @@

All log runs

+ +
+
+
@@ -26833,10 +26833,6 @@

All log runs

- -
-
-
@@ -26913,6 +26909,10 @@

All log runs

+ +
+
+
@@ -26929,10 +26929,6 @@

All log runs

- -
-
-
@@ -27009,6 +27005,10 @@

All log runs

+ +
+
+
@@ -27025,10 +27025,6 @@

All log runs

- -
-
-
@@ -27105,6 +27101,10 @@

All log runs

+ +
+
+
@@ -27121,10 +27121,6 @@

All log runs

- -
-
-
@@ -27201,6 +27197,10 @@

All log runs

+ +
+
+
@@ -27217,10 +27217,6 @@

All log runs

- -
-
-
@@ -27297,6 +27293,10 @@

All log runs

+ +
+
+
@@ -27313,10 +27313,6 @@

All log runs

- -
-
-
@@ -27393,6 +27389,10 @@

All log runs

+ +
+
+
@@ -27409,10 +27409,6 @@

All log runs

- -
-
-
@@ -27489,6 +27485,10 @@

All log runs

+ +
+
+
@@ -27505,10 +27505,6 @@

All log runs

- -
-
-
@@ -27585,6 +27581,10 @@

All log runs

+ +
+
+
@@ -27601,10 +27601,6 @@

All log runs

- -
-
-
@@ -27681,6 +27677,10 @@

All log runs

+ +
+
+
@@ -27697,10 +27697,6 @@

All log runs

- -
-
-
@@ -27777,6 +27773,10 @@

All log runs

+ +
+
+
@@ -27793,10 +27793,6 @@

All log runs

- -
-
-
@@ -27873,6 +27869,10 @@

All log runs

+ +
+
+
@@ -27889,10 +27889,6 @@

All log runs

- -
-
-
@@ -27969,6 +27965,10 @@

All log runs

+ +
+
+
@@ -27985,10 +27985,6 @@

All log runs

- -
-
-
@@ -28065,6 +28061,10 @@

All log runs

+ +
+
+
@@ -28081,10 +28081,6 @@

All log runs

- -
-
-
@@ -28161,6 +28157,10 @@

All log runs

+ +
+
+
@@ -28177,10 +28177,6 @@

All log runs

- -
-
-
@@ -28257,6 +28253,10 @@

All log runs

+ +
+
+
@@ -28273,10 +28273,6 @@

All log runs

- -
-
-
@@ -28353,6 +28349,10 @@

All log runs

+ +
+
+
@@ -28369,10 +28369,6 @@

All log runs

- -
-
-
@@ -28449,6 +28445,10 @@

All log runs

+ +
+
+
@@ -28465,10 +28465,6 @@

All log runs

- -
-
-
@@ -28545,6 +28541,10 @@

All log runs

+ +
+
+
@@ -28561,10 +28561,6 @@

All log runs

- -
-
-
@@ -28641,6 +28637,10 @@

All log runs

+ +
+
+
@@ -28657,10 +28657,6 @@

All log runs

- -
-
-
@@ -28737,6 +28733,10 @@

All log runs

+ +
+
+
@@ -28753,10 +28753,6 @@

All log runs

- -
-
-
@@ -28833,6 +28829,10 @@

All log runs

+ +
+
+
@@ -28849,10 +28849,6 @@

All log runs

- -
-
-
@@ -28929,6 +28925,10 @@

All log runs

+ +
+
+
@@ -28945,10 +28945,6 @@

All log runs

- -
-
-
@@ -29025,6 +29021,10 @@

All log runs

+ +
+
+
@@ -29041,10 +29041,6 @@

All log runs

- -
-
-
@@ -29121,6 +29117,10 @@

All log runs

+ +
+
+
@@ -29137,10 +29137,6 @@

All log runs

- -
-
-
@@ -29217,6 +29213,10 @@

All log runs

+ +
+
+
@@ -29233,10 +29233,6 @@

All log runs

- -
-
-
@@ -29313,6 +29309,10 @@

All log runs

+ +
+
+
@@ -29329,10 +29329,6 @@

All log runs

- -
-
-
@@ -29409,6 +29405,10 @@

All log runs

+ +
+
+
@@ -29425,10 +29425,6 @@

All log runs

- -
-
-
@@ -29505,6 +29501,10 @@

All log runs

+ +
+
+
@@ -29521,10 +29521,6 @@

All log runs

- -
-
-
@@ -29601,6 +29597,10 @@

All log runs

+ +
+
+
@@ -29617,10 +29617,6 @@

All log runs

- -
-
-
@@ -29697,26 +29693,26 @@

All log runs

+ +
+
+
- + TRF - ERROR + OK
- -
-
-
@@ -29793,6 +29789,10 @@

All log runs

+ +
+
+
@@ -29809,10 +29809,6 @@

All log runs

- -
-
-
@@ -29889,6 +29885,10 @@

All log runs

+ +
+
+
@@ -29905,10 +29905,6 @@

All log runs

- -
-
-
@@ -29985,6 +29981,10 @@

All log runs

+ +
+
+
@@ -30001,10 +30001,6 @@

All log runs

- -
-
-
@@ -30081,6 +30077,10 @@

All log runs

+ +
+
+
@@ -30097,10 +30097,6 @@

All log runs

- -
-
-
@@ -30177,6 +30173,10 @@

All log runs

+ +
+
+
@@ -30193,10 +30193,6 @@

All log runs

- -
-
-
@@ -30273,6 +30269,10 @@

All log runs

+ +
+
+
@@ -30289,10 +30289,6 @@

All log runs

- -
-
-
@@ -30369,6 +30365,10 @@

All log runs

+ +
+
+
@@ -30385,10 +30385,6 @@

All log runs

- -
-
-
@@ -30465,6 +30461,10 @@

All log runs

+ +
+
+
@@ -30481,10 +30481,6 @@

All log runs

- -
-
-
@@ -30561,6 +30557,10 @@

All log runs

+ +
+
+
@@ -30577,10 +30577,6 @@

All log runs

- -
-
-
@@ -30657,6 +30653,10 @@

All log runs

+ +
+
+
@@ -30673,10 +30673,6 @@

All log runs

- -
-
-
@@ -30753,6 +30749,10 @@

All log runs

+ +
+
+
@@ -30769,10 +30769,6 @@

All log runs

- -
-
-
@@ -30849,6 +30845,10 @@

All log runs

+ +
+
+
@@ -30865,10 +30865,6 @@

All log runs

- -
-
-
@@ -30945,6 +30941,10 @@

All log runs

+ +
+
+
@@ -30961,10 +30961,6 @@

All log runs

- -
-
-
@@ -31041,6 +31037,10 @@

All log runs

+ +
+
+
@@ -31057,10 +31057,6 @@

All log runs

- -
-
-
@@ -31137,6 +31133,10 @@

All log runs

+ +
+
+
@@ -31153,10 +31153,6 @@

All log runs

- -
-
-
@@ -31233,6 +31229,10 @@

All log runs

+ +
+
+
@@ -31249,10 +31249,6 @@

All log runs

- -
-
-
@@ -31329,6 +31325,10 @@

All log runs

+ +
+
+
@@ -31345,10 +31345,6 @@

All log runs

- -
-
-
@@ -31425,6 +31421,10 @@

All log runs

+ +
+
+
@@ -31441,10 +31441,6 @@

All log runs

- -
-
-
@@ -31521,6 +31517,10 @@

All log runs

+ +
+
+
@@ -31537,10 +31537,6 @@

All log runs

- -
-
-
@@ -31617,6 +31613,10 @@

All log runs

+ +
+
+
@@ -31633,10 +31633,6 @@

All log runs

- -
-
-
@@ -31713,6 +31709,10 @@

All log runs

+ +
+
+
@@ -31729,10 +31729,6 @@

All log runs

- -
-
-
@@ -31809,6 +31805,10 @@

All log runs

+ +
+
+
@@ -31825,10 +31825,6 @@

All log runs

- -
-
-
@@ -31905,6 +31901,10 @@

All log runs

+ +
+
+
@@ -31921,10 +31921,6 @@

All log runs

- -
-
-
@@ -32001,6 +31997,10 @@

All log runs

+ +
+
+
@@ -32017,10 +32017,6 @@

All log runs

- -
-
-
@@ -32097,6 +32093,10 @@

All log runs

+ +
+
+
@@ -32113,10 +32113,6 @@

All log runs

- -
-
-
@@ -32193,6 +32189,10 @@

All log runs

+ +
+
+
@@ -32209,10 +32209,6 @@

All log runs

- -
-
-
@@ -32289,6 +32285,10 @@

All log runs

+ +
+
+
@@ -32305,10 +32305,6 @@

All log runs

- -
-
-
@@ -32385,6 +32381,10 @@

All log runs

+ +
+
+
@@ -32401,10 +32401,6 @@

All log runs

- -
-
-
@@ -32481,6 +32477,10 @@

All log runs

+ +
+
+
@@ -32497,10 +32497,6 @@

All log runs

- -
-
-
@@ -32577,6 +32573,10 @@

All log runs

+ +
+
+
@@ -32593,10 +32593,6 @@

All log runs

- -
-
-
@@ -32673,6 +32669,10 @@

All log runs

+ +
+
+
@@ -32689,10 +32689,6 @@

All log runs

- -
-
-
@@ -32769,6 +32765,10 @@

All log runs

+ +
+
+
@@ -32785,10 +32785,6 @@

All log runs

- -
-
-
@@ -32865,6 +32861,10 @@

All log runs

+ +
+
+
@@ -32881,10 +32881,6 @@

All log runs

- -
-
-
@@ -32961,6 +32957,10 @@

All log runs

+ +
+
+
@@ -32977,10 +32977,6 @@

All log runs

- -
-
-
@@ -33057,6 +33053,10 @@

All log runs

+ +
+
+
@@ -33073,10 +33073,6 @@

All log runs

- -
-
-
@@ -33153,6 +33149,10 @@

All log runs

+ +
+
+
@@ -33169,10 +33169,6 @@

All log runs

- -
-
-
@@ -33249,6 +33245,10 @@

All log runs

+ +
+
+
@@ -33265,10 +33265,6 @@

All log runs

- -
-
-
@@ -33345,6 +33341,10 @@

All log runs

+ +
+
+
@@ -33361,10 +33361,6 @@

All log runs

- -
-
-
@@ -33441,6 +33437,10 @@

All log runs

+ +
+
+
@@ -33457,10 +33457,6 @@

All log runs

- -
-
-
@@ -33537,6 +33533,10 @@

All log runs

+ +
+
+
@@ -33553,10 +33553,6 @@

All log runs

- -
-
-
@@ -33633,6 +33629,10 @@

All log runs

+ +
+
+
@@ -33649,10 +33649,6 @@

All log runs

- -
-
-
@@ -33729,6 +33725,10 @@

All log runs

+ +
+
+
@@ -33745,10 +33745,6 @@

All log runs

- -
-
-
@@ -33825,6 +33821,10 @@

All log runs

+ +
+
+
@@ -33841,10 +33841,6 @@

All log runs

- -
-
-
@@ -33921,6 +33917,10 @@

All log runs

+ +
+
+
diff --git a/logbooks/ABC/index.html b/logbooks/ABC/index.html index 8f8f7b2762..266353fb06 100644 --- a/logbooks/ABC/index.html +++ b/logbooks/ABC/index.html @@ -37,6 +37,56 @@ +

2024-12-11

+
+
+
+
+
Duration
+
7 seconds
+
+
+
Start
+
2024-12-11 08:54:57.309663
+
+
+
End
+
2024-12-11 08:55:04.968858
+
+
+
Status code
+
0
+
+
+
Error
+
+
+
+
+
+ +

Run log

+
[08:54:57] Fetching Scraper for: ABC                              handlers.py:23
+           Begin attempting to scrape: ABC                        handlers.py:27
+           Deleting existing data...                                 base.py:257
+[08:54:58] Getting all files in Councillors...                       base.py:209
+           Getting all files in Councillors/json...                  base.py:209
+           ...found 16 files in Councillors/json                     base.py:225
+           Getting all files in Councillors/raw...                   base.py:209
+           ...found 16 files in Councillors/raw                      base.py:225
+           ...found 33 files in Councillors                          base.py:225
+           Deleting batch no. 1 consisting of 33 files               base.py:236
+[08:54:59] ...data deleted.                                          base.py:264
+           Scraping from                                              base.py:49
+           https://www.armaghbanbridgecraigavon.gov.uk/councillors/             
+[08:55:03] Committing batch 1 consisting of 32 files                 base.py:297
+[08:55:04] Finished attempting to scrape: ABC                        base.py:345
+
+ + + + +

2024-12-10

@@ -984,56 +1034,6 @@

Run log

- - - -

2024-11-21

-
-
-
-
-
Duration
-
6 seconds
-
-
-
Start
-
2024-11-21 08:26:42.024600
-
-
-
End
-
2024-11-21 08:26:48.808169
-
-
-
Status code
-
0
-
-
-
Error
-
-
-
-
-
- -

Run log

-
[08:26:42] Fetching Scraper for: ABC                              handlers.py:23
-           Begin attempting to scrape: ABC                        handlers.py:27
-           Deleting existing data...                                 base.py:257
-           Getting all files in Councillors...                       base.py:209
-[08:26:43] Getting all files in Councillors/json...                  base.py:209
-           ...found 16 files in Councillors/json                     base.py:225
-           Getting all files in Councillors/raw...                   base.py:209
-           ...found 16 files in Councillors/raw                      base.py:225
-           ...found 33 files in Councillors                          base.py:225
-           Deleting batch no. 1 consisting of 33 files               base.py:236
-[08:26:44] ...data deleted.                                          base.py:264
-           Scraping from                                              base.py:49
-           https://www.armaghbanbridgecraigavon.gov.uk/councillors/             
-[08:26:47] Committing batch 1 consisting of 32 files                 base.py:297
-[08:26:48] Finished attempting to scrape: ABC                        base.py:345
-
- -