mirror of
https://github.com/doms9/iptv.git
synced 2026-04-21 19:46:59 +02:00
e
- misc edits.
This commit is contained in:
parent
714b781f4f
commit
00000d9cf1
2 changed files with 17 additions and 7 deletions
|
|
@ -45,6 +45,8 @@ async def get_events() -> list[dict[str, str]]:
|
||||||
if not (html_data := await network.request(BASE_URL, log=log)):
|
if not (html_data := await network.request(BASE_URL, log=log)):
|
||||||
return events
|
return events
|
||||||
|
|
||||||
|
now = Time.clean(Time.now())
|
||||||
|
|
||||||
soup = HTMLParser(html_data.content)
|
soup = HTMLParser(html_data.content)
|
||||||
|
|
||||||
for row in soup.css(".row"):
|
for row in soup.css(".row"):
|
||||||
|
|
@ -55,9 +57,14 @@ async def get_events() -> list[dict[str, str]]:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
for a in row.css("a.list-group-item[href]"):
|
for a in row.css("a.list-group-item[href]"):
|
||||||
splits = a.text(strip=True).split(":")
|
x, y = a.text(strip=True).split(":", 1)
|
||||||
|
|
||||||
event_name = ":".join(splits[:2]).split("@")[0].strip()
|
event_name = x.split("@")[0].strip()
|
||||||
|
|
||||||
|
event_dt = Time.from_str(y.split(":", 1)[-1], timezone="UTC")
|
||||||
|
|
||||||
|
if event_dt.date() != now.date():
|
||||||
|
continue
|
||||||
|
|
||||||
if not (href := a.attributes.get("href")):
|
if not (href := a.attributes.get("href")):
|
||||||
continue
|
continue
|
||||||
|
|
@ -67,6 +74,7 @@ async def get_events() -> list[dict[str, str]]:
|
||||||
"sport": sport,
|
"sport": sport,
|
||||||
"event": event_name,
|
"event": event_name,
|
||||||
"link": urljoin(f"{html_data.url}", href),
|
"link": urljoin(f"{html_data.url}", href),
|
||||||
|
"timestamp": now.timestamp(),
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
@ -86,8 +94,6 @@ async def scrape() -> None:
|
||||||
if events := await get_events():
|
if events := await get_events():
|
||||||
log.info(f"Processing {len(events)} URL(s)")
|
log.info(f"Processing {len(events)} URL(s)")
|
||||||
|
|
||||||
now = Time.clean(Time.now())
|
|
||||||
|
|
||||||
for i, ev in enumerate(events, start=1):
|
for i, ev in enumerate(events, start=1):
|
||||||
handler = partial(
|
handler = partial(
|
||||||
process_event,
|
process_event,
|
||||||
|
|
@ -102,7 +108,11 @@ async def scrape() -> None:
|
||||||
log=log,
|
log=log,
|
||||||
)
|
)
|
||||||
|
|
||||||
sport, event = ev["sport"], ev["event"]
|
sport, event, ts = (
|
||||||
|
ev["sport"],
|
||||||
|
ev["event"],
|
||||||
|
ev["timestamp"],
|
||||||
|
)
|
||||||
|
|
||||||
key = f"[{sport}] {event} ({TAG})"
|
key = f"[{sport}] {event} ({TAG})"
|
||||||
|
|
||||||
|
|
@ -112,7 +122,7 @@ async def scrape() -> None:
|
||||||
"url": url,
|
"url": url,
|
||||||
"logo": logo,
|
"logo": logo,
|
||||||
"base": BASE_URL,
|
"base": BASE_URL,
|
||||||
"timestamp": now.timestamp(),
|
"timestamp": ts,
|
||||||
"id": tvg_id or "Live.Event.us",
|
"id": tvg_id or "Live.Event.us",
|
||||||
"link": link,
|
"link": link,
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -97,7 +97,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
|
||||||
|
|
||||||
if (
|
if (
|
||||||
not (live_badge := card.css_first("span.bg-green-600"))
|
not (live_badge := card.css_first("span.bg-green-600"))
|
||||||
or live_badge.text(strip=True) != "LIVE"
|
or live_badge.text(strip=True).lower().strip() != "live"
|
||||||
):
|
):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue