e
This commit is contained in:
parent
e92fb81e94
commit
00000d9fd3
2 changed files with 6 additions and 4 deletions
|
|
@ -55,7 +55,8 @@ async def get_events(
|
|||
|
||||
soup = HTMLParser(r.text)
|
||||
|
||||
pattern = re.compile(r"\d{1,2}:\d{1,2}")
|
||||
valid_event = re.compile(r"\d{1,2}:\d{1,2}")
|
||||
clean_event = re.compile(r"\s+-+\s+CH")
|
||||
|
||||
events = []
|
||||
|
||||
|
|
@ -77,15 +78,15 @@ async def get_events(
|
|||
|
||||
event_name, details = text.text(strip=True), subtext.text(strip=True)
|
||||
|
||||
if not (pattern.search(details)):
|
||||
if not (valid_event.search(details)):
|
||||
continue
|
||||
|
||||
sport = pattern.split(details)[0].strip()
|
||||
sport = valid_event.split(details)[0].strip()
|
||||
|
||||
events.append(
|
||||
{
|
||||
"sport": sport,
|
||||
"event": event_name.replace(" --- CH", ""),
|
||||
"event": clean_event.sub("", event_name),
|
||||
"link": urljoin(BASE_URL, href),
|
||||
"href": href,
|
||||
}
|
||||
|
|
|
|||
|
|
@ -261,6 +261,7 @@ async def scrape(client: httpx.AsyncClient) -> None:
|
|||
"base": base_url,
|
||||
"timestamp": ts,
|
||||
"id": tvg_id or "Live.Event.us",
|
||||
"link": ev["link"],
|
||||
}
|
||||
|
||||
cached_urls[key] = entry
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue