mirror of
https://github.com/doms9/iptv.git
synced 2026-04-21 19:46:59 +02:00
e
- misc edits.
This commit is contained in:
parent
edc80fba5f
commit
00000d9e8c
4 changed files with 15 additions and 9 deletions
|
|
@ -227,6 +227,7 @@ async def scrape(browser: Browser) -> None:
|
||||||
|
|
||||||
if events := await get_events(cached_urls.keys()):
|
if events := await get_events(cached_urls.keys()):
|
||||||
log.info(f"Processing {len(events)} new URL(s)")
|
log.info(f"Processing {len(events)} new URL(s)")
|
||||||
|
|
||||||
async with network.event_context(browser) as context:
|
async with network.event_context(browser) as context:
|
||||||
for i, ev in enumerate(events, start=1):
|
for i, ev in enumerate(events, start=1):
|
||||||
async with network.event_page(context) as page:
|
async with network.event_page(context) as page:
|
||||||
|
|
|
||||||
|
|
@ -45,11 +45,11 @@ class Cache:
|
||||||
if per_entry:
|
if per_entry:
|
||||||
return {k: v for k, v in data.items() if self.is_fresh(v)}
|
return {k: v for k, v in data.items() if self.is_fresh(v)}
|
||||||
|
|
||||||
if index:
|
ts: float | int = (
|
||||||
ts: float | int = data[index].get("timestamp", Time.default_8())
|
data[index].get("timestamp", Time.default_8())
|
||||||
|
if index
|
||||||
else:
|
else data.get("timestamp", Time.default_8())
|
||||||
ts: float | int = data.get("timestamp", Time.default_8())
|
)
|
||||||
|
|
||||||
dt_ts = Time.clean(Time.from_ts(ts)).timestamp()
|
dt_ts = Time.clean(Time.from_ts(ts)).timestamp()
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -153,7 +153,7 @@ class Leagues:
|
||||||
pattern = re.compile(r"\s+(?:-|vs\.?|at|@)\s+", re.I)
|
pattern = re.compile(r"\s+(?:-|vs\.?|at|@)\s+", re.I)
|
||||||
|
|
||||||
if pattern.search(event):
|
if pattern.search(event):
|
||||||
t1, t2 = re.split(pattern, event)
|
t1, t2 = pattern.split(event)
|
||||||
|
|
||||||
return any(t in self.teams(league) for t in (t1.strip(), t2.strip()))
|
return any(t in self.teams(league) for t in (t1.strip(), t2.strip()))
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -217,9 +217,14 @@ class Network:
|
||||||
got_one: asyncio.Event,
|
got_one: asyncio.Event,
|
||||||
) -> None:
|
) -> None:
|
||||||
|
|
||||||
invalids = ["amazonaws", "knitcdn", "jwpltx"]
|
escaped = [
|
||||||
|
re.escape(i)
|
||||||
escaped = [re.escape(i) for i in invalids]
|
for i in {
|
||||||
|
"amazonaws",
|
||||||
|
"knitcdn",
|
||||||
|
"jwpltx",
|
||||||
|
}
|
||||||
|
]
|
||||||
|
|
||||||
pattern = re.compile(rf"^(?!.*({'|'.join(escaped)})).*\.m3u8", re.I)
|
pattern = re.compile(rf"^(?!.*({'|'.join(escaped)})).*\.m3u8", re.I)
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue