From 00000d9e8c74263ed2cba19a7699b47e843aac7b Mon Sep 17 00:00:00 2001 From: doms9 <96013514+doms9@users.noreply.github.com> Date: Sat, 18 Apr 2026 13:53:53 -0400 Subject: [PATCH] e - misc edits. --- M3U8/scrapers/streamhub.py | 1 + M3U8/scrapers/utils/caching.py | 10 +++++----- M3U8/scrapers/utils/config.py | 2 +- M3U8/scrapers/utils/webwork.py | 11 ++++++++--- 4 files changed, 15 insertions(+), 9 deletions(-) diff --git a/M3U8/scrapers/streamhub.py b/M3U8/scrapers/streamhub.py index 7d0ebf35..4d4b4c29 100644 --- a/M3U8/scrapers/streamhub.py +++ b/M3U8/scrapers/streamhub.py @@ -227,6 +227,7 @@ async def scrape(browser: Browser) -> None: if events := await get_events(cached_urls.keys()): log.info(f"Processing {len(events)} new URL(s)") + async with network.event_context(browser) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: diff --git a/M3U8/scrapers/utils/caching.py b/M3U8/scrapers/utils/caching.py index 3a86f5db..3fe23923 100644 --- a/M3U8/scrapers/utils/caching.py +++ b/M3U8/scrapers/utils/caching.py @@ -45,11 +45,11 @@ class Cache: if per_entry: return {k: v for k, v in data.items() if self.is_fresh(v)} - if index: - ts: float | int = data[index].get("timestamp", Time.default_8()) - - else: - ts: float | int = data.get("timestamp", Time.default_8()) + ts: float | int = ( + data[index].get("timestamp", Time.default_8()) + if index + else data.get("timestamp", Time.default_8()) + ) dt_ts = Time.clean(Time.from_ts(ts)).timestamp() diff --git a/M3U8/scrapers/utils/config.py b/M3U8/scrapers/utils/config.py index 0aba68dc..3d9c5fc6 100644 --- a/M3U8/scrapers/utils/config.py +++ b/M3U8/scrapers/utils/config.py @@ -153,7 +153,7 @@ class Leagues: pattern = re.compile(r"\s+(?:-|vs\.?|at|@)\s+", re.I) if pattern.search(event): - t1, t2 = re.split(pattern, event) + t1, t2 = pattern.split(event) return any(t in self.teams(league) for t in (t1.strip(), t2.strip())) diff --git a/M3U8/scrapers/utils/webwork.py b/M3U8/scrapers/utils/webwork.py index 618ae4be..044c16cc 100644 --- a/M3U8/scrapers/utils/webwork.py +++ b/M3U8/scrapers/utils/webwork.py @@ -217,9 +217,14 @@ class Network: got_one: asyncio.Event, ) -> None: - invalids = ["amazonaws", "knitcdn", "jwpltx"] - - escaped = [re.escape(i) for i in invalids] + escaped = [ + re.escape(i) + for i in { + "amazonaws", + "knitcdn", + "jwpltx", + } + ] pattern = re.compile(rf"^(?!.*({'|'.join(escaped)})).*\.m3u8", re.I)