diff --git a/M3U8/scrapers/livetvsx.py b/M3U8/scrapers/livetvsx.py index 198acfa..d29afb8 100644 --- a/M3U8/scrapers/livetvsx.py +++ b/M3U8/scrapers/livetvsx.py @@ -126,15 +126,16 @@ async def refresh_xml_cache(now_ts: float) -> dict[str, dict[str, str | float]]: feed = feedparser.parse(xml_data.content) for entry in feed.entries: - title = entry.get("title") + if not (date := entry.get("published")): + continue - link = entry.get("link") + if not (link := entry.get("link")): + continue - sport_sum = entry.get("summary") + if not (title := entry.get("title")): + continue - date = entry.get("published") - - if not all([title, link, sport_sum, date]): + if not (sport_sum := entry.get("summary")): continue sprt = sport_sum.split(".", 1) diff --git a/M3U8/scrapers/roxie.py b/M3U8/scrapers/roxie.py index 4c3b9da..33f58cd 100644 --- a/M3U8/scrapers/roxie.py +++ b/M3U8/scrapers/roxie.py @@ -78,7 +78,7 @@ async def process_event( url: str, url_num: int, page: Page, -) -> tuple[str | None, str | None]: +) -> str | None: captured: list[str] = [] diff --git a/M3U8/scrapers/streambtw.py b/M3U8/scrapers/streambtw.py index 4c7eaf0..a64476e 100644 --- a/M3U8/scrapers/streambtw.py +++ b/M3U8/scrapers/streambtw.py @@ -15,7 +15,7 @@ CACHE_FILE = Cache(TAG, exp=3_600) API_FILE = Cache(f"{TAG}-api", exp=19_800) -BASE_URL = "https://hiteasport.info" +BASE_URL = "https://streambtw.com" def fix_league(s: str) -> str: diff --git a/M3U8/scrapers/watchfooty.py b/M3U8/scrapers/watchfooty.py index 8d69929..a91fdcc 100644 --- a/M3U8/scrapers/watchfooty.py +++ b/M3U8/scrapers/watchfooty.py @@ -102,7 +102,7 @@ async def process_event( await page.wait_for_timeout(2_000) try: - header = await page.wait_for_selector("text=/Stream Links/i", timeout=5_000) + header = await page.wait_for_selector("text=/Stream Links/i", timeout=4_000) text = await header.inner_text() except TimeoutError: