diff --git a/M3U8/scrapers/livetvsx.py b/M3U8/scrapers/livetvsx.py index 76866f9f..90087d24 100644 --- a/M3U8/scrapers/livetvsx.py +++ b/M3U8/scrapers/livetvsx.py @@ -3,7 +3,7 @@ import re from functools import partial import feedparser -from playwright.async_api import Browser, Page +from playwright.async_api import Browser, Page, TimeoutError from .utils import Cache, Time, get_logger, leagues, network @@ -47,6 +47,8 @@ async def process_event( page: Page, ) -> str | None: + event_id_pattern = re.compile(r"&c=(\d*)", re.I) + captured: list[str] = [] got_one = asyncio.Event() @@ -57,8 +59,6 @@ async def process_event( got_one=got_one, ) - event_id_pattern = re.compile(r"&c=(\d*)", re.I) - page.on("request", handler) try: @@ -69,23 +69,21 @@ async def process_event( ) if resp.status != 200: - log.warning(f"URL {url_num}) status code: {resp.status}") + log.warning(f"URL {url_num}) Status Code: {resp.status}") return try: event_a = page.locator('a[title*="Aliez"]').first href = await event_a.get_attribute("href", timeout=1_250) - except TimeoutError: log.warning(f"URL {url_num}) No valid sources found.") return - if match := event_id_pattern.search(href): - event_id = match[1] - + if (match := event_id_pattern.search(href)) and ( + event_id := match[1] + ).isalnum(): event_url = f"https://emb.apl392.me/player/live.php?id={event_id}" - else: event_url = href if href.startswith("http") else f"https:{href}" diff --git a/M3U8/scrapers/pixel.py b/M3U8/scrapers/pixel.py index 28a47b97..d6902921 100644 --- a/M3U8/scrapers/pixel.py +++ b/M3U8/scrapers/pixel.py @@ -26,7 +26,7 @@ async def get_api_data(page: Page) -> dict[str, list[dict, str, str]]: ) if resp.status != 200: - log.warning(f"{url} status code: {resp.status}") + log.warning(f"{url} Status Code: {resp.status}") return {} diff --git a/M3U8/scrapers/roxie.py b/M3U8/scrapers/roxie.py index 96d553b9..d927ebf2 100644 --- a/M3U8/scrapers/roxie.py +++ b/M3U8/scrapers/roxie.py @@ -100,7 +100,7 @@ async def process_event( ) if resp.status != 200: - log.warning(f"URL {url_num}) status code: {resp.status}") + log.warning(f"URL {url_num}) Status Code: {resp.status}") return try: diff --git a/M3U8/scrapers/utils/webwork.py b/M3U8/scrapers/utils/webwork.py index 31f26b32..0dff8150 100644 --- a/M3U8/scrapers/utils/webwork.py +++ b/M3U8/scrapers/utils/webwork.py @@ -257,7 +257,7 @@ class Network: ) if resp.status != 200: - log.warning(f"URL {url_num}) status code: {resp.status}") + log.warning(f"URL {url_num}) Status Code: {resp.status}") return diff --git a/M3U8/scrapers/watchfooty.py b/M3U8/scrapers/watchfooty.py index dbd5aa24..f31943ef 100644 --- a/M3U8/scrapers/watchfooty.py +++ b/M3U8/scrapers/watchfooty.py @@ -96,7 +96,7 @@ async def process_event( ) if resp.status != 200: - log.warning(f"URL {url_num}) status code: {resp.status}") + log.warning(f"URL {url_num}) Status Code: {resp.status}") return await page.wait_for_timeout(2_000)