diff --git a/M3U8/scrapers/fawa.py b/M3U8/scrapers/fawa.py index 83db6b8..fa6f1fb 100644 --- a/M3U8/scrapers/fawa.py +++ b/M3U8/scrapers/fawa.py @@ -113,7 +113,7 @@ async def scrape(client: httpx.AsyncClient) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - now = Time.now().timestamp() + now = Time.clean(Time.now()).timestamp() for i, ev in enumerate(events, start=1): handler = partial( diff --git a/M3U8/scrapers/lotus.py b/M3U8/scrapers/lotus.py index 831d674..33936fa 100644 --- a/M3U8/scrapers/lotus.py +++ b/M3U8/scrapers/lotus.py @@ -111,7 +111,7 @@ async def scrape(client: httpx.AsyncClient) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - now = Time.now().timestamp() + now = Time.clean(Time.now()).timestamp() async with async_playwright() as p: browser, context = await network.browser(p) diff --git a/M3U8/scrapers/sport9.py b/M3U8/scrapers/sport9.py index 8347da3..fe40803 100644 --- a/M3U8/scrapers/sport9.py +++ b/M3U8/scrapers/sport9.py @@ -118,7 +118,7 @@ async def scrape(client: httpx.AsyncClient) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - now = Time.now().timestamp() + now = Time.clean(Time.now()).timestamp() async with async_playwright() as p: browser, context = await network.browser(p, browser="brave") diff --git a/M3U8/scrapers/streamfree.py b/M3U8/scrapers/streamfree.py index 2a1ba2f..809b43e 100644 --- a/M3U8/scrapers/streamfree.py +++ b/M3U8/scrapers/streamfree.py @@ -32,7 +32,7 @@ async def get_events(client: httpx.AsyncClient) -> dict[str, dict[str, str | flo events = {} - now = Time.now().timestamp() + now = Time.clean(Time.now()).timestamp() for streams in api_data.get("streams", {}).values(): if not streams: diff --git a/M3U8/scrapers/streamhub.py b/M3U8/scrapers/streamhub.py index 6b259e5..b9b218a 100644 --- a/M3U8/scrapers/streamhub.py +++ b/M3U8/scrapers/streamhub.py @@ -172,7 +172,7 @@ async def scrape(client: httpx.AsyncClient) -> None: if events: async with async_playwright() as p: - browser, context = await network.browser(p) + browser, context = await network.browser(p, browser="brave") for i, ev in enumerate(events, start=1): handler = partial( diff --git a/M3U8/scrapers/streamsgate.py b/M3U8/scrapers/streamsgate.py index 39037cc..dd86b30 100644 --- a/M3U8/scrapers/streamsgate.py +++ b/M3U8/scrapers/streamsgate.py @@ -153,7 +153,7 @@ async def scrape(client: httpx.AsyncClient) -> None: if events: async with async_playwright() as p: - browser, context = await network.browser(p) + browser, context = await network.browser(p, browser="brave") for i, ev in enumerate(events, start=1): handler = partial( diff --git a/M3U8/scrapers/tvpass.py b/M3U8/scrapers/tvpass.py index 2fb9b32..1e6cb8f 100644 --- a/M3U8/scrapers/tvpass.py +++ b/M3U8/scrapers/tvpass.py @@ -28,7 +28,7 @@ async def get_data(client: httpx.AsyncClient) -> list[str]: async def get_events(client: httpx.AsyncClient) -> dict[str, dict[str, str | float]]: - now = Time.now().timestamp() + now = Time.clean(Time.now()).timestamp() events = {} diff --git a/M3U8/scrapers/watchfooty.py b/M3U8/scrapers/watchfooty.py index fcc9ce7..e15a41e 100644 --- a/M3U8/scrapers/watchfooty.py +++ b/M3U8/scrapers/watchfooty.py @@ -72,7 +72,7 @@ async def refresh_api_cache( for ev in data: ev["ts"] = ev.pop("timestamp") - data[-1]["timestamp"] = Time.now().timestamp() + data[-1]["timestamp"] = Time.clean(Time.now()).timestamp() return data diff --git a/M3U8/scrapers/webcast.py b/M3U8/scrapers/webcast.py index 61e2e0a..6343ea3 100644 --- a/M3U8/scrapers/webcast.py +++ b/M3U8/scrapers/webcast.py @@ -35,7 +35,7 @@ async def refresh_html_cache( return {} - now = Time.now() + now = Time.clean(Time.now()) soup = HTMLParser(r.content)