From 00000d956aa9bdfa64ed3eb607145879a52c82d2 Mon Sep 17 00:00:00 2001 From: doms9 <96013514+doms9@users.noreply.github.com> Date: Tue, 17 Mar 2026 22:26:10 -0400 Subject: [PATCH] e - rename glisco.py to s2watch.py --- M3U8/fetch.py | 8 ++++---- M3U8/scrapers/{glisco.py => s2watch.py} | 6 +++--- 2 files changed, 7 insertions(+), 7 deletions(-) rename M3U8/scrapers/{glisco.py => s2watch.py} (97%) diff --git a/M3U8/fetch.py b/M3U8/fetch.py index b12f72c6..04e147f1 100644 --- a/M3U8/fetch.py +++ b/M3U8/fetch.py @@ -8,13 +8,13 @@ from scrapers import ( cdnlivetv, embedhd, fawa, - glisco, istreameast, livetvsx, ovogoal, pawa, ppv, roxie, + s2watch, shark, sportzone, streambtw, @@ -76,10 +76,10 @@ async def main() -> None: httpx_tasks = [ asyncio.create_task(fawa.scrape()), - asyncio.create_task(glisco.scrape()), asyncio.create_task(istreameast.scrape()), asyncio.create_task(ovogoal.scrape()), # asyncio.create_task(pawa.scrape()), + asyncio.create_task(s2watch.scrape()), asyncio.create_task(shark.scrape()), asyncio.create_task(streambtw.scrape()), asyncio.create_task(totalsportek1.scrape()), @@ -92,7 +92,7 @@ async def main() -> None: await asyncio.gather(*(pw_tasks + httpx_tasks)) # others - await livetvsx.scrape(xtrnl_brwsr) + # await livetvsx.scrape(xtrnl_brwsr) finally: await hdl_brwsr.close() @@ -105,13 +105,13 @@ async def main() -> None: cdnlivetv.urls | embedhd.urls | fawa.urls - | glisco.urls | istreameast.urls | livetvsx.urls | ovogoal.urls | pawa.urls | ppv.urls | roxie.urls + | s2watch.urls | shark.urls | sportzone.urls | streambtw.urls diff --git a/M3U8/scrapers/glisco.py b/M3U8/scrapers/s2watch.py similarity index 97% rename from M3U8/scrapers/glisco.py rename to M3U8/scrapers/s2watch.py index 7f18578d..ce23416a 100644 --- a/M3U8/scrapers/glisco.py +++ b/M3U8/scrapers/s2watch.py @@ -10,7 +10,7 @@ log = get_logger(__name__) urls: dict[str, dict[str, str | float]] = {} -TAG = "GLISCO" +TAG = "S2WATCH" CACHE_FILE = Cache(TAG, exp=10_800) @@ -79,7 +79,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]: events = [] - start_dt = now.delta(hours=-3) + start_dt = now.delta(minutes=-30) end_dt = now.delta(minutes=30) for info in api_data: @@ -127,7 +127,7 @@ async def scrape() -> None: log.info(f"Loaded {cached_count} event(s) from cache") - log.info('Scraping from "https://www.glisco.link/"') + log.info('Scraping from "https://s2watch.me"') if events := await get_events(cached_urls.keys()): log.info(f"Processing {len(events)} new URL(s)")