- rename glisco.py to s2watch.py
This commit is contained in:
doms9 2026-03-17 22:26:10 -04:00
parent 00000d9da6
commit 00000d956a
2 changed files with 7 additions and 7 deletions

View file

@ -8,13 +8,13 @@ from scrapers import (
cdnlivetv, cdnlivetv,
embedhd, embedhd,
fawa, fawa,
glisco,
istreameast, istreameast,
livetvsx, livetvsx,
ovogoal, ovogoal,
pawa, pawa,
ppv, ppv,
roxie, roxie,
s2watch,
shark, shark,
sportzone, sportzone,
streambtw, streambtw,
@ -76,10 +76,10 @@ async def main() -> None:
httpx_tasks = [ httpx_tasks = [
asyncio.create_task(fawa.scrape()), asyncio.create_task(fawa.scrape()),
asyncio.create_task(glisco.scrape()),
asyncio.create_task(istreameast.scrape()), asyncio.create_task(istreameast.scrape()),
asyncio.create_task(ovogoal.scrape()), asyncio.create_task(ovogoal.scrape()),
# asyncio.create_task(pawa.scrape()), # asyncio.create_task(pawa.scrape()),
asyncio.create_task(s2watch.scrape()),
asyncio.create_task(shark.scrape()), asyncio.create_task(shark.scrape()),
asyncio.create_task(streambtw.scrape()), asyncio.create_task(streambtw.scrape()),
asyncio.create_task(totalsportek1.scrape()), asyncio.create_task(totalsportek1.scrape()),
@ -92,7 +92,7 @@ async def main() -> None:
await asyncio.gather(*(pw_tasks + httpx_tasks)) await asyncio.gather(*(pw_tasks + httpx_tasks))
# others # others
await livetvsx.scrape(xtrnl_brwsr) # await livetvsx.scrape(xtrnl_brwsr)
finally: finally:
await hdl_brwsr.close() await hdl_brwsr.close()
@ -105,13 +105,13 @@ async def main() -> None:
cdnlivetv.urls cdnlivetv.urls
| embedhd.urls | embedhd.urls
| fawa.urls | fawa.urls
| glisco.urls
| istreameast.urls | istreameast.urls
| livetvsx.urls | livetvsx.urls
| ovogoal.urls | ovogoal.urls
| pawa.urls | pawa.urls
| ppv.urls | ppv.urls
| roxie.urls | roxie.urls
| s2watch.urls
| shark.urls | shark.urls
| sportzone.urls | sportzone.urls
| streambtw.urls | streambtw.urls

View file

@ -10,7 +10,7 @@ log = get_logger(__name__)
urls: dict[str, dict[str, str | float]] = {} urls: dict[str, dict[str, str | float]] = {}
TAG = "GLISCO" TAG = "S2WATCH"
CACHE_FILE = Cache(TAG, exp=10_800) CACHE_FILE = Cache(TAG, exp=10_800)
@ -79,7 +79,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
events = [] events = []
start_dt = now.delta(hours=-3) start_dt = now.delta(minutes=-30)
end_dt = now.delta(minutes=30) end_dt = now.delta(minutes=30)
for info in api_data: for info in api_data:
@ -127,7 +127,7 @@ async def scrape() -> None:
log.info(f"Loaded {cached_count} event(s) from cache") log.info(f"Loaded {cached_count} event(s) from cache")
log.info('Scraping from "https://www.glisco.link/"') log.info('Scraping from "https://s2watch.me"')
if events := await get_events(cached_urls.keys()): if events := await get_events(cached_urls.keys()):
log.info(f"Processing {len(events)} new URL(s)") log.info(f"Processing {len(events)} new URL(s)")