mirror of
https://github.com/doms9/iptv.git
synced 2026-04-21 19:46:59 +02:00
e
- rename glisco.py to s2watch.py
This commit is contained in:
parent
00000d9da6
commit
00000d956a
2 changed files with 7 additions and 7 deletions
|
|
@ -8,13 +8,13 @@ from scrapers import (
|
||||||
cdnlivetv,
|
cdnlivetv,
|
||||||
embedhd,
|
embedhd,
|
||||||
fawa,
|
fawa,
|
||||||
glisco,
|
|
||||||
istreameast,
|
istreameast,
|
||||||
livetvsx,
|
livetvsx,
|
||||||
ovogoal,
|
ovogoal,
|
||||||
pawa,
|
pawa,
|
||||||
ppv,
|
ppv,
|
||||||
roxie,
|
roxie,
|
||||||
|
s2watch,
|
||||||
shark,
|
shark,
|
||||||
sportzone,
|
sportzone,
|
||||||
streambtw,
|
streambtw,
|
||||||
|
|
@ -76,10 +76,10 @@ async def main() -> None:
|
||||||
|
|
||||||
httpx_tasks = [
|
httpx_tasks = [
|
||||||
asyncio.create_task(fawa.scrape()),
|
asyncio.create_task(fawa.scrape()),
|
||||||
asyncio.create_task(glisco.scrape()),
|
|
||||||
asyncio.create_task(istreameast.scrape()),
|
asyncio.create_task(istreameast.scrape()),
|
||||||
asyncio.create_task(ovogoal.scrape()),
|
asyncio.create_task(ovogoal.scrape()),
|
||||||
# asyncio.create_task(pawa.scrape()),
|
# asyncio.create_task(pawa.scrape()),
|
||||||
|
asyncio.create_task(s2watch.scrape()),
|
||||||
asyncio.create_task(shark.scrape()),
|
asyncio.create_task(shark.scrape()),
|
||||||
asyncio.create_task(streambtw.scrape()),
|
asyncio.create_task(streambtw.scrape()),
|
||||||
asyncio.create_task(totalsportek1.scrape()),
|
asyncio.create_task(totalsportek1.scrape()),
|
||||||
|
|
@ -92,7 +92,7 @@ async def main() -> None:
|
||||||
await asyncio.gather(*(pw_tasks + httpx_tasks))
|
await asyncio.gather(*(pw_tasks + httpx_tasks))
|
||||||
|
|
||||||
# others
|
# others
|
||||||
await livetvsx.scrape(xtrnl_brwsr)
|
# await livetvsx.scrape(xtrnl_brwsr)
|
||||||
|
|
||||||
finally:
|
finally:
|
||||||
await hdl_brwsr.close()
|
await hdl_brwsr.close()
|
||||||
|
|
@ -105,13 +105,13 @@ async def main() -> None:
|
||||||
cdnlivetv.urls
|
cdnlivetv.urls
|
||||||
| embedhd.urls
|
| embedhd.urls
|
||||||
| fawa.urls
|
| fawa.urls
|
||||||
| glisco.urls
|
|
||||||
| istreameast.urls
|
| istreameast.urls
|
||||||
| livetvsx.urls
|
| livetvsx.urls
|
||||||
| ovogoal.urls
|
| ovogoal.urls
|
||||||
| pawa.urls
|
| pawa.urls
|
||||||
| ppv.urls
|
| ppv.urls
|
||||||
| roxie.urls
|
| roxie.urls
|
||||||
|
| s2watch.urls
|
||||||
| shark.urls
|
| shark.urls
|
||||||
| sportzone.urls
|
| sportzone.urls
|
||||||
| streambtw.urls
|
| streambtw.urls
|
||||||
|
|
|
||||||
|
|
@ -10,7 +10,7 @@ log = get_logger(__name__)
|
||||||
|
|
||||||
urls: dict[str, dict[str, str | float]] = {}
|
urls: dict[str, dict[str, str | float]] = {}
|
||||||
|
|
||||||
TAG = "GLISCO"
|
TAG = "S2WATCH"
|
||||||
|
|
||||||
CACHE_FILE = Cache(TAG, exp=10_800)
|
CACHE_FILE = Cache(TAG, exp=10_800)
|
||||||
|
|
||||||
|
|
@ -79,7 +79,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
|
||||||
|
|
||||||
events = []
|
events = []
|
||||||
|
|
||||||
start_dt = now.delta(hours=-3)
|
start_dt = now.delta(minutes=-30)
|
||||||
end_dt = now.delta(minutes=30)
|
end_dt = now.delta(minutes=30)
|
||||||
|
|
||||||
for info in api_data:
|
for info in api_data:
|
||||||
|
|
@ -127,7 +127,7 @@ async def scrape() -> None:
|
||||||
|
|
||||||
log.info(f"Loaded {cached_count} event(s) from cache")
|
log.info(f"Loaded {cached_count} event(s) from cache")
|
||||||
|
|
||||||
log.info('Scraping from "https://www.glisco.link/"')
|
log.info('Scraping from "https://s2watch.me"')
|
||||||
|
|
||||||
if events := await get_events(cached_urls.keys()):
|
if events := await get_events(cached_urls.keys()):
|
||||||
log.info(f"Processing {len(events)} new URL(s)")
|
log.info(f"Processing {len(events)} new URL(s)")
|
||||||
Loading…
Add table
Add a link
Reference in a new issue