mirror of
https://github.com/doms9/iptv.git
synced 2026-04-21 19:46:59 +02:00
e
- rename glisco.py to s2watch.py
This commit is contained in:
parent
00000d9da6
commit
00000d956a
2 changed files with 7 additions and 7 deletions
|
|
@ -8,13 +8,13 @@ from scrapers import (
|
|||
cdnlivetv,
|
||||
embedhd,
|
||||
fawa,
|
||||
glisco,
|
||||
istreameast,
|
||||
livetvsx,
|
||||
ovogoal,
|
||||
pawa,
|
||||
ppv,
|
||||
roxie,
|
||||
s2watch,
|
||||
shark,
|
||||
sportzone,
|
||||
streambtw,
|
||||
|
|
@ -76,10 +76,10 @@ async def main() -> None:
|
|||
|
||||
httpx_tasks = [
|
||||
asyncio.create_task(fawa.scrape()),
|
||||
asyncio.create_task(glisco.scrape()),
|
||||
asyncio.create_task(istreameast.scrape()),
|
||||
asyncio.create_task(ovogoal.scrape()),
|
||||
# asyncio.create_task(pawa.scrape()),
|
||||
asyncio.create_task(s2watch.scrape()),
|
||||
asyncio.create_task(shark.scrape()),
|
||||
asyncio.create_task(streambtw.scrape()),
|
||||
asyncio.create_task(totalsportek1.scrape()),
|
||||
|
|
@ -92,7 +92,7 @@ async def main() -> None:
|
|||
await asyncio.gather(*(pw_tasks + httpx_tasks))
|
||||
|
||||
# others
|
||||
await livetvsx.scrape(xtrnl_brwsr)
|
||||
# await livetvsx.scrape(xtrnl_brwsr)
|
||||
|
||||
finally:
|
||||
await hdl_brwsr.close()
|
||||
|
|
@ -105,13 +105,13 @@ async def main() -> None:
|
|||
cdnlivetv.urls
|
||||
| embedhd.urls
|
||||
| fawa.urls
|
||||
| glisco.urls
|
||||
| istreameast.urls
|
||||
| livetvsx.urls
|
||||
| ovogoal.urls
|
||||
| pawa.urls
|
||||
| ppv.urls
|
||||
| roxie.urls
|
||||
| s2watch.urls
|
||||
| shark.urls
|
||||
| sportzone.urls
|
||||
| streambtw.urls
|
||||
|
|
|
|||
|
|
@ -10,7 +10,7 @@ log = get_logger(__name__)
|
|||
|
||||
urls: dict[str, dict[str, str | float]] = {}
|
||||
|
||||
TAG = "GLISCO"
|
||||
TAG = "S2WATCH"
|
||||
|
||||
CACHE_FILE = Cache(TAG, exp=10_800)
|
||||
|
||||
|
|
@ -79,7 +79,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
|
|||
|
||||
events = []
|
||||
|
||||
start_dt = now.delta(hours=-3)
|
||||
start_dt = now.delta(minutes=-30)
|
||||
end_dt = now.delta(minutes=30)
|
||||
|
||||
for info in api_data:
|
||||
|
|
@ -127,7 +127,7 @@ async def scrape() -> None:
|
|||
|
||||
log.info(f"Loaded {cached_count} event(s) from cache")
|
||||
|
||||
log.info('Scraping from "https://www.glisco.link/"')
|
||||
log.info('Scraping from "https://s2watch.me"')
|
||||
|
||||
if events := await get_events(cached_urls.keys()):
|
||||
log.info(f"Processing {len(events)} new URL(s)")
|
||||
Loading…
Add table
Add a link
Reference in a new issue