mirror of
https://github.com/doms9/iptv.git
synced 2026-04-22 19:57:00 +02:00
e
- rename glisco.py to s2watch.py
This commit is contained in:
parent
00000d9da6
commit
00000d956a
2 changed files with 7 additions and 7 deletions
|
|
@ -10,7 +10,7 @@ log = get_logger(__name__)
|
|||
|
||||
urls: dict[str, dict[str, str | float]] = {}
|
||||
|
||||
TAG = "GLISCO"
|
||||
TAG = "S2WATCH"
|
||||
|
||||
CACHE_FILE = Cache(TAG, exp=10_800)
|
||||
|
||||
|
|
@ -79,7 +79,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
|
|||
|
||||
events = []
|
||||
|
||||
start_dt = now.delta(hours=-3)
|
||||
start_dt = now.delta(minutes=-30)
|
||||
end_dt = now.delta(minutes=30)
|
||||
|
||||
for info in api_data:
|
||||
|
|
@ -127,7 +127,7 @@ async def scrape() -> None:
|
|||
|
||||
log.info(f"Loaded {cached_count} event(s) from cache")
|
||||
|
||||
log.info('Scraping from "https://www.glisco.link/"')
|
||||
log.info('Scraping from "https://s2watch.me"')
|
||||
|
||||
if events := await get_events(cached_urls.keys()):
|
||||
log.info(f"Processing {len(events)} new URL(s)")
|
||||
Loading…
Add table
Add a link
Reference in a new issue