edit typing
move epg fetching into M3U8 folder
edit workflows
This commit is contained in:
doms9 2026-01-24 00:48:30 -05:00
parent 00000d9638
commit 00000d98e3
16 changed files with 376740 additions and 377279 deletions

View file

@ -1,6 +1,6 @@
from functools import partial
from playwright.async_api import BrowserContext
from playwright.async_api import Browser
from .utils import Cache, Time, get_logger, leagues, network
@ -85,7 +85,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
return events
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
cached_urls = CACHE_FILE.load()
cached_count = len(cached_urls)

View file

@ -1,6 +1,6 @@
from functools import partial
from playwright.async_api import BrowserContext
from playwright.async_api import Browser
from .utils import Cache, Time, get_logger, leagues, network
@ -75,7 +75,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
return events
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
cached_urls = CACHE_FILE.load()
cached_count = len(cached_urls)

View file

@ -1,7 +1,7 @@
import json
from functools import partial
from playwright.async_api import BrowserContext, Page
from playwright.async_api import Browser, Page
from .utils import Cache, Time, get_logger, leagues, network
@ -73,7 +73,7 @@ async def get_events(page: Page) -> dict[str, dict[str, str | float]]:
return events
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
if cached := CACHE_FILE.load():
urls.update(cached)

View file

@ -1,6 +1,6 @@
from functools import partial
from playwright.async_api import BrowserContext
from playwright.async_api import Browser
from .utils import Cache, Time, get_logger, leagues, network
@ -78,7 +78,7 @@ async def get_events(url: str, cached_keys: list[str]) -> list[dict[str, str]]:
return events
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
cached_urls = CACHE_FILE.load()
cached_count = len(cached_urls)

View file

@ -2,7 +2,7 @@ import asyncio
from functools import partial
from urllib.parse import urljoin
from playwright.async_api import BrowserContext
from playwright.async_api import Browser
from selectolax.parser import HTMLParser
from .utils import Cache, Time, get_logger, leagues, network
@ -88,7 +88,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
return events
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
cached_urls = CACHE_FILE.load()
cached_count = len(cached_urls)

View file

@ -1,6 +1,6 @@
from functools import partial
from playwright.async_api import BrowserContext
from playwright.async_api import Browser
from .utils import Cache, Time, get_logger, leagues, network
@ -90,7 +90,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
return events
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
cached_urls = CACHE_FILE.load()
cached_count = len(cached_urls)

View file

@ -2,7 +2,7 @@ import asyncio
from functools import partial
from urllib.parse import urljoin
from playwright.async_api import BrowserContext
from playwright.async_api import Browser
from selectolax.parser import HTMLParser
from .utils import Cache, Time, get_logger, leagues, network
@ -132,7 +132,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
return live
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
cached_urls = CACHE_FILE.load()
valid_urls = {k: v for k, v in cached_urls.items() if v["url"]}

View file

@ -4,7 +4,7 @@ from itertools import chain
from typing import Any
from urllib.parse import urljoin
from playwright.async_api import BrowserContext
from playwright.async_api import Browser
from .utils import Cache, Time, get_logger, leagues, network
@ -120,7 +120,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
return events
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
cached_urls = CACHE_FILE.load()
cached_count = len(cached_urls)

View file

@ -3,7 +3,7 @@ from functools import partial
from urllib.parse import urljoin
import feedparser
from playwright.async_api import BrowserContext, Error, Page, TimeoutError
from playwright.async_api import Browser, Error, Page, TimeoutError
from .utils import Cache, Time, get_logger, leagues, network
@ -161,7 +161,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
return events
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
cached_urls = CACHE_FILE.load()
valid_urls = {k: v for k, v in cached_urls.items() if v["url"]}

View file

@ -5,7 +5,7 @@ from itertools import chain
from typing import Any
from urllib.parse import urljoin
from playwright.async_api import BrowserContext, Page, TimeoutError
from playwright.async_api import Browser, Page, TimeoutError
from .utils import Cache, Time, get_logger, leagues, network
@ -232,7 +232,7 @@ async def get_events(base_url: str, cached_keys: list[str]) -> list[dict[str, st
return events
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
cached_urls = CACHE_FILE.load()
valid_urls = {k: v for k, v in cached_urls.items() if v["url"]}

View file

@ -1,7 +1,7 @@
import asyncio
from functools import partial
from playwright.async_api import BrowserContext
from playwright.async_api import Browser
from selectolax.parser import HTMLParser
from .utils import Cache, Time, get_logger, leagues, network
@ -110,7 +110,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
return live
async def scrape(browser: BrowserContext) -> None:
async def scrape(browser: Browser) -> None:
cached_urls = CACHE_FILE.load()
cached_count = len(cached_urls)