This commit is contained in:
doms9 2025-12-19 17:58:37 -05:00
parent 3780698908
commit 00000d9924
9 changed files with 15 additions and 20 deletions

View file

@ -108,7 +108,7 @@ async def scrape() -> None:
if events: if events:
async with async_playwright() as p: async with async_playwright() as p:
browser, context = await network.browser(p, browser="brave") browser, context = await network.browser(p, browser="external")
for i, ev in enumerate(events, start=1): for i, ev in enumerate(events, start=1):
handler = partial( handler = partial(

View file

@ -107,7 +107,7 @@ async def scrape() -> None:
now = Time.clean(Time.now()).timestamp() now = Time.clean(Time.now()).timestamp()
async with async_playwright() as p: async with async_playwright() as p:
browser, context = await network.browser(p, browser="brave") browser, context = await network.browser(p, browser="external")
for i, ev in enumerate(events, start=1): for i, ev in enumerate(events, start=1):
handler = partial( handler = partial(

View file

@ -105,7 +105,7 @@ async def scrape() -> None:
if events: if events:
async with async_playwright() as p: async with async_playwright() as p:
browser, context = await network.browser(p, browser="brave") browser, context = await network.browser(p, browser="external")
for i, ev in enumerate(events, start=1): for i, ev in enumerate(events, start=1):
handler = partial( handler = partial(

View file

@ -149,7 +149,7 @@ async def scrape() -> None:
if events: if events:
async with async_playwright() as p: async with async_playwright() as p:
browser, context = await network.browser(p, browser="brave") browser, context = await network.browser(p, browser="external")
for i, ev in enumerate(events, start=1): for i, ev in enumerate(events, start=1):
handler = partial( handler = partial(

View file

@ -140,7 +140,7 @@ async def scrape() -> None:
if events: if events:
async with async_playwright() as p: async with async_playwright() as p:
browser, context = await network.browser(p, browser="brave") browser, context = await network.browser(p, browser="external")
for i, ev in enumerate(events, start=1): for i, ev in enumerate(events, start=1):
handler = partial( handler = partial(

View file

@ -137,7 +137,7 @@ async def scrape() -> None:
if events: if events:
async with async_playwright() as p: async with async_playwright() as p:
browser, context = await network.browser(p, browser="brave") browser, context = await network.browser(p, browser="external")
for i, ev in enumerate(events, start=1): for i, ev in enumerate(events, start=1):
handler = partial( handler = partial(

View file

@ -6,7 +6,7 @@ log = get_logger(__name__)
urls: dict[str, dict[str, str | float]] = {} urls: dict[str, dict[str, str | float]] = {}
TAG = "TVP" TAG = "TVPASS"
CACHE_FILE = Cache(f"{TAG.lower()}.json", exp=86_400) CACHE_FILE = Cache(f"{TAG.lower()}.json", exp=86_400)

View file

@ -18,14 +18,14 @@ T = TypeVar("T")
class Network: class Network:
proxy_base = "https://stream.nvrmind.xyz"
UA = ( UA = (
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) " "Mozilla/5.0 (Windows NT 10.0; Win64; x64) "
"AppleWebKit/537.36 (KHTML, like Gecko) " "AppleWebKit/537.36 (KHTML, like Gecko) "
"Chrome/134.0.0.0 Safari/537.36 Edg/134.0.0.0" "Chrome/134.0.0.0 Safari/537.36 Edg/134.0.0.0"
) )
proxy_base = "https://stream.nvrmind.xyz"
def __init__(self) -> None: def __init__(self) -> None:
self.client = httpx.AsyncClient( self.client = httpx.AsyncClient(
timeout=5, timeout=5,
@ -41,14 +41,12 @@ class Network:
query: dict | None = None, query: dict | None = None,
) -> str: ) -> str:
base = network.proxy_base
tag = tag.lower() tag = tag.lower()
return ( return (
f"{urljoin(base, f'{tag}/{path}')}?{urlencode(query)}" f"{urljoin(network.proxy_base, f'{tag}/{path}')}?{urlencode(query)}"
if query if query
else urljoin(base, f"{tag}/{path}") else urljoin(network.proxy_base, f"{tag}/{path}")
) )
async def request( async def request(
@ -207,12 +205,9 @@ class Network:
@staticmethod @staticmethod
async def browser( async def browser(
playwright: Playwright, playwright: Playwright, browser: str = "internal"
browser: str = "firefox",
ignore_https_errors: bool = False,
) -> tuple[Browser, BrowserContext]: ) -> tuple[Browser, BrowserContext]:
if browser == "external":
if browser == "brave":
brwsr = await playwright.chromium.connect_over_cdp("http://localhost:9222") brwsr = await playwright.chromium.connect_over_cdp("http://localhost:9222")
context = brwsr.contexts[0] context = brwsr.contexts[0]
@ -222,7 +217,7 @@ class Network:
context = await brwsr.new_context( context = await brwsr.new_context(
user_agent=Network.UA, user_agent=Network.UA,
ignore_https_errors=ignore_https_errors, ignore_https_errors=False,
viewport={"width": 1366, "height": 768}, viewport={"width": 1366, "height": 768},
device_scale_factor=1, device_scale_factor=1,
locale="en-US", locale="en-US",

View file

@ -261,7 +261,7 @@ async def scrape() -> None:
if events: if events:
async with async_playwright() as p: async with async_playwright() as p:
browser, context = await network.browser(p, browser="brave") browser, context = await network.browser(p, browser="external")
for i, ev in enumerate(events, start=1): for i, ev in enumerate(events, start=1):
handler = partial( handler = partial(