From 00000d9a013d09414f83075e2364e8b68a81e3a0 Mon Sep 17 00:00:00 2001 From: doms9 <96013514+doms9@users.noreply.github.com> Date: Sat, 24 Jan 2026 11:50:43 -0500 Subject: [PATCH] e fix contexts --- M3U8/scrapers/ppv.py | 2 +- M3U8/scrapers/sport9.py | 2 +- M3U8/scrapers/streamcenter.py | 2 +- M3U8/scrapers/streamhub.py | 2 +- M3U8/scrapers/streamsgate.py | 2 +- M3U8/scrapers/tflix.py | 2 +- M3U8/scrapers/watchfooty.py | 2 +- 7 files changed, 7 insertions(+), 7 deletions(-) diff --git a/M3U8/scrapers/ppv.py b/M3U8/scrapers/ppv.py index 3d9e4c15..08568a8a 100644 --- a/M3U8/scrapers/ppv.py +++ b/M3U8/scrapers/ppv.py @@ -101,7 +101,7 @@ async def scrape(browser: Browser) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/sport9.py b/M3U8/scrapers/sport9.py index e402dd12..fc3e4466 100644 --- a/M3U8/scrapers/sport9.py +++ b/M3U8/scrapers/sport9.py @@ -106,7 +106,7 @@ async def scrape(browser: Browser) -> None: if events: now = Time.clean(Time.now()).timestamp() - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/streamcenter.py b/M3U8/scrapers/streamcenter.py index 3b93cc5e..f2973dd1 100644 --- a/M3U8/scrapers/streamcenter.py +++ b/M3U8/scrapers/streamcenter.py @@ -106,7 +106,7 @@ async def scrape(browser: Browser) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/streamhub.py b/M3U8/scrapers/streamhub.py index 8ef7eddb..730b7863 100644 --- a/M3U8/scrapers/streamhub.py +++ b/M3U8/scrapers/streamhub.py @@ -150,7 +150,7 @@ async def scrape(browser: Browser) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/streamsgate.py b/M3U8/scrapers/streamsgate.py index 9cc3331a..10a183dd 100644 --- a/M3U8/scrapers/streamsgate.py +++ b/M3U8/scrapers/streamsgate.py @@ -136,7 +136,7 @@ async def scrape(browser: Browser) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/tflix.py b/M3U8/scrapers/tflix.py index bfc875c7..d883130c 100644 --- a/M3U8/scrapers/tflix.py +++ b/M3U8/scrapers/tflix.py @@ -181,7 +181,7 @@ async def scrape(browser: Browser) -> None: if events: now = Time.clean(Time.now()).timestamp() - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/watchfooty.py b/M3U8/scrapers/watchfooty.py index 54c4c140..2c18682a 100644 --- a/M3U8/scrapers/watchfooty.py +++ b/M3U8/scrapers/watchfooty.py @@ -257,7 +257,7 @@ async def scrape(browser: Browser) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial(