diff --git a/M3U8/scrapers/ppv.py b/M3U8/scrapers/ppv.py index 3d9e4c1..08568a8 100644 --- a/M3U8/scrapers/ppv.py +++ b/M3U8/scrapers/ppv.py @@ -101,7 +101,7 @@ async def scrape(browser: Browser) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/sport9.py b/M3U8/scrapers/sport9.py index e402dd1..fc3e446 100644 --- a/M3U8/scrapers/sport9.py +++ b/M3U8/scrapers/sport9.py @@ -106,7 +106,7 @@ async def scrape(browser: Browser) -> None: if events: now = Time.clean(Time.now()).timestamp() - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/streamcenter.py b/M3U8/scrapers/streamcenter.py index 3b93cc5..f2973dd 100644 --- a/M3U8/scrapers/streamcenter.py +++ b/M3U8/scrapers/streamcenter.py @@ -106,7 +106,7 @@ async def scrape(browser: Browser) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/streamhub.py b/M3U8/scrapers/streamhub.py index 8ef7edd..730b786 100644 --- a/M3U8/scrapers/streamhub.py +++ b/M3U8/scrapers/streamhub.py @@ -150,7 +150,7 @@ async def scrape(browser: Browser) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/streamsgate.py b/M3U8/scrapers/streamsgate.py index 9cc3331..10a183d 100644 --- a/M3U8/scrapers/streamsgate.py +++ b/M3U8/scrapers/streamsgate.py @@ -136,7 +136,7 @@ async def scrape(browser: Browser) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/tflix.py b/M3U8/scrapers/tflix.py index bfc875c..d883130 100644 --- a/M3U8/scrapers/tflix.py +++ b/M3U8/scrapers/tflix.py @@ -181,7 +181,7 @@ async def scrape(browser: Browser) -> None: if events: now = Time.clean(Time.now()).timestamp() - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial( diff --git a/M3U8/scrapers/watchfooty.py b/M3U8/scrapers/watchfooty.py index 54c4c14..2c18682 100644 --- a/M3U8/scrapers/watchfooty.py +++ b/M3U8/scrapers/watchfooty.py @@ -257,7 +257,7 @@ async def scrape(browser: Browser) -> None: log.info(f"Processing {len(events)} new URL(s)") if events: - async with network.event_context(browser) as context: + async with network.event_context(browser, stealth=False) as context: for i, ev in enumerate(events, start=1): async with network.event_page(context) as page: handler = partial(