Compare commits

..

No commits in common. "c901ed52e8aa0e88fc926cfaf3a8d7bad5d544a8" and "8c544f3a6de1c3d53cd41e24bb3d00d9f97a1f95" have entirely different histories.

8 changed files with 46859 additions and 47358 deletions

File diff suppressed because it is too large Load diff

86792
M3U8/TV.xml

File diff suppressed because one or more lines are too long

View file

@ -43,7 +43,7 @@ http://212.102.60.231/BET/index.m3u8
http://23.237.104.106:8080/USA_BTN/index.m3u8
#EXTINF:-1 tvg-chno="15" tvg-id="Bloomberg.HD.us2" tvg-name="Bloomberg TV" tvg-logo="http://schedulesdirect-api20141201-logos.s3.dualstack.us-east-1.amazonaws.com/stationLogos/s71799_dark_360w_270h.png" group-title="TV",Bloomberg TV
https://live-manifest.production-public.tubi.io/live/a5aabf5a-9615-4987-8163-18edfa1360c6/playlist.m3u8
http://41.205.93.154/BLOOMBERG/index.m3u8
#EXTINF:-1 tvg-chno="16" tvg-id="Boomerang.us2" tvg-name="Boomerang" tvg-logo="http://schedulesdirect-api20141201-logos.s3.dualstack.us-east-1.amazonaws.com/stationLogos/s21883_dark_360w_270h.png" group-title="TV",Boomerang
http://23.237.104.106:8080/USA_BOOMERANG/index.m3u8

File diff suppressed because it is too large Load diff

View file

@ -72,7 +72,7 @@ async def main() -> None:
asyncio.create_task(roxie.scrape(hdl_brwsr)),
asyncio.create_task(sportzone.scrape(xtrnl_brwsr)),
asyncio.create_task(streamcenter.scrape(hdl_brwsr)),
asyncio.create_task(streamhub.scrape(xtrnl_brwsr)),
# asyncio.create_task(streamhub.scrape(xtrnl_brwsr)),
asyncio.create_task(streamsgate.scrape(xtrnl_brwsr)),
# asyncio.create_task(timstreams.scrape(xtrnl_brwsr)),
]

View file

@ -17,19 +17,19 @@ CACHE_FILE = Cache(TAG, exp=10_800)
HTML_CACHE = Cache(f"{TAG}-html", exp=19_800)
BASE_URL = "https://streamhub.pro"
BASE_URL = "https://livesports4u.net"
SPORT_ENDPOINTS = [
f"sport_{sport_id}"
for sport_id in [
# "68c02a4465113", # American Football
# "68c02a446582f", # Baseball
"68c02a446582f", # Baseball
"68c02a4466011", # Basketball
"68c02a4466f56", # Hockey
# "68c02a44674e9", # MMA
# "68c02a4467a48", # Racing
"68c02a44674e9", # MMA
"68c02a4467a48", # Racing
"68c02a4464a38", # Soccer
# "68c02a4468cf7", # Tennis
"68c02a4468cf7", # Tennis
]
]
@ -107,7 +107,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
now.timestamp(),
)
for date in [now.date(), now.delta(days=1).date()]
for sport_id in SPORT_ENDPOINTS
for sport_id in SPORT_ENDPOINTS.values()
]
results = await asyncio.gather(*tasks)
@ -193,8 +193,6 @@ async def scrape(browser: Browser) -> None:
if url:
valid_count += 1
entry["url"] = url.split("?")[0]
urls[key] = entry
log.info(f"Collected and cached {valid_count - cached_count} new event(s)")

View file

@ -19,9 +19,13 @@ CACHE_FILE = Cache(TAG, exp=10_800)
API_FILE = Cache(f"{TAG}-api", exp=19_800)
BASE_DOMAIN = "watchfooty.pw"
API_URL = "https://api.watchfooty.st"
API_URL, BASE_URL = f"https://api.{BASE_DOMAIN}", f"https://www.{BASE_DOMAIN}"
BASE_MIRRORS = [
"https://www.watchfooty.top",
"https://www.watchfooty.st",
"https://www.watchfooty.su",
]
VALID_SPORTS = [
# "american-football",
@ -119,7 +123,7 @@ async def process_event(
embed = re.sub(
pattern=r"^.*\/stream",
repl="https://pikachusports.top/embed",
repl="https://spiderembed.top/embed",
string=stream_url,
)
@ -165,7 +169,7 @@ async def process_event(
page.remove_listener("request", handler)
async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
async def get_events(base_url: str, cached_keys: list[str]) -> list[dict[str, str]]:
now = Time.clean(Time.now())
if not (api_data := API_FILE.load(per_entry=False, index=-1)):
@ -214,7 +218,7 @@ async def get_events(cached_keys: list[str]) -> list[dict[str, str]]:
{
"sport": sport,
"event": name,
"link": urljoin(BASE_URL, f"stream/{match_id}"),
"link": urljoin(base_url, f"stream/{match_id}"),
"match-id": match_id,
"logo": logo,
"timestamp": event_dt.timestamp(),
@ -235,9 +239,16 @@ async def scrape(browser: Browser) -> None:
log.info(f"Loaded {cached_count} event(s) from cache")
log.info(f'Scraping from "{BASE_URL}"')
if not (base_url := await network.get_base(BASE_MIRRORS)):
log.warning("No working Watch Footy mirrors")
if events := await get_events(cached_urls.keys()):
CACHE_FILE.write(cached_urls)
return
log.info(f'Scraping from "{base_url}"')
if events := await get_events(base_url, cached_urls.keys()):
log.info(f"Processing {len(events)} new URL(s)")
async with network.event_context(browser, stealth=False) as context:

View file

@ -1,12 +1,12 @@
## Base Log @ 2026-03-08 08:53 UTC
## Base Log @ 2026-03-07 08:53 UTC
### ✅ Working Streams: 158<br>❌ Dead Streams: 3
| Channel | Error (Code) | Link |
| ------- | ------------ | ---- |
| MSNBC | HTTP Error (404) | `http://41.205.93.154/MSNBC/index.m3u8` |
| NBC Sports Bay Area | HTTP Error (403) | `http://mytvstream.net:8080/live/30550113/30550113/9900.m3u8` |
| Spectrum SportsNet Lakers | HTTP Error (403) | `http://mytvstream.net:8080/live/30550113/30550113/21842.m3u8` |
| Bloomberg TV | HTTP Error (404) | `http://41.205.93.154/BLOOMBERG/index.m3u8` |
| Discovery Science | HTTP Error (403) | `http://mytvstream.net:8080/live/30550113/30550113/20866.m3u8` |
| Showtime Extreme | HTTP Error (403) | `http://mytvstream.net:8080/live/30550113/30550113/13220.m3u8` |
---
#### Base Channels URL
```