e
This commit is contained in:
parent
94096b613c
commit
00000d97e8
3 changed files with 3 additions and 10 deletions
|
|
@ -8,7 +8,6 @@ from scrapers import (
|
||||||
lotus,
|
lotus,
|
||||||
pixel,
|
pixel,
|
||||||
ppv,
|
ppv,
|
||||||
roxie,
|
|
||||||
streambtw,
|
streambtw,
|
||||||
streameast,
|
streameast,
|
||||||
strmd,
|
strmd,
|
||||||
|
|
@ -47,7 +46,6 @@ async def main() -> None:
|
||||||
asyncio.create_task(lotus.scrape(network.client)),
|
asyncio.create_task(lotus.scrape(network.client)),
|
||||||
asyncio.create_task(pixel.scrape(network.client)),
|
asyncio.create_task(pixel.scrape(network.client)),
|
||||||
asyncio.create_task(ppv.scrape(network.client)),
|
asyncio.create_task(ppv.scrape(network.client)),
|
||||||
asyncio.create_task(roxie.scrape(network.client)),
|
|
||||||
asyncio.create_task(streambtw.scrape(network.client)),
|
asyncio.create_task(streambtw.scrape(network.client)),
|
||||||
asyncio.create_task(streameast.scrape(network.client)),
|
asyncio.create_task(streameast.scrape(network.client)),
|
||||||
asyncio.create_task(strmd.scrape(network.client)),
|
asyncio.create_task(strmd.scrape(network.client)),
|
||||||
|
|
@ -63,7 +61,6 @@ async def main() -> None:
|
||||||
| lotus.urls
|
| lotus.urls
|
||||||
| pixel.urls
|
| pixel.urls
|
||||||
| ppv.urls
|
| ppv.urls
|
||||||
| roxie.urls
|
|
||||||
| streambtw.urls
|
| streambtw.urls
|
||||||
| streameast.urls
|
| streameast.urls
|
||||||
| strmd.urls
|
| strmd.urls
|
||||||
|
|
|
||||||
|
|
@ -77,14 +77,11 @@ async def get_events(
|
||||||
|
|
||||||
stream_url = stream["stream_key"]
|
stream_url = stream["stream_key"]
|
||||||
|
|
||||||
stream_logo = stream["thumbnail_url"]
|
|
||||||
|
|
||||||
events.append(
|
events.append(
|
||||||
{
|
{
|
||||||
"sport": sport,
|
"sport": sport,
|
||||||
"event": name,
|
"event": name,
|
||||||
"link": urljoin(url, f"player/{category}/{stream_url}"),
|
"link": urljoin(url, f"player/{category}/{stream_url}"),
|
||||||
"logo": urljoin(url, stream_logo),
|
|
||||||
"timestamp": event_dt.timestamp(),
|
"timestamp": event_dt.timestamp(),
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
@ -130,20 +127,19 @@ async def scrape(client: httpx.AsyncClient) -> None:
|
||||||
)
|
)
|
||||||
|
|
||||||
if url:
|
if url:
|
||||||
sport, event, logo, ts = (
|
sport, event, ts = (
|
||||||
ev["sport"],
|
ev["sport"],
|
||||||
ev["event"],
|
ev["event"],
|
||||||
ev["logo"],
|
|
||||||
ev["timestamp"],
|
ev["timestamp"],
|
||||||
)
|
)
|
||||||
|
|
||||||
key = f"[{sport}] {event} (STRMFR)"
|
key = f"[{sport}] {event} (STRMFR)"
|
||||||
|
|
||||||
tvg_id, pic = leagues.get_tvg_info(sport, event)
|
tvg_id, logo = leagues.get_tvg_info(sport, event)
|
||||||
|
|
||||||
entry = {
|
entry = {
|
||||||
"url": url.replace("540p", "720p"),
|
"url": url.replace("540p", "720p"),
|
||||||
"logo": logo or pic,
|
"logo": logo,
|
||||||
"base": "",
|
"base": "",
|
||||||
"timestamp": ts,
|
"timestamp": ts,
|
||||||
"id": tvg_id or "Live.Event.us",
|
"id": tvg_id or "Live.Event.us",
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue