mirror of
https://github.com/spaam/svtplay-dl.git
synced 2024-11-25 12:45:42 +01:00
baa8d76551
self.get_urldata() is eqivalent to get_http_data(self.url), but also caches the data, so no additional requests are made if it is called multiple times (e.g when grabbing title or downloading thumbnail). Generic().get(url) still causes it to be fetched an extra time.
53 lines
1.9 KiB
Python
53 lines
1.9 KiB
Python
# ex:ts=4:sw=4:sts=4:et
|
|
# -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*-
|
|
from __future__ import absolute_import
|
|
import re
|
|
import sys
|
|
import json
|
|
|
|
from svtplay_dl.service import Service, OpenGraphThumbMixin
|
|
from svtplay_dl.utils import get_http_data, subtitle_tt
|
|
from svtplay_dl.utils.urllib import urlparse
|
|
from svtplay_dl.fetcher.hds import download_hds
|
|
from svtplay_dl.fetcher.hls import download_hls
|
|
from svtplay_dl.log import log
|
|
|
|
class Nrk(Service, OpenGraphThumbMixin):
|
|
supported_domains = ['nrk.no', 'tv.nrk.no']
|
|
|
|
def get(self, options):
|
|
data = self.get_urldata()
|
|
match = re.search(r'data-media="(.*manifest.f4m)"', data)
|
|
if match:
|
|
manifest_url = match.group(1)
|
|
else:
|
|
match = re.search(r'data-video-id="(\d+)"', data)
|
|
if match is None:
|
|
log.error("Can't find video id.")
|
|
sys.exit(2)
|
|
vid = match.group(1)
|
|
match = re.search(r"PS_VIDEO_API_URL : '([^']*)',", data)
|
|
if match is None:
|
|
log.error("Can't find server address with media info")
|
|
sys.exit(2)
|
|
dataurl = "%smediaelement/%s" % (match.group(1), vid)
|
|
data = json.loads(get_http_data(dataurl))
|
|
manifest_url = data["mediaUrl"]
|
|
options.live = data["isLive"]
|
|
if options.hls:
|
|
manifest_url = manifest_url.replace("/z/", "/i/").replace("manifest.f4m", "master.m3u8")
|
|
download_hls(options, manifest_url)
|
|
else:
|
|
manifest_url = "%s?hdcore=2.8.0&g=hejsan" % manifest_url
|
|
download_hds(options, manifest_url)
|
|
|
|
|
|
def get_subtitle(self, options):
|
|
match = re.search("data-subtitlesurl = \"(/.*)\"", self.get_urldata())
|
|
if match:
|
|
parse = urlparse(self.url)
|
|
subtitle = "%s://%s%s" % (parse.scheme, parse.netloc, match.group(1))
|
|
data = get_http_data(subtitle)
|
|
subtitle_tt(options, data)
|
|
|