# ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- from __future__ import absolute_import import copy import json import logging import re from urllib.parse import urljoin from urllib.parse import urlparse from svtplay_dl.error import ServiceError from svtplay_dl.fetcher.hls import hlsparse from svtplay_dl.service import OpenGraphThumbMixin from svtplay_dl.service import Service from svtplay_dl.subtitle import subtitle class Urplay(Service, OpenGraphThumbMixin): supported_domains = ["urplay.se", "ur.se", "betaplay.ur.se", "urskola.se"] def get(self): data = self.get_urldata() match = re.search(r"urPlayer.init\((.*)\);", data) if not match: yield ServiceError("Can't find json info") return data = match.group(1) jsondata = json.loads(data) if len(jsondata["subtitles"]) > 0: for sub in jsondata["subtitles"]: if "label" in sub: absurl = urljoin(self.url, sub["file"].split(",")[0]) if absurl.endswith("vtt"): subtype = "wrst" else: subtype = "tt" if self.config.get("get_all_subtitles"): yield subtitle(copy.copy(self.config), subtype, absurl, sub["label"], output=self.output) else: yield subtitle(copy.copy(self.config), subtype, absurl, output=self.output) if "streamer" in jsondata["streaming_config"]: basedomain = jsondata["streaming_config"]["streamer"]["redirect"] else: url = jsondata["streaming_config"]["loadbalancer"] if url[:1] == "/": url = "https:{}".format(url) lbjson = self.http.request("get", url).text lbjson = json.loads(lbjson) basedomain = lbjson["redirect"] http = "https://{}/{}".format(basedomain, jsondata["file_http"]) hd = None if len(jsondata["file_http_hd"]) > 0: http_hd = "https://{}/{}".format(basedomain, jsondata["file_http_hd"]) hls_hd = "{}{}".format(http_hd, jsondata["streaming_config"]["http_streaming"]["hls_file"]) hd = True hls = "{}{}".format(http, jsondata["streaming_config"]["http_streaming"]["hls_file"]) streams = hlsparse(self.config, self.http.request("get", hls), hls, output=self.output) for n in list(streams.keys()): yield streams[n] if hd: streams = hlsparse(self.config, self.http.request("get", hls_hd), hls_hd, output=self.output) for n in list(streams.keys()): yield streams[n] def find_all_episodes(self, config): parse = urlparse(self.url) episodes = [] if parse.netloc == "urskola.se": data = self.get_urldata() match = re.search('data-limit="[^"]+" href="([^"]+)"', data) if match: res = self.http.get(urljoin("https://urskola.se", match.group(1))) data = res.text tags = re.findall('