2013-03-02 21:26:28 +01:00
|
|
|
# ex:ts=4:sw=4:sts=4:et
|
|
|
|
# -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*-
|
2013-03-01 23:39:42 +01:00
|
|
|
from __future__ import absolute_import
|
2016-06-29 23:58:37 +02:00
|
|
|
import base64
|
2013-02-12 19:43:37 +01:00
|
|
|
import re
|
|
|
|
import json
|
2014-06-07 20:43:40 +02:00
|
|
|
import copy
|
2019-03-24 21:06:23 +01:00
|
|
|
import binascii
|
|
|
|
import hashlib
|
2018-01-30 22:07:21 +01:00
|
|
|
from urllib.parse import urljoin, urlparse
|
2013-02-12 19:43:37 +01:00
|
|
|
|
2019-03-24 21:06:23 +01:00
|
|
|
from cryptography.hazmat.primitives.ciphers import Cipher, algorithms, modes
|
|
|
|
from cryptography.hazmat.backends import default_backend
|
|
|
|
|
2014-01-26 01:51:53 +01:00
|
|
|
from svtplay_dl.service import Service, OpenGraphThumbMixin
|
2015-10-04 14:37:16 +02:00
|
|
|
from svtplay_dl.fetcher.hls import hlsparse
|
2014-06-02 21:43:22 +02:00
|
|
|
from svtplay_dl.fetcher.hds import hdsparse
|
2015-03-01 21:46:22 +01:00
|
|
|
from svtplay_dl.subtitle import subtitle
|
2015-09-06 14:19:10 +02:00
|
|
|
from svtplay_dl.error import ServiceError
|
2015-09-15 20:10:32 +02:00
|
|
|
|
2018-01-30 20:11:37 +01:00
|
|
|
|
2014-01-26 01:51:53 +01:00
|
|
|
class Dr(Service, OpenGraphThumbMixin):
|
2019-08-25 00:27:31 +02:00
|
|
|
supported_domains = ["dr.dk"]
|
2013-01-17 00:21:47 +01:00
|
|
|
|
2015-12-26 11:46:14 +01:00
|
|
|
def get(self):
|
2015-08-30 00:06:20 +02:00
|
|
|
data = self.get_urldata()
|
2014-12-22 17:41:40 +01:00
|
|
|
|
2013-01-17 00:21:47 +01:00
|
|
|
match = re.search(r'resource:[ ]*"([^"]*)",', data)
|
2013-09-14 22:39:37 +02:00
|
|
|
if match:
|
|
|
|
resource_url = match.group(1)
|
2015-08-31 19:45:15 +02:00
|
|
|
resource_data = self.http.request("get", resource_url).content
|
2013-09-14 22:39:37 +02:00
|
|
|
resource = json.loads(resource_data)
|
2018-05-22 00:02:20 +02:00
|
|
|
streams = self.find_stream(self.config, resource)
|
2019-08-25 00:33:51 +02:00
|
|
|
yield from streams
|
2013-09-14 21:58:55 +02:00
|
|
|
else:
|
2013-09-14 22:39:37 +02:00
|
|
|
match = re.search(r'resource="([^"]*)"', data)
|
|
|
|
if not match:
|
2015-09-06 14:19:10 +02:00
|
|
|
yield ServiceError("Cant find resource info for this video")
|
2014-10-06 23:21:43 +02:00
|
|
|
return
|
2015-10-08 09:52:36 +02:00
|
|
|
if match.group(1)[:4] != "http":
|
2019-08-25 00:33:51 +02:00
|
|
|
resource_url = "http:{}".format(match.group(1))
|
2015-10-08 09:52:36 +02:00
|
|
|
else:
|
|
|
|
resource_url = match.group(1)
|
2015-09-01 00:37:06 +02:00
|
|
|
resource_data = self.http.request("get", resource_url).text
|
2013-09-14 22:39:37 +02:00
|
|
|
resource = json.loads(resource_data)
|
2013-09-14 21:58:55 +02:00
|
|
|
|
2015-10-08 09:54:07 +02:00
|
|
|
if "Links" not in resource:
|
|
|
|
yield ServiceError("Cant access this video. its geoblocked.")
|
|
|
|
return
|
2015-11-15 15:21:42 +01:00
|
|
|
if "SubtitlesList" in resource and len(resource["SubtitlesList"]) > 0:
|
2015-03-01 21:46:22 +01:00
|
|
|
suburl = resource["SubtitlesList"][0]["Uri"]
|
2018-06-03 01:18:15 +02:00
|
|
|
yield subtitle(copy.copy(self.config), "wrst", suburl, output=self.output)
|
2014-08-11 19:46:56 +02:00
|
|
|
if "Data" in resource:
|
2018-05-13 13:06:45 +02:00
|
|
|
streams = self.find_stream(self.config, resource)
|
2019-08-25 00:33:51 +02:00
|
|
|
yield from streams
|
2014-08-11 19:46:56 +02:00
|
|
|
else:
|
2019-08-25 00:27:31 +02:00
|
|
|
for stream in resource["Links"]:
|
2019-03-24 21:06:23 +01:00
|
|
|
uri = stream["Uri"]
|
|
|
|
if uri is None:
|
|
|
|
uri = self._decrypt(stream["EncryptedUri"])
|
|
|
|
|
2014-08-11 19:46:56 +02:00
|
|
|
if stream["Target"] == "HDS":
|
2019-08-25 00:27:31 +02:00
|
|
|
streams = hdsparse(copy.copy(self.config), self.http.request("get", uri, params={"hdcore": "3.7.0"}), uri, output=self.output)
|
2014-10-12 23:31:02 +02:00
|
|
|
if streams:
|
|
|
|
for n in list(streams.keys()):
|
|
|
|
yield streams[n]
|
2014-08-11 19:46:56 +02:00
|
|
|
if stream["Target"] == "HLS":
|
2019-03-24 21:06:23 +01:00
|
|
|
streams = hlsparse(self.config, self.http.request("get", uri), uri, output=self.output)
|
2014-08-11 19:46:56 +02:00
|
|
|
for n in list(streams.keys()):
|
2015-10-04 14:37:16 +02:00
|
|
|
yield streams[n]
|
2014-08-11 19:46:56 +02:00
|
|
|
|
2018-05-22 00:02:20 +02:00
|
|
|
def find_all_episodes(self, config):
|
2016-06-29 23:58:37 +02:00
|
|
|
episodes = []
|
2019-08-25 00:27:31 +02:00
|
|
|
matches = re.findall(r'<button class="show-more" data-url="([^"]+)" data-partial="([^"]+)"', self.get_urldata())
|
2016-06-29 23:58:37 +02:00
|
|
|
for encpath, enccomp in matches:
|
2019-08-25 00:27:31 +02:00
|
|
|
newstyle = "_" in encpath
|
2016-06-29 23:58:37 +02:00
|
|
|
if newstyle:
|
2019-08-25 00:27:31 +02:00
|
|
|
encbasepath = encpath.split("_")[0]
|
|
|
|
path = base64.b64decode(encbasepath + "===").decode("latin1")
|
2016-06-29 23:58:37 +02:00
|
|
|
else:
|
2019-08-25 00:27:31 +02:00
|
|
|
path = base64.b64decode(encpath + "===").decode("latin1")
|
2016-06-29 23:58:37 +02:00
|
|
|
|
2019-08-25 00:27:31 +02:00
|
|
|
if "/view/" in path:
|
2016-06-29 23:58:37 +02:00
|
|
|
continue
|
|
|
|
|
2019-08-25 00:27:31 +02:00
|
|
|
params = "offset=0&limit=1000"
|
2016-06-29 23:58:37 +02:00
|
|
|
if newstyle:
|
2019-08-25 00:27:31 +02:00
|
|
|
encparams = base64.b64encode(params.encode("latin1")).decode("latin1").rstrip("=")
|
2019-08-25 00:33:51 +02:00
|
|
|
encpath = "{}_{}".format(encbasepath, encparams)
|
2016-06-29 23:58:37 +02:00
|
|
|
else:
|
2019-08-25 00:33:51 +02:00
|
|
|
path = "{}?{}".format(urlparse(path).path, params)
|
2019-08-25 00:27:31 +02:00
|
|
|
encpath = base64.b64encode(path.encode("latin1")).decode("latin1").rstrip("=")
|
2016-06-29 23:58:37 +02:00
|
|
|
|
2019-08-25 00:33:51 +02:00
|
|
|
url = urljoin("https://www.dr.dk/tv/partial/", "{}/{}".format(enccomp, encpath))
|
2019-08-25 00:27:31 +02:00
|
|
|
data = self.http.request("get", url).content.decode("latin1")
|
2016-06-29 23:58:37 +02:00
|
|
|
|
|
|
|
matches = re.findall(r'"program-link" href="([^"]+)">', data)
|
2019-08-25 00:27:31 +02:00
|
|
|
episodes = [urljoin("https://www.dr.dk/", url) for url in matches]
|
2016-06-29 23:58:37 +02:00
|
|
|
break
|
|
|
|
|
|
|
|
if not episodes:
|
2019-08-25 00:27:31 +02:00
|
|
|
prefix = "/".join(urlparse(self.url).path.rstrip("/").split("/")[:-1])
|
2016-06-29 23:58:37 +02:00
|
|
|
matches = re.findall(r'"program-link" href="([^"]+)">', self.get_urldata())
|
2019-08-25 00:27:31 +02:00
|
|
|
episodes = [urljoin("https://www.dr.dk/", url) for url in matches if url.startswith(prefix)]
|
2016-06-29 23:58:37 +02:00
|
|
|
|
2018-05-22 00:02:20 +02:00
|
|
|
if config.get("all_last") != -1:
|
2019-08-25 00:27:31 +02:00
|
|
|
episodes = episodes[: config.get("all_last")]
|
2016-06-29 23:58:37 +02:00
|
|
|
else:
|
|
|
|
episodes.reverse()
|
|
|
|
|
|
|
|
return episodes
|
|
|
|
|
2018-05-13 13:06:45 +02:00
|
|
|
def find_stream(self, config, resource):
|
2019-08-25 00:27:31 +02:00
|
|
|
tempresource = resource["Data"][0]["Assets"]
|
2015-09-01 00:37:06 +02:00
|
|
|
# To find the VideoResource, they have Images as well
|
|
|
|
for resources in tempresource:
|
2019-08-25 00:27:31 +02:00
|
|
|
if resources["Kind"] == "VideoResource":
|
|
|
|
links = resources["Links"]
|
2015-09-01 00:37:06 +02:00
|
|
|
break
|
|
|
|
for i in links:
|
|
|
|
if i["Target"] == "Ios" or i["Target"] == "HLS":
|
2018-05-21 00:56:22 +02:00
|
|
|
streams = hlsparse(config, self.http.request("get", i["Uri"]), i["Uri"], output=self.output)
|
2015-09-01 00:37:06 +02:00
|
|
|
for n in list(streams.keys()):
|
2015-10-04 14:37:16 +02:00
|
|
|
yield streams[n]
|
2019-03-24 21:06:23 +01:00
|
|
|
|
|
|
|
def _decrypt(self, url):
|
|
|
|
n = int(url[2:10], 16)
|
2019-08-25 00:27:31 +02:00
|
|
|
iv_hex = url[10 + n :]
|
|
|
|
data = binascii.a2b_hex(url[10 : 10 + n].encode("ascii"))
|
|
|
|
key = hashlib.sha256(("%s:sRBzYNXBzkKgnjj8pGtkACch" % iv_hex).encode("utf-8")).digest()
|
2019-03-24 21:06:23 +01:00
|
|
|
iv = bytes.fromhex(iv_hex)
|
|
|
|
|
|
|
|
backend = default_backend()
|
|
|
|
cipher = Cipher(algorithms.AES(key), modes.CBC(iv), backend=backend)
|
|
|
|
decryptor = cipher.decryptor()
|
|
|
|
decrypted = decryptor.update(data)
|
2019-08-25 00:27:31 +02:00
|
|
|
return decrypted[: -decrypted[-1]].decode("utf-8")
|