# ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # pylint has issues with urlparse: "some types could not be inferred" # pylint: disable=E1103 from __future__ import absolute_import import re import json import copy from urllib.parse import urlparse from svtplay_dl.utils.text import filenamify from svtplay_dl.service import Service, OpenGraphThumbMixin from svtplay_dl.fetcher.rtmp import RTMP from svtplay_dl.fetcher.hds import hdsparse from svtplay_dl.fetcher.hls import hlsparse from svtplay_dl.subtitle import subtitle from svtplay_dl.error import ServiceError class Viaplay(Service, OpenGraphThumbMixin): supported_domains = [ 'tv3play.se', 'tv6play.se', 'tv8play.se', 'tv10play.se', 'tv3play.no', 'tv3play.dk', 'tv6play.no', 'viasat4play.no', 'tv3play.ee', 'tv3play.lv', 'tv3play.lt', 'tvplay.lv', 'viagame.com', 'juicyplay.se', 'viafree.se', 'viafree.dk', 'viafree.no', 'viafree.fi', 'play.tv3.lt', 'tv3play.tv3.ee', 'tvplay.skaties.lv' ] def _get_video_id(self, url=None): """ Extract video id. It will try to avoid making an HTTP request if it can find the ID in the URL, but otherwise it will try to scrape it from the HTML document. Returns None in case it's unable to extract the ID at all. """ if url: html_data = self.http.request("get", url).text else: html_data = self.get_urldata() html_data = self.get_urldata() match = re.search(r'data-video-id="([0-9]+)"', html_data) if match: return match.group(1) match = re.search(r'data-videoid="([0-9]+)', html_data) if match: return match.group(1) clips = False slug = None match = re.search('params":({.*}),"query', self.get_urldata()) if match: jansson = json.loads(match.group(1)) if "seasonNumberOrVideoId" in jansson: season = jansson["seasonNumberOrVideoId"] match = re.search("\w-(\d+)$", season) if match: season = match.group(1) else: match = self._conentpage(self.get_urldata()) if match: print("printa") return None if "videoIdOrEpisodeNumber" in jansson: videp = jansson["videoIdOrEpisodeNumber"] match = re.search('(\w+)-(\d+)', videp) if match: episodenr = match.group(2) else: episodenr = videp clips = True match = re.search('(s\w+)-(\d+)', season) if match: season = match.group(2) else: # sometimes videoIdOrEpisodeNumber does not work.. this is a workaround match = re.search('(episode|avsnitt)-(\d+)', self.url) if match: episodenr = match.group(2) else: episodenr = season if "slug" in jansson: slug = jansson["slug"] if clips: return episodenr else: match = self._conentpage(self.get_urldata()) if match: janson = json.loads(match.group(1)) for i in janson["formatPage"]["format"]["videos"].keys(): if "program" in janson["formatPage"]["format"]["videos"][str(i)]: for n in janson["formatPage"]["format"]["videos"][i]["program"]: if str(n["episodeNumber"]) and int(episodenr) == n["episodeNumber"] and int(season) == n["seasonNumber"]: if slug is None or slug == n["formatSlug"]: return n["id"] elif n["id"] == episodenr: return episodenr parse = urlparse(self.url) match = re.search(r'/\w+/(\d+)', parse.path) if match: return match.group(1) match = re.search(r'iframe src="http://play.juicyplay.se[^\"]+id=(\d+)', html_data) if match: return match.group(1) match = re.search(r' 0: return episodes[-config.get("all_last"):] return sorted(episodes) def _grab_episodes(self, config, seasons): episodes = [] baseurl = self.url match = re.search("(saeson|sasong|sesong)-\d+", urlparse(self.url).path) if match: baseurl = self.url[:self.url.rfind("/")] baseurl = baseurl[:baseurl.rfind("/")] for i in seasons: url = "{0}/{1}-{2}".format(baseurl, self._isswe(self.url), i) res = self.http.get(url) if res: match = self._conentpage(res.text) if match: janson = json.loads(match.group(1)) if "program" in janson["formatPage"]["format"]["videos"][str(i)]: for n in janson["formatPage"]["format"]["videos"][str(i)]["program"]: episodes = self._videos_to_list(n["sharingUrl"], n["id"], episodes) if config.get("include_clips"): if "clip" in janson["formatPage"]["format"]["videos"][str(i)]: for n in janson["formatPage"]["format"]["videos"][str(i)]["clip"]: episodes = self._videos_to_list(n["sharingUrl"], n["id"], episodes) return episodes def _isswe(self, url): if re.search(".se$", urlparse(url).netloc): return "sasong" elif re.search(".dk$", urlparse(url).netloc): return "saeson" else: return "sesong" def _conentpage(self, data): return re.search('=({"sportsPlayer.*}); window.__config', data) def _videos_to_list(self, url, vid, episodes): dataj = json.loads(self._get_video_data(vid).text) if "msg" not in dataj: if url not in episodes: episodes.append(url) return episodes def _get_video_data(self, vid): url = "http://playapi.mtgx.tv/v3/videos/{0}".format(vid) data = self.http.request("get", url) return data def _autoname(self, dataj): program = dataj["format_slug"] season = None episode = None title = None if "season" in dataj["format_position"]: if dataj["format_position"]["season"] > 0: season = dataj["format_position"]["season"] if season: if len(dataj["format_position"]["episode"]) > 0: episode = dataj["format_position"]["episode"] if episode: try: episode = int(episode) except TypeError: title = episode episode = None else: title = filenamify(dataj["title"]) if dataj["type"] == "clip": # Removes the show name from the end of the filename # e.g. Showname.S0X.title instead of Showname.S07.title-showname match = re.search(r'(.+)-', dataj["title"]) if match: title = match.group(1) else: title = dataj["title"] if "derived_from_id" in dataj: if dataj["derived_from_id"]: parent_id = dataj["derived_from_id"] parent_episode = self.http.request("get", "http://playapi.mtgx.tv/v3/videos/{0}".format(parent_id)) if parent_episode.status_code != 403: # if not geoblocked datajparent = json.loads(parent_episode.text) if not season and datajparent["format_position"]["season"] > 0: season = datajparent["format_position"]["season"] if len(datajparent["format_position"]["episode"]) > 0: episode = datajparent["format_position"]["episode"] self.output["title"] = program self.output["season"] = int(season) self.output["episode"] = int(episode) self.output["episodename"] = title return True def _sport(self): content = self._conentpage(self.get_urldata()) if not content: yield ServiceError("Can't find video file for: {0}".format(self.url)) return janson = json.loads(content.group(1)) if not janson["sportsPlayer"]["currentVideo"]: yield ServiceError("Can't find video file for: {0}".format(self.url)) return self.output["title"] = janson["sportsPlayer"]["currentVideo"]["title"] res = self.http.request("get", janson["sportsPlayer"]["currentVideo"]["_links"]["streamLink"]["href"]) if res.status_code == 403: yield ServiceError("Can't play this because the video is geoblocked.") return for i in res.json()["embedded"]["prioritizedStreams"]: streams = hlsparse(self.config, self.http.request("get", i["links"]["stream"]["href"]), i["links"]["stream"]["href"], output=self.output) if streams: for n in list(streams.keys()): yield streams[n]