New
This commit is contained in:
parent
98dd671f16
commit
4297654052
2
DOWNLOAD.cmd
Normal file
2
DOWNLOAD.cmd
Normal file
@ -0,0 +1,2 @@
|
||||
python bad37.py --url https://www.paramountplus.com/shows/mayor-of-kingstown/ -s 1 -e 3 --alang es-la en --slang es-la en --flang es-la
|
||||
pause
|
1
KEYS/PARAMOUNTPLUS.txt
Normal file
1
KEYS/PARAMOUNTPLUS.txt
Normal file
@ -0,0 +1 @@
|
||||
##### One KEY per line. #####
|
40
README.md
40
README.md
@ -1,2 +1,38 @@
|
||||
# Paramount Plus 4k Downloader
|
||||
Tool To Get Downloads up to 4k from Paramount+
|
||||
|
||||
<div size='20px'> Paramount 4K Downloader
|
||||
</div>
|
||||
|
||||
<div size='20px'> Tool To Get Downloads up to 4k from Paramount+ :smile:
|
||||
</div>
|
||||
|
||||
<p align="center">
|
||||
<img width="200" src="https://github.com/Kathryn-Jie/Kathryn-Jie/blob/main/kathryn.png">
|
||||
</p>
|
||||
|
||||
<h1> Hello Fellow < Developers/ >! <img src = "https://raw.githubusercontent.com/MartinHeinz/MartinHeinz/master/wave.gif" width = 30px> </h1>
|
||||
<p align='center'>
|
||||
</p>
|
||||
|
||||
|
||||
|
||||
<div size='20px'> Hi! My name is WVDUMP. I am Leaking the scripts to punish few idiots :smile:
|
||||
</div>
|
||||
|
||||
<h2> About Me <img src = "https://media0.giphy.com/media/KDDpcKigbfFpnejZs6/giphy.gif?cid=ecf05e47oy6f4zjs8g1qoiystc56cu7r9tb8a1fe76e05oty&rid=giphy.gif" width = 100px></h2>
|
||||
|
||||
<img width="55%" align="right" alt="Github" src="https://raw.githubusercontent.com/onimur/.github/master/.resources/git-header.svg" />
|
||||
|
||||
- 🔭 I’m currently working on Java scripts
|
||||
|
||||
- 🌱 I’m currently learning Python
|
||||
|
||||
- 👯 Sharing is caring
|
||||
|
||||
|
||||
- ⚡ CDM IS NOT INCLUDED BUY it from wvfuck@cyberfiends.net ⚡
|
||||
|
||||
|
||||
<br>
|
||||
<br>
|
||||
<br>
|
||||
|
||||
|
128
bad37.py
Normal file
128
bad37.py
Normal file
@ -0,0 +1,128 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
# Module: BAD Project
|
||||
# Created on: 01-06-2021
|
||||
# Authors: JUNi
|
||||
# Version: 1.0
|
||||
|
||||
import argparse
|
||||
import os
|
||||
import sys
|
||||
|
||||
|
||||
parser = argparse.ArgumentParser()
|
||||
#Common:
|
||||
parser.add_argument("content", nargs="?", help="Content URL or ID")
|
||||
parser.add_argument("--url", dest="url_season", help="If set, it will download all assets from the season provided.")
|
||||
parser.add_argument("--tqdm", dest="tqmd_mode", help="If set, will download with threading", action="store_true")
|
||||
parser.add_argument("--nv", "--no-video", dest="novideo", help="If set, don't download video", action="store_true")
|
||||
parser.add_argument("--na", "--no-audio", dest="noaudio", help="If set, don't download audio", action="store_true")
|
||||
parser.add_argument("--ns", "--no-subs", dest="nosubs", help="If set, don't download subs", action="store_true")
|
||||
parser.add_argument("--all-season", dest="all_season", help="If set, active download mode.", action="store_true")
|
||||
parser.add_argument("-e", "--episode", dest="episodeStart", help="If set, it will start downloading the season from that episode.")
|
||||
parser.add_argument("-s", dest="season", help="If set, it will download all assets from the season provided.")
|
||||
parser.add_argument("--tag", type=str, required=False, help="Release group tag to use for filenames")
|
||||
parser.add_argument("-q", "--quality", dest="customquality", type=lambda x: [x.rstrip('p')], help="For configure quality of video.", default=[])
|
||||
parser.add_argument("-o", "--output", dest="output", default="downloads", help="If set, it will download all assets to directory provided.")
|
||||
parser.add_argument("--keep", dest="keep", help="If set, it will list all formats available.", action="store_true")
|
||||
parser.add_argument("--info", help="If set, it will print manifest infos and exit.", action="store_true")
|
||||
parser.add_argument("--no-mux", dest="nomux", help="If set, dont mux.", action="store_true")
|
||||
#parser.add_argument("--force-mux", dest="force_mux", nargs=1, help="If set, force mux.", default=[])
|
||||
#parser.add_argument("--langtag", dest="langtag", nargs=1, help="For configure language tag of MKV.", default=[])
|
||||
parser.add_argument("--only-2ch-audio", dest="only_2ch_audio", help="If set, no clean tag subtitles.", action="store_true")
|
||||
parser.add_argument("--alang", "--audio-language", dest="audiolang", nargs="*", help="If set, download only selected audio languages", default=[])
|
||||
parser.add_argument("--slang", "--subtitle-language", dest="sublang", nargs="*", help="If set, download only selected subtitle languages", default=[])
|
||||
parser.add_argument("--flang", "--forced-language", dest="forcedlang", nargs="*", help="If set, download only selected forced subtitle languages", default=[])
|
||||
parser.add_argument("--no-cleansubs", dest="nocleansubs", help="If set, no clean tag subtitles.", action="store_true")
|
||||
parser.add_argument("--hevc", dest="hevc", help="If set, it will return HEVC manifest", action="store_true")
|
||||
parser.add_argument("--uhd", dest="uhd", help="If set, it will return UHD manifest", action="store_true")
|
||||
parser.add_argument("--license", dest="license", help="Only print keys, don't download", action="store_true")
|
||||
parser.add_argument("-licenses-as-json", help="Save the wv keys as json instead", action="store_true")
|
||||
parser.add_argument("--debug", action="store_true", help="Enable debug logging")
|
||||
parser.add_argument("--aformat-51ch", "--audio-format-51ch", dest="aformat_51ch", help="For configure format of audio.")
|
||||
parser.add_argument("--nc", "--no-chapters", dest="nochpaters", help="If set, don't download chapters", action="store_true")
|
||||
parser.add_argument("-c", "--codec", choices=["widevine", "playready"], help="Video type to download", default="playready")
|
||||
parser.add_argument("--ap", dest="audiocodec", default="atmos", choices=["aac", "ac3", "atmos"], help="audio codec profile")
|
||||
|
||||
#HBOMAX
|
||||
parser.add_argument("--atmos", dest="atmos", help="If set, it will return Atmos MPDs", action="store_true")
|
||||
parser.add_argument("--ad", "--desc-audio", action="store_true", dest="desc_audio", help="Download descriptive audio instead of normal dialogue")
|
||||
parser.add_argument("--hdr", dest="hdr", help="If set, it will return HDR manifest", action="store_true")
|
||||
parser.add_argument("-r", "--region", choices=["la", "us"], help="HBO Max video region", default="us")
|
||||
parser.add_argument("--vp", dest="videocodec", default="h264", choices=["h264", "hevc", "hdr"], help="video codec profile")
|
||||
|
||||
#Clarovideo:
|
||||
parser.add_argument("--m3u8", dest="m3u8mode", help="If set, it will return M3U8 manifest", action="store_true")
|
||||
parser.add_argument("--file", dest="txtpath", help="If set, it will download links of an txt file")
|
||||
|
||||
#DisneyPlus:
|
||||
parser.add_argument("--tlang", "--title-language", dest="titlelang", nargs=1, help="If set, it will change title language", default="es-419")
|
||||
parser.add_argument("--scenario1", dest="scenarioDSNP", help="Video API from DisneyPlus", default="chromecast-drm-cbcs")
|
||||
parser.add_argument("--scenario2", dest="scenarioSTAR", help="Video API from DisneyPlus", default="restricted-drm-ctr-sw")
|
||||
|
||||
#PROXY:
|
||||
parser.add_argument("--proxy", dest="proxy", help="Proxy URL to use for both fetching metadata and downloading")
|
||||
#proxy format: http://email@email:password@host:port
|
||||
args = parser.parse_args()
|
||||
|
||||
|
||||
if args.debug:
|
||||
import logging
|
||||
logging.basicConfig(level=logging.DEBUG)
|
||||
|
||||
currentFile = '__main__'
|
||||
realPath = os.path.realpath(currentFile)
|
||||
dirPath = os.path.dirname(realPath)
|
||||
dirName = os.path.basename(dirPath)
|
||||
|
||||
if __name__ == "__main__":
|
||||
if args.content:
|
||||
args.url_season = args.content
|
||||
|
||||
if not args.url_season:
|
||||
print('Please specify the URL of the content to download.')
|
||||
sys.exit(1)
|
||||
|
||||
if (args.url_season and 'hbomax' in args.url_season):
|
||||
mode = 'hbomax'
|
||||
import hbomax
|
||||
hbomax.main(args)
|
||||
elif (args.url_season and 'clarovideo' in args.url_season):
|
||||
mode = 'clarovideo'
|
||||
import clarovideo
|
||||
clarovideo.main(args)
|
||||
elif (args.url_season and 'blim' in args.url_season):
|
||||
mode = 'blimtv'
|
||||
import blimtv
|
||||
blimtv.main(args)
|
||||
elif (args.url_season and 'nowonline' in args.url_season):
|
||||
mode = 'nowonline'
|
||||
import nowonline
|
||||
nowonline.main(args)
|
||||
elif (args.url_season and 'globo' in args.url_season):
|
||||
mode = 'globoplay'
|
||||
import globoplay
|
||||
globoplay.main(args)
|
||||
elif (args.url_season and 'paramountplus.com' in args.url_season):
|
||||
mode = 'paramountplus'
|
||||
import paramountplus
|
||||
paramountplus.main(args)
|
||||
elif (args.url_season and 'disneyplus' in args.url_season):
|
||||
mode = 'disneyplus'
|
||||
import disneyplus
|
||||
disneyplus.main(args)
|
||||
elif (args.url_season and 'starplus.com' in args.url_season):
|
||||
mode = 'starplus'
|
||||
import starplus
|
||||
starplus.main(args)
|
||||
elif (args.url_season and 'tv.apple.com' in args.url_season):
|
||||
mode = 'appletv'
|
||||
import appletv
|
||||
appletv.main(args)
|
||||
elif (args.url_season and 'telecine' in args.url_season):
|
||||
mode = 'telecine'
|
||||
import telecineplay
|
||||
telecineplay.main(args)
|
||||
|
||||
else:
|
||||
print("Error! This url or mode is not recognized.")
|
||||
sys.exit(0)
|
0
cookies/cookies_pmnp.txt
Normal file
0
cookies/cookies_pmnp.txt
Normal file
910
paramountplus.py
Normal file
910
paramountplus.py
Normal file
@ -0,0 +1,910 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
# Module: Paramount Plus Downloader
|
||||
# Created on: 19-02-2021
|
||||
# Authors: JUNi
|
||||
# Version: 2.0
|
||||
|
||||
import urllib.parse
|
||||
import re, base64, requests, sys, os
|
||||
import subprocess, shutil
|
||||
import xmltodict, isodate
|
||||
import json, ffmpy, math
|
||||
import http, html, time, pathlib, glob
|
||||
|
||||
from unidecode import unidecode
|
||||
from http.cookiejar import MozillaCookieJar
|
||||
from titlecase import titlecase
|
||||
from pymediainfo import MediaInfo
|
||||
from m3u8 import parse as m3u8parser
|
||||
|
||||
import pywidevine.clients.paramountplus.config as pmnp_cfg
|
||||
from pywidevine.clients.proxy_config import ProxyConfig
|
||||
from pywidevine.muxer.muxer import Muxer
|
||||
|
||||
from pywidevine.clients.paramountplus.downloader import WvDownloader
|
||||
from pywidevine.clients.paramountplus.config import WvDownloaderConfig
|
||||
|
||||
|
||||
currentFile = 'paramountplus'
|
||||
realPath = os.path.realpath(currentFile)
|
||||
dirPath = os.path.dirname(realPath)
|
||||
USER_AGENT = 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/90.0.4430.93 Safari/537.36'
|
||||
SESSION = requests.Session()
|
||||
|
||||
def main(args):
|
||||
global _id
|
||||
|
||||
proxies = {}
|
||||
proxy_meta = args.proxy
|
||||
if proxy_meta == 'none':
|
||||
proxies['meta'] = {'http': None, 'https': None}
|
||||
elif proxy_meta:
|
||||
proxies['meta'] = {'http': proxy_meta, 'https': proxy_meta}
|
||||
SESSION.proxies = proxies.get('meta')
|
||||
proxy_cfg = ProxyConfig(proxies)
|
||||
|
||||
if not os.path.exists(dirPath + '/KEYS'):
|
||||
os.makedirs(dirPath + '/KEYS')
|
||||
else:
|
||||
keys_file = dirPath + '/KEYS/PARAMOUNTPLUS.txt'
|
||||
try:
|
||||
keys_file_pmnp = open(keys_file, 'r')
|
||||
keys_file_txt = keys_file_pmnp.readlines()
|
||||
except Exception:
|
||||
with open(keys_file, 'a', encoding='utf8') as (file):
|
||||
file.write('##### One KEY per line. #####\n')
|
||||
keys_file_pmnp = open(keys_file, 'r', encoding='utf8')
|
||||
keys_file_txt = keys_file_pmnp.readlines()
|
||||
|
||||
def alphanumericSort(l):
|
||||
def convert(text):
|
||||
if text.isdigit():
|
||||
return int(text)
|
||||
else:
|
||||
return text
|
||||
|
||||
def alphanum_key(key):
|
||||
return [convert(c) for c in re.split('([0-9]+)', key)]
|
||||
|
||||
return sorted(l, key=alphanum_key)
|
||||
|
||||
def convert_size(size_bytes):
|
||||
if size_bytes == 0:
|
||||
return '0bps'
|
||||
else:
|
||||
s = round(size_bytes / 1000, 0)
|
||||
return '%ikbps' % s
|
||||
|
||||
def get_size(size):
|
||||
power = 1024
|
||||
n = 0
|
||||
Dic_powerN = {0:'', 1:'K', 2:'M', 3:'G', 4:'T'}
|
||||
while size > power:
|
||||
size /= power
|
||||
n += 1
|
||||
return str(round(size, 2)) + Dic_powerN[n] + 'B'
|
||||
|
||||
def getKeyId(name):
|
||||
mp4dump = subprocess.Popen([pmnp_cfg.MP4DUMP, name], stdout=(subprocess.PIPE))
|
||||
mp4dump = str(mp4dump.stdout.read())
|
||||
A = find_str(mp4dump, 'default_KID')
|
||||
KEY_ID_ORI = ''
|
||||
KEY_ID_ORI = mp4dump[A:A + 63].replace('default_KID = ', '').replace('[', '').replace(']', '').replace(' ', '')
|
||||
if KEY_ID_ORI == '' or KEY_ID_ORI == "'":
|
||||
KEY_ID_ORI = 'nothing'
|
||||
return KEY_ID_ORI
|
||||
|
||||
def find_str(s, char):
|
||||
index = 0
|
||||
if char in s:
|
||||
c = char[0]
|
||||
for ch in s:
|
||||
if ch == c:
|
||||
if s[index:index + len(char)] == char:
|
||||
return index
|
||||
index += 1
|
||||
|
||||
return -1
|
||||
|
||||
def mediainfo_(file):
|
||||
mediainfo_output = subprocess.Popen([pmnp_cfg.MEDIAINFO, '--Output=JSON', '-f', file], stdout=(subprocess.PIPE))
|
||||
mediainfo_json = json.load(mediainfo_output.stdout)
|
||||
return mediainfo_json
|
||||
|
||||
def replace_words(x):
|
||||
x = re.sub(r'[]¡!"#$%\'()*+,:;<=>¿?@\\^_`{|}~[-]', '', x)
|
||||
x = x.replace('/', '-')
|
||||
return unidecode(x)
|
||||
|
||||
def ReplaceSubs1(X):
|
||||
pattern1 = re.compile('(?!<i>|<b>|<u>|<\\/i>|<\\/b>|<\\/u>)(<)(?:[A-Za-z0-9_ -=]*)(>)')
|
||||
pattern2 = re.compile('(?!<\\/i>|<\\/b>|<\\/u>)(<\\/)(?:[A-Za-z0-9_ -=]*)(>)')
|
||||
X = X.replace('‏', '').replace('{\\an1}', '').replace('{\\an2}', '').replace('{\\an3}', '').replace('{\\an4}', '').replace('{\\an5}', '').replace('{\\an6}', '').replace('{\\an7}', '').replace('{\\an8}', '').replace('{\\an9}', '').replace('‎', '')
|
||||
X = pattern1.sub('', X)
|
||||
X = pattern2.sub('', X)
|
||||
return X
|
||||
|
||||
def replace_code_lang(X):
|
||||
X = X.lower()
|
||||
X = X.replace('es-mx', 'es-la').replace('pt-BR', 'pt-br').replace('dolby digital', 'en').replace('dd+', 'en')
|
||||
return X
|
||||
|
||||
def get_cookies(file_path):
|
||||
try:
|
||||
cj = http.cookiejar.MozillaCookieJar(file_path)
|
||||
cj.load()
|
||||
except Exception:
|
||||
print('\nCookies not found! Please dump the cookies with the Chrome extension https://chrome.google.com/webstore/detail/cookiestxt/njabckikapfpffapmjgojcnbfjonfjfg and place the generated file in ' + file_path)
|
||||
print('\nWarning, do not click on "download all cookies", you have to click on "click here".\n')
|
||||
sys.exit(0)
|
||||
|
||||
cookies = str()
|
||||
for cookie in cj:
|
||||
cookie.value = urllib.parse.unquote(html.unescape(cookie.value))
|
||||
cookies = cookies + cookie.name + '=' + cookie.value + ';'
|
||||
|
||||
cookies = list(cookies)
|
||||
del cookies[-1]
|
||||
cookies = ''.join(cookies)
|
||||
return cookies
|
||||
|
||||
cookies_file = 'cookies_pmnp.txt'
|
||||
cookies = get_cookies(dirPath + '/cookies/' + cookies_file)
|
||||
pmnp_headers = {
|
||||
'Accept':'application/json, text/plain, */*',
|
||||
'Access-Control-Allow-Origin':'*',
|
||||
'cookie':cookies,
|
||||
'User-Agent':USER_AGENT
|
||||
}
|
||||
|
||||
def mpd_parsing(mpd_url):
|
||||
base_url = mpd_url.split('stream.mpd')[0]
|
||||
r = SESSION.get(url=mpd_url)
|
||||
r.raise_for_status()
|
||||
xml = xmltodict.parse(r.text)
|
||||
mpdf = json.loads(json.dumps(xml))
|
||||
length = isodate.parse_duration(mpdf['MPD']['@mediaPresentationDuration']).total_seconds()
|
||||
tracks = mpdf['MPD']['Period']['AdaptationSet']
|
||||
|
||||
def get_pssh(track):
|
||||
pssh = ''
|
||||
for t in track["ContentProtection"]:
|
||||
if t['@schemeIdUri'].lower() == 'urn:uuid:edef8ba9-79d6-4ace-a3c8-27dcd51d21ed':
|
||||
pssh = t["cenc:pssh"]
|
||||
return pssh
|
||||
|
||||
def force_instance(x):
|
||||
if isinstance(x['Representation'], list):
|
||||
X = x['Representation']
|
||||
else:
|
||||
X = [x['Representation']]
|
||||
return X
|
||||
|
||||
baseUrl = ''
|
||||
video_list = []
|
||||
for video_tracks in tracks:
|
||||
if video_tracks['@contentType'] == 'video':
|
||||
pssh = get_pssh(video_tracks)
|
||||
for x in force_instance(video_tracks):
|
||||
try:
|
||||
codecs = x['@codecs']
|
||||
except (KeyError, TypeError):
|
||||
codecs = video_tracks['@codecs']
|
||||
try:
|
||||
baseUrl = x["BaseURL"]
|
||||
except (KeyError, TypeError):
|
||||
pass
|
||||
video_dict = {
|
||||
'Height':x['@height'],
|
||||
'Width':x['@width'],
|
||||
'Bandwidth':x['@bandwidth'],
|
||||
'ID':x['@id'],
|
||||
'TID':video_tracks['@id'],
|
||||
'URL':baseUrl,
|
||||
'Codec':codecs}
|
||||
video_list.append(video_dict)
|
||||
|
||||
video_list = sorted(video_list, key=(lambda k: int(k['Bandwidth'])))
|
||||
|
||||
while args.customquality != [] and int(video_list[(-1)]['Height']) > int(args.customquality[0]):
|
||||
video_list.pop(-1)
|
||||
|
||||
audio_list = []
|
||||
for audio_tracks in tracks:
|
||||
if audio_tracks['@contentType'] == 'audio':
|
||||
for x in force_instance(audio_tracks):
|
||||
try:
|
||||
codecs = x['@codecs']
|
||||
except (KeyError, TypeError):
|
||||
codecs = audio_tracks['@codecs']
|
||||
try:
|
||||
baseUrl = x["BaseURL"]
|
||||
except (KeyError, TypeError):
|
||||
pass
|
||||
|
||||
audio_dict = {
|
||||
'Bandwidth':x['@bandwidth'],
|
||||
'ID':x['@id'],
|
||||
'TID':audio_tracks['@id'],
|
||||
'Language':replace_code_lang(audio_tracks['@lang']),
|
||||
'URL':baseUrl,
|
||||
'Codec':codecs}
|
||||
audio_list.append(audio_dict)
|
||||
|
||||
audio_list = sorted(audio_list, key=(lambda k: (str(k['Language']))), reverse=True)
|
||||
|
||||
if args.only_2ch_audio:
|
||||
c = 0
|
||||
while c != len(audio_list):
|
||||
if '-3' in audio_list[c]['Codec'].split('=')[0]:
|
||||
audio_list.remove(audio_list[c])
|
||||
else:
|
||||
c += 1
|
||||
|
||||
BitrateList = []
|
||||
AudioLanguageList = []
|
||||
for x in audio_list:
|
||||
BitrateList.append(x['Bandwidth'])
|
||||
AudioLanguageList.append(x['Language'])
|
||||
|
||||
BitrateList = alphanumericSort(list(set(BitrateList)))
|
||||
AudioLanguageList = alphanumericSort(list(set(AudioLanguageList)))
|
||||
audioList_new = []
|
||||
audio_Dict_new = {}
|
||||
for y in AudioLanguageList:
|
||||
counter = 0
|
||||
for x in audio_list:
|
||||
if x['Language'] == y and counter == 0:
|
||||
audio_Dict_new = {
|
||||
'Language':x['Language'],
|
||||
'Bandwidth':x['Bandwidth'],
|
||||
'Codec': x['Codec'],
|
||||
'TID':x['TID'],
|
||||
'URL':x['URL'],
|
||||
'ID':x['ID']}
|
||||
audioList_new.append(audio_Dict_new)
|
||||
counter = counter + 1
|
||||
|
||||
audioList = audioList_new
|
||||
audio_list = sorted(audioList, key=(lambda k: (int(k['Bandwidth']), str(k['Language']))))
|
||||
|
||||
audioList_new = []
|
||||
if args.audiolang:
|
||||
for x in audio_list:
|
||||
langAbbrev = x['Language']
|
||||
if langAbbrev in list(args.audiolang):
|
||||
audioList_new.append(x)
|
||||
audio_list = audioList_new
|
||||
|
||||
if 'precon' in mpd_url:
|
||||
sub_url = mpd_url.replace('_cenc_precon_dash/stream.mpd', '_fp_precon_hls/master.m3u8')
|
||||
else:
|
||||
sub_url = mpd_url.replace('_cenc_dash/stream.mpd', '_fp_hls/master.m3u8')
|
||||
print(sub_url)
|
||||
|
||||
return base_url, length, video_list, audio_list, get_subtitles(sub_url), pssh, mpdf
|
||||
|
||||
def download_subs(filename, sub_url):
|
||||
urlm3u8_base_url = re.split('(/)(?i)', sub_url)
|
||||
del urlm3u8_base_url[-1]
|
||||
urlm3u8_base_url = ''.join(urlm3u8_base_url)
|
||||
urlm3u8_request = requests.get(sub_url).text
|
||||
m3u8_json = m3u8parser(urlm3u8_request)
|
||||
|
||||
urls = []
|
||||
for segment in m3u8_json['segments']:
|
||||
if 'https://' not in segment['uri']:
|
||||
segment_url = urlm3u8_base_url + segment['uri']
|
||||
urls.append(segment_url)
|
||||
|
||||
print('\n' + filename)
|
||||
aria2c_infile = ""
|
||||
num_segments = len(urls)
|
||||
digits = math.floor(math.log10(num_segments)) + 1
|
||||
for (i, url) in enumerate(urls):
|
||||
aria2c_infile += f"{url}\n"
|
||||
aria2c_infile += f"\tout={filename}.{i:0{digits}d}.vtt\n"
|
||||
aria2c_infile += f"\tdir={filename}\n"
|
||||
subprocess.run([pmnp_cfg.ARIA2C, "--allow-overwrite=true", "--file-allocation=none",
|
||||
"--console-log-level=warn", "--download-result=hide", "--summary-interval=0",
|
||||
"-x16", "-j16", "-s1", "-i-"],
|
||||
input=aria2c_infile.encode("utf-8"))
|
||||
|
||||
source_files = pathlib.Path(filename).rglob(r'./*.vtt')
|
||||
with open(filename + '.vtt', mode='wb') as (destination):
|
||||
for vtt in source_files:
|
||||
with open(vtt, mode='rb') as (source):
|
||||
shutil.copyfileobj(source, destination)
|
||||
|
||||
if os.path.exists(filename):
|
||||
shutil.rmtree(filename)
|
||||
|
||||
print('\nConverting subtitles...')
|
||||
for f in glob.glob(f'{filename}*.vtt'):
|
||||
with open(f, 'r+', encoding='utf-8-sig') as (x):
|
||||
old = x.read().replace('STYLE\n::cue() {\nfont-family: Arial, Helvetica, sans-serif;\n}', '').replace('WEBVTT', '').replace('X-TIMESTAMP-MAP=LOCAL:00:00:00.000,MPEGTS:9000', '').replace('\n\n\n', '\n')
|
||||
with open(f, 'w+', encoding='utf-8-sig') as (x):
|
||||
x.write(ReplaceSubs1(old))
|
||||
SubtitleEdit_process = subprocess.Popen([pmnp_cfg.SUBTITLE_EDIT, '/convert', filename + ".vtt", "srt", "/MergeSameTexts"], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE).wait()
|
||||
for f in glob.glob(f'{filename}*.vtt'):
|
||||
os.remove(f)
|
||||
|
||||
'''
|
||||
for f in glob.glob(f'{filename}*.srt'):
|
||||
with open(f, 'r+', encoding='utf-8-sig') as (x):
|
||||
old = x.read().replace('STYLE\n::cue() {\nfont-family: Arial, Helvetica, sans-serif;\n}', '').replace('WEBVTT', '').replace('\nX-TIMESTAMP-MAP=LOCAL:00:00:00.000,MPEGTS:9000\n', '').replace('\n\n\n', '\n')
|
||||
with open(f, 'w+', encoding='utf-8-sig') as (x):
|
||||
if not args.nocleansubs:
|
||||
x.write(ReplaceSubs1(old))
|
||||
else:
|
||||
x.write(ReplaceSubs2(old))
|
||||
for f in glob.glob(f'{filename}*.vtt'):
|
||||
os.remove(f)
|
||||
'''
|
||||
|
||||
print('Done!')
|
||||
|
||||
def get_episodes(ep_str, num_eps):
|
||||
eps = ep_str.split(',')
|
||||
eps_final = []
|
||||
|
||||
for ep in eps:
|
||||
if '-' in ep:
|
||||
(start, end) = ep.split('-')
|
||||
start = int(start)
|
||||
end = int(end or num_eps)
|
||||
eps_final += list(range(start, end + 1))
|
||||
else:
|
||||
eps_final.append(int(ep))
|
||||
|
||||
return eps_final
|
||||
|
||||
_id = args.url_season.split('/')[-2]
|
||||
if '/video/' in args.url_season:
|
||||
content_regex = r'(\/shows\/)([\w-]+)(\/video\/)([\w-]+)'
|
||||
url_match = re.search(content_regex, args.url_season)
|
||||
_id = url_match[2]
|
||||
|
||||
def get_content_info():
|
||||
if 'shows' in args.url_season:
|
||||
pmnp_season_url = 'https://www.paramountplus.com/shows/{}/xhr/episodes/page/0/size/100/xs/0/season/{}/'.format(_id, '')
|
||||
season_req = requests.get(url=pmnp_season_url, headers=pmnp_headers, proxies=proxy_cfg.get_proxy('meta'))
|
||||
|
||||
if not args.season:
|
||||
args.season = 'all'
|
||||
|
||||
seasons = []
|
||||
if args.season:
|
||||
if args.season == 'all':
|
||||
seasons = 'all'
|
||||
elif ',' in args.season:
|
||||
seasons = [int(x) for x in args.season.split(',')]
|
||||
elif '-' in args.season:
|
||||
(start, end) = args.season.split('-')
|
||||
seasons = list(range(int(start), int(end) + 1))
|
||||
else:
|
||||
seasons = [int(args.season)]
|
||||
|
||||
if seasons == 'all':
|
||||
seasons_list = [x['season_number'] for x in season_req.json()['result']['data']]
|
||||
seasons = sorted(set(seasons_list))
|
||||
|
||||
for season_num in seasons:
|
||||
pmnp_season_url = 'https://www.paramountplus.com/shows/{}/xhr/episodes/page/0/size/500/xs/0/season/{}/'.format(_id, season_num)
|
||||
season_req = requests.get(url=pmnp_season_url, headers=pmnp_headers, proxies=proxy_cfg.get_proxy('meta'))
|
||||
if season_req.json()['result']['total'] < 1:
|
||||
print('This season doesnt exist!')
|
||||
exit()
|
||||
|
||||
for num, ep in enumerate(season_req.json()['result']['data'], start=1):
|
||||
episodeNumber = ep['episode_number']
|
||||
seasonNumber = ep['season_number']
|
||||
|
||||
if ' - ' in ep['series_title']:
|
||||
seriesTitle = ep['series_title'].split(' - ')[0]
|
||||
else:
|
||||
seriesTitle = ep['series_title']
|
||||
episodeTitle = replace_words(ep['label'])
|
||||
seriesName = f'{replace_words(seriesTitle)} S{seasonNumber:0>2}E{episodeNumber:0>2}'
|
||||
folderName = f'{replace_words(seriesTitle)} S{seasonNumber:0>2}'
|
||||
raw_url = urllib.parse.urljoin('https://www.paramountplus.com', ep['metaData']['contentUrl'])
|
||||
|
||||
episodes_list_new = []
|
||||
episodes_dict = {
|
||||
'id': ep['content_id'],
|
||||
'raw_url': raw_url,
|
||||
'pid':ep['metaData']['pid'],
|
||||
'seriesName':seriesName,
|
||||
'folderName':folderName,
|
||||
'episodeNumber': num,
|
||||
'seasonNumber':seasonNumber,
|
||||
'pmnpType': 'show'}
|
||||
episodes_list_new.append(episodes_dict)
|
||||
episodes_list = []
|
||||
for x in episodes_list_new:
|
||||
episodes_list.append(x)
|
||||
#episodes_list = sorted(episodes_list, key=lambda x: x['episodeNumber'])
|
||||
|
||||
if args.episodeStart:
|
||||
eps = get_episodes(args.episodeStart, len(episodes_list))
|
||||
episodes_list = [x for x in episodes_list if x['episodeNumber'] in eps]
|
||||
|
||||
if 'video' in args.url_season:
|
||||
episodes_list = [x for x in episodes_list if x['id'] in url_match.group(4)]
|
||||
|
||||
for content_json in episodes_list:
|
||||
start_process(content_json)
|
||||
|
||||
if 'movies' in args.url_season:
|
||||
while 1:
|
||||
resp = requests.get(url=args.url_season + '/', headers=pmnp_headers, proxies=proxy_cfg.get_proxy('meta'))
|
||||
if resp.ok:
|
||||
break
|
||||
|
||||
html_data = resp
|
||||
html_data = html_data.text.replace('\r\n', '').replace('\n', '').replace('\r', '').replace('\t', '').replace(' ', '')
|
||||
html_data_list = re.split('(</div>)(?i)', html_data)
|
||||
json_web = []
|
||||
for div in html_data_list:
|
||||
if 'player.paramsVO.adCallParams' in div:
|
||||
print()
|
||||
rg = re.compile('(player.metaData = )(.*)(;player.tms_program_id)')
|
||||
m = rg.search(div)
|
||||
if m:
|
||||
json_web = m.group(2)
|
||||
json_web = json.loads(json_web)
|
||||
|
||||
content_dict = {}
|
||||
episodes_list = []
|
||||
year_regex = r'(\d{4})'
|
||||
movieTitle = replace_words(json_web['seriesTitle'])
|
||||
try:
|
||||
r = re.search(year_regex, json_web['airdate'])
|
||||
except KeyError:
|
||||
r = re.search(year_regex, json_web['airdate_tv'])
|
||||
seriesName = f'{movieTitle} ({r.group(0)})'
|
||||
|
||||
content_dict = {
|
||||
'id':json_web['contentId'],
|
||||
'raw_url': str(args.url_season),
|
||||
'pid': json_web['pid'],
|
||||
'seriesName':seriesName,
|
||||
'folderName':None,
|
||||
'episodeNumber':1,
|
||||
'seasonNumber':1,
|
||||
'pmnpType': 'movie'}
|
||||
episodes_list.append(content_dict)
|
||||
|
||||
for content_json in episodes_list:
|
||||
start_process(content_json)
|
||||
|
||||
def get_license(id_json):
|
||||
while 1:
|
||||
resp = requests.get(url=id_json['raw_url'], headers=pmnp_headers, proxies=proxy_cfg.get_proxy('meta'))
|
||||
if resp.ok:
|
||||
break
|
||||
|
||||
html_data = resp
|
||||
html_data = html_data.text.replace('\r\n', '').replace('\n', '').replace('\r', '').replace('\t', '').replace(' ', '')
|
||||
html_data_list = re.split('(</div>)(?i)', html_data)
|
||||
json_web = []
|
||||
for div in html_data_list:
|
||||
if '(!window.CBS.Registry.drmPromise) {' in div:
|
||||
rg = re.compile('(player.drm = )(.*)(;}player.enableCP)')
|
||||
m = rg.search(div)
|
||||
if m:
|
||||
json_web = m.group(2)
|
||||
json_web = json.loads(json_web)
|
||||
|
||||
lic_url = json_web['widevine']['url']
|
||||
header_auth = json_web['widevine']['header']['Authorization']
|
||||
if not lic_url:
|
||||
print('Too many requests...')
|
||||
return lic_url, header_auth
|
||||
|
||||
global folderdownloader
|
||||
if args.output:
|
||||
if not os.path.exists(args.output):
|
||||
os.makedirs(args.output)
|
||||
os.chdir(args.output)
|
||||
if ":" in str(args.output):
|
||||
folderdownloader = str(args.output).replace('/','\\').replace('.\\','\\')
|
||||
else:
|
||||
folderdownloader = dirPath + '\\' + str(args.output).replace('/','\\').replace('.\\','\\')
|
||||
else:
|
||||
folderdownloader = dirPath.replace('/','\\').replace('.\\','\\')
|
||||
|
||||
def get_subtitles(url):
|
||||
master_base_url = re.split('(/)(?i)', url)
|
||||
del master_base_url[-1]
|
||||
master_base_url = ''.join(master_base_url)
|
||||
urlm3u8_request = requests.get(url).text
|
||||
m3u8_json = m3u8parser(urlm3u8_request)
|
||||
|
||||
subs_list = []
|
||||
for media in m3u8_json['media']:
|
||||
if media['type'] == 'SUBTITLES':
|
||||
if 'https://' not in media['uri']:
|
||||
full_url = master_base_url + media['uri']
|
||||
Full_URL_Type = False
|
||||
else:
|
||||
full_url = media['uri']
|
||||
Full_URL_Type = True
|
||||
subs_dict = {
|
||||
'Type':'subtitles',
|
||||
'trackType':'NORMAL',
|
||||
'Language':media['name'],
|
||||
'LanguageID':replace_code_lang(media['language']),
|
||||
'Profile':media['group_id'],
|
||||
'URL':full_url}
|
||||
subs_list.append(subs_dict)
|
||||
|
||||
subsList_new = []
|
||||
if args.sublang:
|
||||
for x in subs_list:
|
||||
sub_lang = x['LanguageID']
|
||||
if sub_lang in list(args.sublang):
|
||||
subsList_new.append(x)
|
||||
|
||||
subs_list = subsList_new
|
||||
|
||||
return subs_list
|
||||
|
||||
def get_manifest(id_json):
|
||||
api_manifest = 'https://link.theplatform.com/s/dJ5BDC/{}?format=SMIL&manifest=m3u&Tracking=true&mbr=true'.format(id_json['pid'])
|
||||
r = requests.get(url=api_manifest, headers=pmnp_headers, proxies=proxy_cfg.get_proxy('meta'))
|
||||
smil = json.loads(json.dumps(xmltodict.parse(r.text)))
|
||||
videoSrc = []
|
||||
try:
|
||||
for x in smil['smil']['body']['seq']['switch']:
|
||||
videoSrc = x['video']['@src']
|
||||
except Exception:
|
||||
videoSrc = smil['smil']['body']['seq']['switch']['video']['@src']
|
||||
lic_url, header_auth = get_license(id_json)
|
||||
return {'mpd_url': videoSrc, 'wvLicense': lic_url, 'wvHeader': header_auth}
|
||||
|
||||
def start_process(content_info):
|
||||
drm_info = get_manifest(content_info)
|
||||
base_url, length, video_list, audio_list, subs_list, pssh, xml = mpd_parsing(drm_info['mpd_url'])
|
||||
video_bandwidth = dict(video_list[(-1)])['Bandwidth']
|
||||
video_height = str(dict(video_list[(-1)])['Height'])
|
||||
video_width = str(dict(video_list[(-1)])['Width'])
|
||||
video_codec = str(dict(video_list[(-1)])['Codec'])
|
||||
video_format_id = str(dict(video_list[(-1)])['ID'])
|
||||
video_track_id = str(dict(video_list[(-1)])['TID'])
|
||||
if not args.license:
|
||||
if not args.novideo:
|
||||
print('\nVIDEO - Bitrate: ' + convert_size(int(video_bandwidth)) + ' - Profile: ' + video_codec.split('=')[0] + ' - Size: ' + get_size(length * float(video_bandwidth) * 0.125) + ' - Dimensions: ' + video_width + 'x' + video_height)
|
||||
print()
|
||||
|
||||
if not args.noaudio:
|
||||
if audio_list != []:
|
||||
for x in audio_list:
|
||||
audio_bandwidth = x['Bandwidth']
|
||||
audio_representation_id = str(x['Codec'])
|
||||
audio_lang = x['Language']
|
||||
print('AUDIO - Bitrate: ' + convert_size(int(audio_bandwidth)) + ' - Profile: ' + audio_representation_id.split('=')[0] + ' - Size: ' + get_size(length * float(audio_bandwidth) * 0.125) + ' - Language: ' + audio_lang)
|
||||
print()
|
||||
|
||||
if not args.nosubs:
|
||||
if subs_list != []:
|
||||
for z in subs_list:
|
||||
sub_lang = z['LanguageID']
|
||||
print('SUBTITLE - Profile: NORMAL - Language: ' + sub_lang)
|
||||
print()
|
||||
|
||||
print('Name: ' + content_info['seriesName'])
|
||||
|
||||
if content_info['pmnpType'] == 'show':
|
||||
CurrentName = content_info['seriesName']
|
||||
CurrentHeigh = str(video_height)
|
||||
VideoOutputName = folderdownloader + '\\' + str(content_info['folderName']) + '\\' + str(CurrentName) + ' [' + str(CurrentHeigh) + 'p].mkv'
|
||||
else:
|
||||
CurrentName = content_info['seriesName']
|
||||
CurrentHeigh = str(video_height)
|
||||
VideoOutputName = folderdownloader + '\\' + str(CurrentName) + '\\' + ' [' + str(CurrentHeigh) + 'p].mkv'
|
||||
|
||||
if args.license:
|
||||
keys_all = get_keys(drm_info, pssh)
|
||||
with open(keys_file, 'a', encoding='utf8') as (file):
|
||||
file.write(CurrentName + '\n')
|
||||
print('\n' + CurrentName)
|
||||
for key in keys_all:
|
||||
with open(keys_file, 'a', encoding='utf8') as (file):
|
||||
file.write(key + '\n')
|
||||
print(key)
|
||||
|
||||
else:
|
||||
|
||||
if not args.novideo or (not args.noaudio):
|
||||
print("\nGetting KEYS...")
|
||||
try:
|
||||
keys_all = get_keys(drm_info, pssh)
|
||||
except KeyError:
|
||||
print('License request failed, using keys from txt')
|
||||
keys_all = keys_file_txt
|
||||
print("Done!")
|
||||
|
||||
if not os.path.isfile(VideoOutputName):
|
||||
aria2c_input = ''
|
||||
if not args.novideo:
|
||||
inputVideo = CurrentName + ' [' + str(CurrentHeigh) + 'p].mp4'
|
||||
if os.path.isfile(inputVideo):
|
||||
print('\n' + inputVideo + '\nFile has already been successfully downloaded previously.\n')
|
||||
else:
|
||||
try:
|
||||
wvdl_cfg = WvDownloaderConfig(xml, base_url, inputVideo, video_track_id, video_format_id)
|
||||
wvdownloader = WvDownloader(wvdl_cfg)
|
||||
wvdownloader.run()
|
||||
except KeyError:
|
||||
url = urllib.parse.urljoin(base_url, video_list[(-1)]['URL'])
|
||||
aria2c_input += f'{url}\n'
|
||||
aria2c_input += f'\tdir={folderdownloader}\n'
|
||||
aria2c_input += f'\tout={inputVideo}\n'
|
||||
|
||||
if not args.noaudio:
|
||||
for x in audio_list:
|
||||
langAbbrev = x['Language']
|
||||
format_id = x['ID']
|
||||
inputAudio = CurrentName + ' ' + '(' + langAbbrev + ').mp4'
|
||||
inputAudio_demuxed = CurrentName + ' ' + '(' + langAbbrev + ')' + '.m4a'
|
||||
if os.path.isfile(inputAudio) or os.path.isfile(inputAudio_demuxed):
|
||||
print('\n' + inputAudio + '\nFile has already been successfully downloaded previously.\n')
|
||||
else:
|
||||
try:
|
||||
wvdl_cfg = WvDownloaderConfig(xml, base_url, inputAudio, x['TID'], x['ID'])
|
||||
wvdownloader = WvDownloader(wvdl_cfg)
|
||||
wvdownloader.run()
|
||||
except KeyError:
|
||||
url = urllib.parse.urljoin(base_url, x['URL'])
|
||||
aria2c_input += f'{url}\n'
|
||||
aria2c_input += f'\tdir={folderdownloader}\n'
|
||||
aria2c_input += f'\tout={inputAudio}\n'
|
||||
|
||||
aria2c_infile = os.path.join(folderdownloader, 'aria2c_infile.txt')
|
||||
with open(aria2c_infile, 'w') as fd:
|
||||
fd.write(aria2c_input)
|
||||
aria2c_opts = [
|
||||
pmnp_cfg.ARIA2C,
|
||||
'--allow-overwrite=true',
|
||||
'--download-result=hide',
|
||||
'--console-log-level=warn',
|
||||
'-x16', '-s16', '-j16',
|
||||
'-i', aria2c_infile]
|
||||
subprocess.run(aria2c_opts, check=True)
|
||||
|
||||
if not args.nosubs:
|
||||
if subs_list != []:
|
||||
for z in subs_list:
|
||||
langAbbrev = z['LanguageID']
|
||||
inputSubtitle = CurrentName + ' ' + '(' + langAbbrev + ')'
|
||||
if os.path.isfile(inputSubtitle + '.vtt') or os.path.isfile(inputSubtitle + '.srt'):
|
||||
print('\n' + inputSubtitle + '\nFile has already been successfully downloaded previously.\n')
|
||||
else:
|
||||
download_subs(inputSubtitle, z['URL'])
|
||||
|
||||
CorrectDecryptVideo = False
|
||||
if not args.novideo:
|
||||
inputVideo = CurrentName + ' [' + str(CurrentHeigh) + 'p].mp4'
|
||||
if os.path.isfile(inputVideo):
|
||||
CorrectDecryptVideo = DecryptVideo(inputVideo=inputVideo, keys_video=keys_all)
|
||||
else:
|
||||
CorrectDecryptVideo = True
|
||||
|
||||
CorrectDecryptAudio = False
|
||||
if not args.noaudio:
|
||||
for x in audio_list:
|
||||
langAbbrev = x['Language']
|
||||
inputAudio = CurrentName + ' ' + '(' + langAbbrev + ')' + '.mp4'
|
||||
if os.path.isfile(inputAudio):
|
||||
CorrectDecryptAudio = DecryptAudio(inputAudio=inputAudio, keys_audio=keys_all)
|
||||
else:
|
||||
CorrectDecryptAudio = True
|
||||
|
||||
if not args.nomux:
|
||||
if not args.novideo:
|
||||
if not args.noaudio:
|
||||
if CorrectDecryptVideo == True:
|
||||
if CorrectDecryptAudio == True:
|
||||
print('\nMuxing...')
|
||||
|
||||
pmnpType = content_info['pmnpType']
|
||||
folderName = content_info['folderName']
|
||||
|
||||
if pmnpType=="show":
|
||||
MKV_Muxer=Muxer(CurrentName=CurrentName,
|
||||
SeasonFolder=folderName,
|
||||
CurrentHeigh=CurrentHeigh,
|
||||
Type=pmnpType,
|
||||
mkvmergeexe=pmnp_cfg.MKVMERGE)
|
||||
|
||||
else:
|
||||
MKV_Muxer=Muxer(CurrentName=CurrentName,
|
||||
SeasonFolder=None,
|
||||
CurrentHeigh=CurrentHeigh,
|
||||
Type=pmnpType,
|
||||
mkvmergeexe=pmnp_cfg.MKVMERGE)
|
||||
|
||||
MKV_Muxer.mkvmerge_muxer(lang="English")
|
||||
|
||||
if args.tag:
|
||||
inputName = CurrentName + ' [' + CurrentHeigh + 'p].mkv'
|
||||
release_group(base_filename=inputName,
|
||||
default_filename=CurrentName,
|
||||
folder_name=folderName,
|
||||
type=pmnpType,
|
||||
video_height=CurrentHeigh)
|
||||
|
||||
if not args.keep:
|
||||
for f in os.listdir():
|
||||
if re.fullmatch(re.escape(CurrentName) + r'.*\.(mp4|m4a|h264|h265|eac3|ac3|srt|txt|avs|lwi|mpd)', f):
|
||||
os.remove(f)
|
||||
print("Done!")
|
||||
else:
|
||||
print("\nFile '" + str(VideoOutputName) + "' already exists.")
|
||||
|
||||
def release_group(base_filename, default_filename, folder_name, type, video_height):
|
||||
if type=='show':
|
||||
video_mkv = os.path.join(folder_name, base_filename)
|
||||
else:
|
||||
video_mkv = base_filename
|
||||
|
||||
mediainfo = mediainfo_(video_mkv)
|
||||
for v in mediainfo['media']['track']: # mediainfo do video
|
||||
if v['@type'] == 'Video':
|
||||
video_format = v['Format']
|
||||
|
||||
video_codec = ''
|
||||
if video_format == "AVC":
|
||||
video_codec = 'H.264'
|
||||
elif video_format == "HEVC":
|
||||
video_codec = 'H.265'
|
||||
|
||||
for m in mediainfo['media']['track']: # mediainfo do audio
|
||||
if m['@type'] == 'Audio':
|
||||
codec_name = m['Format']
|
||||
channels_number = m['Channels']
|
||||
|
||||
audio_codec = ''
|
||||
audio_channels = ''
|
||||
if codec_name == "AAC":
|
||||
audio_codec = 'AAC'
|
||||
elif codec_name == "AC-3":
|
||||
audio_codec = "DD"
|
||||
elif codec_name == "E-AC-3":
|
||||
audio_codec = "DDP"
|
||||
elif codec_name == "E-AC-3 JOC":
|
||||
audio_codec = "ATMOS"
|
||||
|
||||
if channels_number == "2":
|
||||
audio_channels = "2.0"
|
||||
elif channels_number == "6":
|
||||
audio_channels = "5.1"
|
||||
|
||||
audio_ = audio_codec + audio_channels
|
||||
|
||||
# renomear arquivo
|
||||
default_filename = default_filename.replace('&', '.and.')
|
||||
default_filename = re.sub(r'[]!"#$%\'()*+,:;<=>?@\\^_`{|}~[-]', '', default_filename)
|
||||
default_filename = default_filename.replace(' ', '.')
|
||||
default_filename = re.sub(r'\.{2,}', '.', default_filename)
|
||||
default_filename = unidecode(default_filename)
|
||||
|
||||
output_name = '{}.{}p.PMNP.WEB-DL.{}.{}-{}'.format(default_filename, video_height, audio_, video_codec, args.tag)
|
||||
if type=='show':
|
||||
outputName = os.path.join(folder_name, output_name + '.mkv')
|
||||
else:
|
||||
outputName = output_name + '.mkv'
|
||||
|
||||
os.rename(video_mkv, outputName)
|
||||
print("{} -> {}".format(base_filename, output_name))
|
||||
|
||||
from pywidevine.decrypt.wvdecryptcustom import WvDecrypt
|
||||
from pywidevine.cdm import cdm, deviceconfig
|
||||
|
||||
def get_keys(licInfo, pssh):
|
||||
device = deviceconfig.device_asus_x00dd
|
||||
wvdecrypt = WvDecrypt(init_data_b64=bytes(pssh.encode()), cert_data_b64=None, device=device)
|
||||
license_req = SESSION.post(url=licInfo['wvLicense'], headers={'authorization':licInfo['wvHeader']}, data=wvdecrypt.get_challenge(), proxies=proxy_cfg.get_proxy('meta')).content
|
||||
license_b64 = base64.b64encode(license_req)
|
||||
|
||||
wvdecrypt.update_license(license_b64)
|
||||
status, keys = wvdecrypt.start_process()
|
||||
return keys
|
||||
|
||||
def DecryptAudio(inputAudio, keys_audio):
|
||||
key_audio_id_original = getKeyId(inputAudio)
|
||||
outputAudioTemp = inputAudio.replace(".mp4", "_dec.mp4")
|
||||
if key_audio_id_original != "nothing":
|
||||
for key in keys_audio:
|
||||
key_id=key[0:32]
|
||||
if key_id == key_audio_id_original:
|
||||
print("\nDecrypting audio...")
|
||||
print ("Using KEY: " + key)
|
||||
wvdecrypt_process = subprocess.Popen([pmnp_cfg.MP4DECRYPT, "--show-progress", "--key", key, inputAudio, outputAudioTemp])
|
||||
stdoutdata, stderrdata = wvdecrypt_process.communicate()
|
||||
wvdecrypt_process.wait()
|
||||
time.sleep (50.0/1000.0)
|
||||
os.remove(inputAudio)
|
||||
print("\nDemuxing audio...")
|
||||
mediainfo = MediaInfo.parse(outputAudioTemp)
|
||||
audio_info = next(x for x in mediainfo.tracks if x.track_type == "Audio")
|
||||
codec_name = audio_info.format
|
||||
|
||||
ext = ''
|
||||
if codec_name == "AAC":
|
||||
ext = '.m4a'
|
||||
elif codec_name == "E-AC-3":
|
||||
ext = ".eac3"
|
||||
elif codec_name == "AC-3":
|
||||
ext = ".ac3"
|
||||
outputAudio = outputAudioTemp.replace("_dec.mp4", ext)
|
||||
print("{} -> {}".format(outputAudioTemp, outputAudio))
|
||||
ff = ffmpy.FFmpeg(executable=pmnp_cfg.FFMPEG, inputs={outputAudioTemp: None}, outputs={outputAudio: '-c copy'}, global_options="-y -hide_banner -loglevel warning")
|
||||
ff.run()
|
||||
time.sleep (50.0/1000.0)
|
||||
os.remove(outputAudioTemp)
|
||||
print("Done!")
|
||||
return True
|
||||
|
||||
elif key_audio_id_original == "nothing":
|
||||
return True
|
||||
|
||||
def DecryptVideo(inputVideo, keys_video):
|
||||
key_video_id_original = getKeyId(inputVideo)
|
||||
inputVideo = inputVideo
|
||||
outputVideoTemp = inputVideo.replace('.mp4', '_dec.mp4')
|
||||
outputVideo = inputVideo
|
||||
if key_video_id_original != 'nothing':
|
||||
for key in keys_video:
|
||||
key_id = key[0:32]
|
||||
if key_id == key_video_id_original:
|
||||
print('\nDecrypting video...')
|
||||
print('Using KEY: ' + key)
|
||||
wvdecrypt_process = subprocess.Popen([pmnp_cfg.MP4DECRYPT, '--show-progress', '--key', key, inputVideo, outputVideoTemp])
|
||||
stdoutdata, stderrdata = wvdecrypt_process.communicate()
|
||||
wvdecrypt_process.wait()
|
||||
print('\nRemuxing video...')
|
||||
ff = ffmpy.FFmpeg(executable=pmnp_cfg.FFMPEG, inputs={outputVideoTemp: None}, outputs={outputVideo: '-c copy'}, global_options='-y -hide_banner -loglevel warning')
|
||||
ff.run()
|
||||
time.sleep(0.05)
|
||||
os.remove(outputVideoTemp)
|
||||
print('Done!')
|
||||
return True
|
||||
|
||||
elif key_video_id_original == 'nothing':
|
||||
return True
|
||||
|
||||
def DemuxAudio(inputAudio):
|
||||
if os.path.isfile(inputAudio):
|
||||
print('\nDemuxing audio...')
|
||||
mediainfo = mediainfo_(inputAudio)
|
||||
for m in mediainfo['media']['track']:
|
||||
if m['@type'] == 'Audio':
|
||||
codec_name = m['Format']
|
||||
try:
|
||||
codec_tag_string = m['Format_Commercial_IfAny']
|
||||
except Exception:
|
||||
codec_tag_string = ''
|
||||
|
||||
ext = ''
|
||||
if codec_name == 'AAC':
|
||||
ext = '.m4a'
|
||||
else:
|
||||
if codec_name == 'E-AC-3':
|
||||
ext = '.eac3'
|
||||
else:
|
||||
if codec_name == 'AC-3':
|
||||
ext = '.ac3'
|
||||
outputAudio = inputAudio.replace('.mp4', ext)
|
||||
print('{} -> {}'.format(inputAudio, outputAudio))
|
||||
ff = ffmpy.FFmpeg(executable=pmnp_cfg.FFMPEG,
|
||||
inputs={inputAudio: None},
|
||||
outputs={outputAudio: '-c copy'},
|
||||
global_options='-y -hide_banner -loglevel warning')
|
||||
ff.run()
|
||||
time.sleep(0.05)
|
||||
os.remove(inputAudio)
|
||||
print('Done!')
|
||||
|
||||
get_content_info()
|
0
pywidevine/cdm/__init__.py
Normal file
0
pywidevine/cdm/__init__.py
Normal file
380
pywidevine/cdm/cdm.py
Normal file
380
pywidevine/cdm/cdm.py
Normal file
@ -0,0 +1,380 @@
|
||||
import base64
|
||||
|
||||
import os
|
||||
import time
|
||||
import binascii
|
||||
|
||||
from google.protobuf.message import DecodeError
|
||||
from google.protobuf import text_format
|
||||
|
||||
from pywidevine.cdm.formats import wv_proto2_pb2 as wv_proto2
|
||||
from pywidevine.cdm.session import Session
|
||||
from pywidevine.cdm.key import Key
|
||||
from Crypto.Random import get_random_bytes
|
||||
from Crypto.Random import random
|
||||
from Crypto import Random
|
||||
from Crypto.Cipher import PKCS1_OAEP, AES
|
||||
from Crypto.Hash import CMAC, SHA256, HMAC, SHA1
|
||||
from Crypto.PublicKey import RSA
|
||||
from Crypto.Signature import pss
|
||||
from Crypto.Util import Padding
|
||||
from pywidevine.cdm import cdmapi
|
||||
import logging
|
||||
|
||||
|
||||
class Cdm:
|
||||
def __init__(self):
|
||||
self.logger = logging.getLogger(__name__)
|
||||
self.sessions = {}
|
||||
|
||||
def open_session(self, init_data_b64, device, raw_init_data=None, offline=False):
|
||||
self.logger.debug("open_session(init_data_b64={}, device={}".format(init_data_b64, device))
|
||||
self.logger.info("opening new cdm session")
|
||||
if device.session_id_type == 'android':
|
||||
# format: 16 random hexdigits, 2 digit counter, 14 0s
|
||||
rand_ascii = ''.join(random.choice('ABCDEF0123456789') for _ in range(16))
|
||||
counter = '01' # this resets regularly so its fine to use 01
|
||||
rest = '00000000000000'
|
||||
session_id = rand_ascii + counter + rest
|
||||
session_id = session_id.encode('ascii')
|
||||
elif device.session_id_type == 'chrome':
|
||||
rand_bytes = get_random_bytes(16)
|
||||
session_id = rand_bytes
|
||||
else:
|
||||
# other formats NYI
|
||||
self.logger.error("device type is unusable")
|
||||
return 1
|
||||
if raw_init_data and isinstance(raw_init_data, (bytes, bytearray)):
|
||||
# used for NF key exchange, where they don't provide a valid PSSH
|
||||
init_data = raw_init_data
|
||||
self.raw_pssh = True
|
||||
else:
|
||||
init_data = self._parse_init_data(init_data_b64)
|
||||
self.raw_pssh = False
|
||||
|
||||
if init_data:
|
||||
new_session = Session(session_id, init_data, device, offline)
|
||||
else:
|
||||
self.logger.error("unable to parse init data")
|
||||
return 1
|
||||
self.sessions[session_id] = new_session
|
||||
self.logger.info("session opened and init data parsed successfully")
|
||||
return session_id
|
||||
|
||||
def _parse_init_data(self, init_data_b64):
|
||||
parsed_init_data = wv_proto2.WidevineCencHeader()
|
||||
try:
|
||||
self.logger.debug("trying to parse init_data directly")
|
||||
parsed_init_data.ParseFromString(base64.b64decode(init_data_b64)[32:])
|
||||
except DecodeError:
|
||||
self.logger.debug("unable to parse as-is, trying with removed pssh box header")
|
||||
try:
|
||||
id_bytes = parsed_init_data.ParseFromString(base64.b64decode(init_data_b64)[32:])
|
||||
except DecodeError:
|
||||
self.logger.error("unable to parse, unsupported init data format")
|
||||
return None
|
||||
self.logger.debug("init_data:")
|
||||
for line in text_format.MessageToString(parsed_init_data).splitlines():
|
||||
self.logger.debug(line)
|
||||
return parsed_init_data
|
||||
|
||||
def close_session(self, session_id):
|
||||
self.logger.debug("close_session(session_id={})".format(session_id))
|
||||
self.logger.info("closing cdm session")
|
||||
if session_id in self.sessions:
|
||||
self.sessions.pop(session_id)
|
||||
self.logger.info("cdm session closed")
|
||||
return 0
|
||||
else:
|
||||
self.logger.info("session {} not found".format(session_id))
|
||||
return 1
|
||||
|
||||
def set_service_certificate(self, session_id, cert_b64):
|
||||
self.logger.debug("set_service_certificate(session_id={}, cert={})".format(session_id, cert_b64))
|
||||
self.logger.info("setting service certificate")
|
||||
|
||||
if session_id not in self.sessions:
|
||||
self.logger.error("session id doesn't exist")
|
||||
return 1
|
||||
|
||||
session = self.sessions[session_id]
|
||||
|
||||
message = wv_proto2.SignedMessage()
|
||||
|
||||
try:
|
||||
message.ParseFromString(base64.b64decode(cert_b64))
|
||||
except DecodeError:
|
||||
self.logger.error("failed to parse cert as SignedMessage")
|
||||
|
||||
service_certificate = wv_proto2.SignedDeviceCertificate()
|
||||
|
||||
if message.Type:
|
||||
self.logger.debug("service cert provided as signedmessage")
|
||||
try:
|
||||
service_certificate.ParseFromString(message.Msg)
|
||||
except DecodeError:
|
||||
self.logger.error("failed to parse service certificate")
|
||||
return 1
|
||||
else:
|
||||
self.logger.debug("service cert provided as signeddevicecertificate")
|
||||
try:
|
||||
service_certificate.ParseFromString(base64.b64decode(cert_b64))
|
||||
except DecodeError:
|
||||
self.logger.error("failed to parse service certificate")
|
||||
return 1
|
||||
|
||||
self.logger.debug("service certificate:")
|
||||
for line in text_format.MessageToString(service_certificate).splitlines():
|
||||
self.logger.debug(line)
|
||||
|
||||
session.service_certificate = service_certificate
|
||||
session.privacy_mode = True
|
||||
return 0
|
||||
|
||||
def sign_license_request(self, data):
|
||||
em = binascii.b2a_hex((pss._EMSA_PSS_ENCODE(data, 2047, Random.get_random_bytes, lambda x, y: pss.MGF1(x, y, data), 20)))
|
||||
sig = cdmapi.encrypt(em.decode('utf-8'))
|
||||
return (binascii.a2b_hex(sig))
|
||||
|
||||
def get_license_request(self, session_id):
|
||||
self.logger.debug("get_license_request(session_id={})".format(session_id))
|
||||
self.logger.info("getting license request")
|
||||
|
||||
if session_id not in self.sessions:
|
||||
self.logger.error("session ID does not exist")
|
||||
return 1
|
||||
|
||||
session = self.sessions[session_id]
|
||||
|
||||
# raw pssh will be treated as bytes and not parsed
|
||||
if self.raw_pssh:
|
||||
license_request = wv_proto2.SignedLicenseRequestRaw()
|
||||
else:
|
||||
license_request = wv_proto2.SignedLicenseRequest()
|
||||
client_id = wv_proto2.ClientIdentification()
|
||||
|
||||
if not os.path.exists(session.device_config.device_client_id_blob_filename):
|
||||
self.logger.error("no client ID blob available for this device")
|
||||
return 1
|
||||
|
||||
with open(session.device_config.device_client_id_blob_filename, "rb") as f:
|
||||
try:
|
||||
cid_bytes = client_id.ParseFromString(f.read())
|
||||
except DecodeError:
|
||||
self.logger.error("client id failed to parse as protobuf")
|
||||
return 1
|
||||
|
||||
self.logger.debug("building license request")
|
||||
if not self.raw_pssh:
|
||||
license_request.Type = wv_proto2.SignedLicenseRequest.MessageType.Value('LICENSE_REQUEST')
|
||||
license_request.Msg.ContentId.CencId.Pssh.CopyFrom(session.init_data)
|
||||
else:
|
||||
license_request.Type = wv_proto2.SignedLicenseRequestRaw.MessageType.Value('LICENSE_REQUEST')
|
||||
license_request.Msg.ContentId.CencId.Pssh = session.init_data # bytes
|
||||
|
||||
if session.offline:
|
||||
license_type = wv_proto2.LicenseType.Value('OFFLINE')
|
||||
else:
|
||||
license_type = wv_proto2.LicenseType.Value('DEFAULT')
|
||||
license_request.Msg.ContentId.CencId.LicenseType = license_type
|
||||
license_request.Msg.ContentId.CencId.RequestId = session_id
|
||||
license_request.Msg.Type = wv_proto2.LicenseRequest.RequestType.Value('NEW')
|
||||
license_request.Msg.RequestTime = int(time.time())
|
||||
license_request.Msg.ProtocolVersion = wv_proto2.ProtocolVersion.Value('CURRENT')
|
||||
if session.device_config.send_key_control_nonce:
|
||||
license_request.Msg.KeyControlNonce = random.randrange(1, 2**31)
|
||||
|
||||
if session.privacy_mode:
|
||||
|
||||
if session.device_config.vmp:
|
||||
self.logger.debug("vmp required, adding to client_id")
|
||||
self.logger.debug("reading vmp hashes")
|
||||
vmp_hashes = wv_proto2.FileHashes()
|
||||
with open(session.device_config.device_vmp_blob_filename, "rb") as f:
|
||||
try:
|
||||
vmp_bytes = vmp_hashes.ParseFromString(f.read())
|
||||
except DecodeError:
|
||||
self.logger.error("vmp hashes failed to parse as protobuf")
|
||||
return 1
|
||||
client_id._FileHashes.CopyFrom(vmp_hashes)
|
||||
self.logger.debug("privacy mode & service certificate loaded, encrypting client id")
|
||||
self.logger.debug("unencrypted client id:")
|
||||
for line in text_format.MessageToString(client_id).splitlines():
|
||||
self.logger.debug(line)
|
||||
cid_aes_key = get_random_bytes(16)
|
||||
cid_iv = get_random_bytes(16)
|
||||
|
||||
cid_cipher = AES.new(cid_aes_key, AES.MODE_CBC, cid_iv)
|
||||