mirror of
https://cdm-project.com/Download-Tools/udemy-downloader.git
synced 2025-04-30 02:14:25 +02:00
Added downloading for multisegmented videos
Download support for multibitstreaming manifests.
This commit is contained in:
parent
03a9a201dd
commit
e2be55b2ae
@ -10,7 +10,7 @@ from mpegdash.utils import (
|
||||
#global ids
|
||||
retry = 3
|
||||
download_dir = os.getcwd() # set the folder to output
|
||||
working_dir = os.getcwd() # set the folder to download emphemeral files
|
||||
working_dir = os.getcwd() # set the folder to download ephemeral files
|
||||
keyfile_path = working_dir + "/keyfile.json"
|
||||
|
||||
#Patching the Mpegdash lib for keyID
|
||||
|
217
dashdownloader_multisegment.py
Normal file
217
dashdownloader_multisegment.py
Normal file
@ -0,0 +1,217 @@
|
||||
#dashdrmmultisegmentdownloader
|
||||
import os,requests,shutil,json,glob,re
|
||||
from mpegdash.parser import MPEGDASHParser
|
||||
from mpegdash.nodes import Descriptor
|
||||
from mpegdash.utils import (
|
||||
parse_attr_value, parse_child_nodes, parse_node_value,
|
||||
write_attr_value, write_child_node, write_node_value
|
||||
)
|
||||
|
||||
#global ids
|
||||
retry = 3
|
||||
download_dir = os.getcwd() # set the folder to output
|
||||
working_dir = os.getcwd() + "\working_dir" # set the folder to download ephemeral files
|
||||
keyfile_path = download_dir + "\keyfile_test.json"
|
||||
|
||||
if not os.path.exists(working_dir):
|
||||
os.makedirs()
|
||||
|
||||
#Get the keys
|
||||
with open(keyfile_path,'r') as keyfile:
|
||||
keyfile = keyfile.read()
|
||||
keyfile = json.loads(keyfile)
|
||||
|
||||
|
||||
#Patching the Mpegdash lib for keyID
|
||||
def __init__(self):
|
||||
self.scheme_id_uri = '' # xs:anyURI (required)
|
||||
self.value = None # xs:string
|
||||
self.id = None # xs:string
|
||||
self.key_id = None # xs:string
|
||||
|
||||
def parse(self, xmlnode):
|
||||
self.scheme_id_uri = parse_attr_value(xmlnode, 'schemeIdUri', str)
|
||||
self.value = parse_attr_value(xmlnode, 'value', str)
|
||||
self.id = parse_attr_value(xmlnode, 'id', str)
|
||||
self.key_id = parse_attr_value(xmlnode, 'cenc:default_KID', str)
|
||||
|
||||
def write(self, xmlnode):
|
||||
write_attr_value(xmlnode, 'schemeIdUri', self.scheme_id_uri)
|
||||
write_attr_value(xmlnode, 'value', self.value)
|
||||
write_attr_value(xmlnode, 'id', self.id)
|
||||
write_attr_value(xmlnode, 'cenc:default_KID', self.key_id)
|
||||
|
||||
Descriptor.__init__ = __init__
|
||||
Descriptor.parse = parse
|
||||
Descriptor.write = write
|
||||
|
||||
#Compiled regex time
|
||||
days = re.compile("([\d.]+)D")
|
||||
hours = re.compile("([\d.]+)H")
|
||||
minutes = re.compile("([\d.]+)M")
|
||||
seconds = re.compile("([\d.]+)S")
|
||||
|
||||
def durationtoseconds(period):
|
||||
#Duration format in PTxDxHxMxS
|
||||
if(period[:2] == "PT"):
|
||||
period = period[2:]
|
||||
day = int(days.match(period).group(1)) if days.match(period) else 0 # Probably never happens
|
||||
hour = int(hours.match(period).group(1)) if hours.match(period) else 0
|
||||
minute = int(minutes.match(period).group(1)) if minutes.match(period) else 0
|
||||
second = seconds.match(period).group(1)
|
||||
total_time = float(str((day * 24 * 60 * 60) + (hour * 60 * 60) + (minute * 60) + (int(second.split('.')[0]))) + '.' + str(int(second.split('.')[-1])))
|
||||
return total_time
|
||||
|
||||
else:
|
||||
print("Duration Format Error")
|
||||
return None
|
||||
|
||||
def download_media(filename,url,epoch = 0):
|
||||
if(os.path.isfile(filename)):
|
||||
media_head = requests.head(url, allow_redirects = True)
|
||||
if media_head.status_code == 200:
|
||||
media_length = int(media_head.headers.get("content-length"))
|
||||
if(os.path.getsize(filename) >= media_length):
|
||||
print("Video already downloaded.. skipping Downloading..")
|
||||
else:
|
||||
print("Redownloading faulty download..")
|
||||
os.remove(filename) #Improve removing logic
|
||||
download_media(filename,url)
|
||||
else:
|
||||
if (epoch > retry):
|
||||
exit("Server doesn't support HEAD.")
|
||||
download_media(filename,url,epoch + 1)
|
||||
else:
|
||||
media = requests.get(url, stream=True)
|
||||
media_length = int(media.headers.get("content-length"))
|
||||
if media.status_code == 200:
|
||||
if(os.path.isfile(filename) and os.path.getsize(filename) >= video_length):
|
||||
print("Video already downloaded.. skipping write to disk..")
|
||||
else:
|
||||
try:
|
||||
with open(filename, 'wb') as video_file:
|
||||
shutil.copyfileobj(video.raw, video_file)
|
||||
return False #Successfully downloaded the file
|
||||
except:
|
||||
print("Connection error: Reattempting download of video..")
|
||||
download_media(filename,url, epoch + 1)
|
||||
|
||||
if os.path.getsize(filename) >= video_length:
|
||||
pass
|
||||
else:
|
||||
print("Error downloaded video is faulty.. Retrying to download")
|
||||
download_media(filename,url, epoch + 1)
|
||||
elif(media.status_code == 404):
|
||||
print("Probably end hit!\n",url)
|
||||
return True #Probably hit the last of the file
|
||||
else:
|
||||
if (epoch > retry):
|
||||
exit("Error Video fetching exceeded retry times.")
|
||||
print("Error fetching video file.. Retrying to download")
|
||||
download_media(filename,url, epoch + 1)
|
||||
|
||||
def cleanup(path):
|
||||
leftover_files = glob.glob(path + '/*.mp4', recursive=True)
|
||||
mpd_files = glob.glob(path + '/*.mpd', recursive=True)
|
||||
leftover_files = leftover_files + mpd_files
|
||||
for file_list in leftover_files:
|
||||
try:
|
||||
os.remove(file_list)
|
||||
except OSError:
|
||||
print(f"Error deleting file: {file_list}")
|
||||
|
||||
def mux_process(video_title,outfile):
|
||||
if os.name == "nt":
|
||||
command = f"ffmpeg -y -i decrypted_audio.mp4 -i decrypted_video.mp4 -acodec copy -vcodec copy -fflags +bitexact -map_metadata -1 -metadata title=\"{video_title}\" -metadata creation_time=2020-00-00T70:05:30.000000Z {outfile}.mp4"
|
||||
else:
|
||||
command = f"nice -n 7 ffmpeg -y -i decrypted_audio.mp4 -i decrypted_video.mp4 -acodec copy -vcodec copy -fflags +bitexact -map_metadata -1 -metadata title=\"{video_title}\" -metadata creation_time=2020-00-00T70:05:30.000000Z {outfile}.mp4"
|
||||
os.system(command)
|
||||
|
||||
def decrypt(kid,filename):
|
||||
try:
|
||||
key = keyfile[kid.lower()]
|
||||
except KeyError as error:
|
||||
exit("Key not found")
|
||||
if(os.name == "nt"):
|
||||
os.system(f"mp4decrypt --key 1:{key} encrypted_{filename}.mp4 decrypted_{filename}.mp4")
|
||||
else:
|
||||
os.system(f"nice -n 7 mp4decrypt --key 1:{key} encrypted_{filename}.mp4 decrypted_{filename}.mp4")
|
||||
|
||||
|
||||
def handle_irregular_segments(media_info,video_title,output_path):
|
||||
no_segment,video_url,video_kid,no_segment,audio_url,audio_kid = media_info
|
||||
video_url = base_url + video_url
|
||||
audio_url = base_url + audio_url
|
||||
video_init = video_url.replace("$Number$","init")
|
||||
audio_init = audio_url.replace("$Number$","init")
|
||||
download_media("video_1.mp4",video_init)
|
||||
download_media("audio_1.mp4",audio_init)
|
||||
for count in range(2,no_segment):
|
||||
video_segment_url = video_url.replace("$Number$",str(count))
|
||||
audio_segment_url = audio_url.replace("$Number$",str(count))
|
||||
video_status = download_media(f"video_{str(count)}.mp4",video_segment_url)
|
||||
audio_status = download_media(f"audio_{str(count)}.mp4",audio_segment_url)
|
||||
if(video_status):
|
||||
if os.name == "nt":
|
||||
video_concat_command = "copy /b " + "+".join([f"video_{i}.mp4" for i in range(1,count)]) + " encrypted_video.mp4"
|
||||
audio_concat_command = "copy /b " + "+".join([f"audio_{i}.mp4" for i in range(1,count)]) + " encrypted_audio.mp4"
|
||||
else:
|
||||
video_concat_command = "cat " + " ".join([f"video_{i}.mp4" for i in range(1,count)]) + " > encrypted_video.mp4"
|
||||
audio_concat_command = "cat " + " ".join([f"audio_{i}.mp4" for i in range(1,count)]) + " > encrypted_audio.mp4"
|
||||
os.system(video_concat_command)
|
||||
os.system(audio_concat_command)
|
||||
break
|
||||
decrypt(video_kid,"video")
|
||||
decrypt(audio_kid,"audio")
|
||||
mux_process(video_title,output_path)
|
||||
|
||||
|
||||
def manifest_parser(mpd_url):
|
||||
video = []
|
||||
audio = []
|
||||
manifest = requests.get(mpd_url).text
|
||||
with open("manifest.mpd",'w') as manifest_handler:
|
||||
manifest_handler.write(manifest)
|
||||
mpd = MPEGDASHParser.parse("./manifest.mpd")
|
||||
running_time = durationtoseconds(mpd.media_presentation_duration)
|
||||
for period in mpd.periods:
|
||||
for adapt_set in period.adaptation_sets:
|
||||
print(adapt_set.mime_type)
|
||||
content_type = adapt_set.mime_type
|
||||
repr = adapt_set.representations[-1] # Max Quality
|
||||
for segment in repr.segment_templates:
|
||||
if(segment.duration):
|
||||
print("Media segments are of equal timeframe")
|
||||
segment_time = segment.duration / segment.timescale
|
||||
total_segments = running_time / segment_time
|
||||
else:
|
||||
print("Media segments are of inequal timeframe")
|
||||
print(segment.media)
|
||||
approx_no_segments = int(running_time // 10) # aproximate of 10 sec per segment
|
||||
print("Expected No of segments:",approx_no_segments)
|
||||
if(content_type == "audio/mp4"):
|
||||
audio.append(approx_no_segments)
|
||||
audio.append(segment.media)
|
||||
elif(content_type == "video/mp4"):
|
||||
video.append(approx_no_segments)
|
||||
video.append(segment.media)
|
||||
for prot in repr.content_protections:
|
||||
if(prot.value == "cenc"):
|
||||
kId = prot.key_id.replace('-','')
|
||||
if(content_type == "audio/mp4"):
|
||||
audio.append(kId)
|
||||
elif(content_type == "video/mp4"):
|
||||
video.append(kId)
|
||||
return video + audio
|
||||
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
mpd = "https://www.example.com/index.mpd"
|
||||
base_url = mpd.split("index.mpd")[0]
|
||||
os.chdir(working_dir)
|
||||
media_info = manifest_parser(mpd)
|
||||
video_title = "Title of the video"
|
||||
output_path = download_dir + "\output_video_name"
|
||||
handle_irregular_segments(media_info,video_title,output_path)
|
||||
cleanup(working_dir)
|
Loading…
x
Reference in New Issue
Block a user