Skip to content

Commit

Permalink
Revert unrelated changes
Browse files Browse the repository at this point in the history
  • Loading branch information
pukkandan committed Jul 15, 2023
1 parent 627df34 commit 591f915
Show file tree
Hide file tree
Showing 17 changed files with 56 additions and 47 deletions.
2 changes: 1 addition & 1 deletion yt_dlp/extractor/abematv.py
Expand Up @@ -17,11 +17,11 @@
from ..aes import aes_ecb_decrypt
from ..utils import (
ExtractorError,
OnDemandPagedList,
bytes_to_intlist,
decode_base_n,
int_or_none,
intlist_to_bytes,
OnDemandPagedList,
time_seconds,
traverse_obj,
update_url_query,
Expand Down
2 changes: 1 addition & 1 deletion yt_dlp/extractor/ceskatelevize.py
Expand Up @@ -282,7 +282,7 @@ def _fix_subtitle(subtitle):
if m:
yield m.group(1)
start, stop = (_msectotimecode(int(t)) for t in m.groups()[1:])
yield f'{start} --> {stop}'
yield '{0} --> {1}'.format(start, stop)
else:
yield line

Expand Down
6 changes: 4 additions & 2 deletions yt_dlp/extractor/globo.py
Expand Up @@ -5,7 +5,9 @@
import re

from .common import InfoExtractor
from ..compat import compat_str
from ..compat import (
compat_str,
)
from ..networking import HEADRequest
from ..utils import (
ExtractorError,
Expand Down Expand Up @@ -228,7 +230,7 @@ class GloboArticleIE(InfoExtractor):

@classmethod
def suitable(cls, url):
return False if GloboIE.suitable(url) else super().suitable(url)
return False if GloboIE.suitable(url) else super(GloboArticleIE, cls).suitable(url)

def _real_extract(self, url):
display_id = self._match_id(url)
Expand Down
2 changes: 1 addition & 1 deletion yt_dlp/extractor/miomio.py
Expand Up @@ -58,7 +58,7 @@ def _extract_mioplayer(self, webpage, video_id, title, http_headers):
video_id)

vid_config_request = Request(
f'http://www.miomio.tv/mioplayer/mioplayerconfigfiles/sina.php?{xml_config}',
'http://www.miomio.tv/mioplayer/mioplayerconfigfiles/sina.php?{0}'.format(xml_config),
headers=http_headers)

# the following xml contains the actual configuration information on the video file(s)
Expand Down
2 changes: 1 addition & 1 deletion yt_dlp/extractor/nbc.py
Expand Up @@ -2,9 +2,9 @@
import json
import re

from .adobepass import AdobePassIE
from .common import InfoExtractor
from .theplatform import ThePlatformIE, default_ns
from .adobepass import AdobePassIE
from ..compat import compat_urllib_parse_unquote
from ..networking import HEADRequest
from ..utils import (
Expand Down
18 changes: 9 additions & 9 deletions yt_dlp/extractor/neteasemusic.py
Expand Up @@ -56,23 +56,23 @@ def make_player_api_request_data_and_headers(self, song_id, bitrate):
'__csrf': '',
'os': 'pc',
'channel': None,
'requestId': f'{now}_{rand:04}',
'requestId': '{0}_{1:04}'.format(now, rand),
}
request_text = json.dumps(
{'ids': f'[{song_id}]', 'br': bitrate, 'header': cookie},
{'ids': '[{0}]'.format(song_id), 'br': bitrate, 'header': cookie},
separators=(',', ':'))
message = 'nobody{}use{}md5forencrypt'.format(
message = 'nobody{0}use{1}md5forencrypt'.format(
URL, request_text).encode('latin1')
msg_digest = md5(message).hexdigest()

data = '{}-36cd479b6b5-{}-36cd479b6b5-{}'.format(
data = '{0}-36cd479b6b5-{1}-36cd479b6b5-{2}'.format(
URL, request_text, msg_digest)
data = pkcs7_padding(bytes_to_intlist(data))
encrypted = intlist_to_bytes(aes_ecb_encrypt(data, bytes_to_intlist(KEY)))
encrypted_params = hexlify(encrypted).decode('ascii').upper()

cookie = '; '.join(
['{}={}'.format(k, v if v is not None else 'undefined')
['{0}={1}'.format(k, v if v is not None else 'undefined')
for [k, v] in cookie.items()])

headers = {
Expand All @@ -81,7 +81,7 @@ def make_player_api_request_data_and_headers(self, song_id, bitrate):
'Referer': 'https://music.163.com',
'Cookie': cookie,
}
return (f'params={encrypted_params}', headers)
return ('params={0}'.format(encrypted_params), headers)

def _call_player_api(self, song_id, bitrate):
url = 'https://interface3.music.163.com/eapi/song/enhance/player/url'
Expand Down Expand Up @@ -215,9 +215,9 @@ def _process_lyrics(self, lyrics_info):

lyrics_expr = r'(\[[0-9]{2}:[0-9]{2}\.[0-9]{2,}\])([^\n]+)'
original_ts_texts = re.findall(lyrics_expr, original)
translation_ts_dict = {
time_stamp: text for time_stamp, text in re.findall(lyrics_expr, translated)
}
translation_ts_dict = dict(
(time_stamp, text) for time_stamp, text in re.findall(lyrics_expr, translated)
)
lyrics = '\n'.join([
'%s%s / %s' % (time_stamp, text, translation_ts_dict.get(time_stamp, ''))
for time_stamp, text in original_ts_texts
Expand Down
5 changes: 4 additions & 1 deletion yt_dlp/extractor/nowness.py
@@ -1,4 +1,7 @@
from .brightcove import BrightcoveLegacyIE, BrightcoveNewIE
from .brightcove import (
BrightcoveLegacyIE,
BrightcoveNewIE,
)
from .common import InfoExtractor
from ..compat import compat_str
from ..networking import Request
Expand Down
4 changes: 2 additions & 2 deletions yt_dlp/extractor/orf.py
Expand Up @@ -4,15 +4,15 @@
from .common import InfoExtractor
from ..networking import HEADRequest
from ..utils import (
InAdvancePagedList,
clean_html,
determine_ext,
float_or_none,
InAdvancePagedList,
int_or_none,
join_nonempty,
make_archive_id,
orderedSet,
remove_end,
make_archive_id,
smuggle_url,
strip_jsonp,
try_call,
Expand Down
9 changes: 5 additions & 4 deletions yt_dlp/extractor/pornhub.py
Expand Up @@ -747,7 +747,7 @@ class PornHubPagedVideoListIE(PornHubPagedPlaylistBaseIE):
def suitable(cls, url):
return (False
if PornHubIE.suitable(url) or PornHubUserIE.suitable(url) or PornHubUserVideosUploadIE.suitable(url)
else super().suitable(url))
else super(PornHubPagedVideoListIE, cls).suitable(url))


class PornHubUserVideosUploadIE(PornHubPagedPlaylistBaseIE):
Expand Down Expand Up @@ -796,8 +796,8 @@ def _entries(self, url, host, item_id):
page_entries = self._extract_entries(webpage, host)

def download_page(page_num):
note = f'Downloading page {page_num}'
page_url = f'https://www.{host}/playlist/viewChunked'
note = 'Downloading page {}'.format(page_num)
page_url = 'https://www.{}/playlist/viewChunked'.format(host)
return self._download_webpage(page_url, item_id, note, query={
'id': playlist_id,
'page': page_num,
Expand All @@ -810,7 +810,8 @@ def download_page(page_num):
page_entries = self._extract_entries(webpage, host)
if not page_entries:
break
yield from page_entries
for e in page_entries:
yield e

def _real_extract(self, url):
mobj = self._match_valid_url(url)
Expand Down
14 changes: 8 additions & 6 deletions yt_dlp/extractor/soundcloud.py
@@ -1,17 +1,21 @@
import itertools
import json
import re
import json
# import random

from .common import InfoExtractor, SearchInfoExtractor
from .common import (
InfoExtractor,
SearchInfoExtractor
)
from ..compat import compat_str
from ..networking import HEADRequest, Request
from ..networking.exceptions import HTTPError
from ..utils import (
KNOWN_EXTENSIONS,
ExtractorError,
error_to_compat_str,
ExtractorError,
float_or_none,
int_or_none,
KNOWN_EXTENSIONS,
mimetype2ext,
parse_qs,
str_or_none,
Expand All @@ -22,8 +26,6 @@
urlhandle_detect_ext,
)

# import random


class SoundcloudEmbedIE(InfoExtractor):
_VALID_URL = r'https?://(?:w|player|p)\.soundcloud\.com/player/?.*?\burl=(?P<id>.+)'
Expand Down
2 changes: 1 addition & 1 deletion yt_dlp/extractor/tenplay.py
@@ -1,5 +1,5 @@
import base64
from datetime import datetime
import base64

from .common import InfoExtractor
from ..networking import HEADRequest
Expand Down
15 changes: 8 additions & 7 deletions yt_dlp/extractor/theplatform.py
@@ -1,23 +1,24 @@
import binascii
import hashlib
import hmac
import re
import time
import hmac
import binascii
import hashlib


from .adobepass import AdobePassIE
from .once import OnceIE
from .adobepass import AdobePassIE
from ..networking import Request
from ..utils import (
ExtractorError,
determine_ext,
find_xpath_attr,
ExtractorError,
float_or_none,
int_or_none,
mimetype2ext,
parse_qs,
unsmuggle_url,
update_url_query,
xpath_with_ns,
mimetype2ext,
find_xpath_attr,
)

default_ns = 'http://www.w3.org/2005/SMIL21/Language'
Expand Down
2 changes: 1 addition & 1 deletion yt_dlp/extractor/triller.py
Expand Up @@ -14,8 +14,8 @@
traverse_obj,
unified_timestamp,
url_basename,
url_or_none,
urljoin,
url_or_none,
)


Expand Down
6 changes: 3 additions & 3 deletions yt_dlp/extractor/udemy.py
Expand Up @@ -126,7 +126,7 @@ def _download_webpage_handle(self, *args, **kwargs):
headers = kwargs.get('headers', {}).copy()
headers['User-Agent'] = 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/72.0.3626.109 Safari/537.36'
kwargs['headers'] = headers
ret = super()._download_webpage_handle(
ret = super(UdemyIE, self)._download_webpage_handle(
*args, **kwargs)
if not ret:
return ret
Expand Down Expand Up @@ -158,7 +158,7 @@ def _download_json(self, url_or_request, *args, **kwargs):
else:
url_or_request = Request(url_or_request, headers=headers)

response = super()._download_json(url_or_request, *args, **kwargs)
response = super(UdemyIE, self)._download_json(url_or_request, *args, **kwargs)
self._handle_error(response)
return response

Expand Down Expand Up @@ -423,7 +423,7 @@ class UdemyCourseIE(UdemyIE): # XXX: Do not subclass from concrete IE

@classmethod
def suitable(cls, url):
return False if UdemyIE.suitable(url) else super().suitable(url)
return False if UdemyIE.suitable(url) else super(UdemyCourseIE, cls).suitable(url)

def _real_extract(self, url):
course_path = self._match_id(url)
Expand Down
2 changes: 1 addition & 1 deletion yt_dlp/extractor/vidlii.py
Expand Up @@ -3,8 +3,8 @@
from .common import InfoExtractor
from ..networking import HEADRequest
from ..utils import (
float_or_none,
format_field,
float_or_none,
get_element_by_id,
int_or_none,
str_to_int,
Expand Down
10 changes: 5 additions & 5 deletions yt_dlp/extractor/vimeo.py
@@ -1,21 +1,21 @@
import base64
import functools
import itertools
import re
import itertools

from .common import InfoExtractor
from ..compat import compat_str, compat_urlparse
from ..networking import HEADRequest, Request
from ..networking.exceptions import HTTPError
from ..utils import (
ExtractorError,
OnDemandPagedList,
clean_html,
determine_ext,
ExtractorError,
get_element_by_class,
int_or_none,
js_to_json,
int_or_none,
merge_dicts,
OnDemandPagedList,
parse_filesize,
parse_iso8601,
parse_qs,
Expand All @@ -25,8 +25,8 @@
unified_timestamp,
unsmuggle_url,
urlencode_postdata,
urlhandle_detect_ext,
urljoin,
urlhandle_detect_ext,
)


Expand Down
2 changes: 1 addition & 1 deletion yt_dlp/extractor/youtube.py
Expand Up @@ -6482,7 +6482,7 @@ def _real_extract(self, url, smuggled_data):

# Handle both video/playlist URLs
qs = parse_qs(url)
video_id, playlist_id = (traverse_obj(qs, (key, 0)) for key in ('v', 'list'))
video_id, playlist_id = [traverse_obj(qs, (key, 0)) for key in ('v', 'list')]
if not video_id and mobj['not_channel'].startswith('watch'):
if not playlist_id:
# If there is neither video or playlist ids, youtube redirects to home page, which is undesirable
Expand Down

0 comments on commit 591f915

Please sign in to comment.