mirror of
https://github.com/ytdl-org/youtube-dl
synced 2025-07-14 15:34:15 +09:00
Compare commits
20 Commits
0c9fc6ba51
...
f51533b3b7
Author | SHA1 | Date | |
---|---|---|---|
![]() |
f51533b3b7 | ||
![]() |
da7223d4aa | ||
![]() |
37c2440d6a | ||
![]() |
420d53387c | ||
![]() |
32f89de92b | ||
![]() |
283dca56fe | ||
![]() |
422b1b31cf | ||
![]() |
1dc27e1c3b | ||
![]() |
af049e309b | ||
![]() |
94849bc997 | ||
![]() |
974c7d7f34 | ||
![]() |
8738407d77 | ||
![]() |
cecaa18b80 | ||
![]() |
673277e510 | ||
![]() |
91b1569f68 | ||
![]() |
72db217289 | ||
![]() |
fc933e686b | ||
![]() |
ea02c40539 | ||
![]() |
7270ecf3d6 | ||
![]() |
dade9111f1 |
@ -11,6 +11,7 @@ sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
|||||||
|
|
||||||
import math
|
import math
|
||||||
import re
|
import re
|
||||||
|
import time
|
||||||
|
|
||||||
from youtube_dl.compat import compat_str as str
|
from youtube_dl.compat import compat_str as str
|
||||||
from youtube_dl.jsinterp import JS_Undefined, JSInterpreter
|
from youtube_dl.jsinterp import JS_Undefined, JSInterpreter
|
||||||
@ -208,6 +209,34 @@ class TestJSInterpreter(unittest.TestCase):
|
|||||||
self._test(jsi, 86000, args=['12/31/1969 18:01:26 MDT'])
|
self._test(jsi, 86000, args=['12/31/1969 18:01:26 MDT'])
|
||||||
# epoch 0
|
# epoch 0
|
||||||
self._test(jsi, 0, args=['1 January 1970 00:00:00 UTC'])
|
self._test(jsi, 0, args=['1 January 1970 00:00:00 UTC'])
|
||||||
|
# undefined
|
||||||
|
self._test(jsi, NaN, args=[JS_Undefined])
|
||||||
|
# y,m,d, ... - may fail with older dates lacking DST data
|
||||||
|
jsi = JSInterpreter(
|
||||||
|
'function f() { return new Date(%s); }'
|
||||||
|
% ('2024, 5, 29, 2, 52, 12, 42',))
|
||||||
|
self._test(jsi, (
|
||||||
|
1719625932042 # UK value
|
||||||
|
+ (
|
||||||
|
+ 3600 # back to GMT
|
||||||
|
+ (time.altzone if time.daylight # host's DST
|
||||||
|
else time.timezone)
|
||||||
|
) * 1000))
|
||||||
|
# no arg
|
||||||
|
self.assertAlmostEqual(JSInterpreter(
|
||||||
|
'function f() { return new Date() - 0; }').call_function('f'),
|
||||||
|
time.time() * 1000, delta=100)
|
||||||
|
# Date.now()
|
||||||
|
self.assertAlmostEqual(JSInterpreter(
|
||||||
|
'function f() { return Date.now(); }').call_function('f'),
|
||||||
|
time.time() * 1000, delta=100)
|
||||||
|
# Date.parse()
|
||||||
|
jsi = JSInterpreter('function f(dt) { return Date.parse(dt); }')
|
||||||
|
self._test(jsi, 0, args=['1 January 1970 00:00:00 UTC'])
|
||||||
|
# Date.UTC()
|
||||||
|
jsi = JSInterpreter('function f() { return Date.UTC(%s); }'
|
||||||
|
% ('1970, 0, 1, 0, 0, 0, 0',))
|
||||||
|
self._test(jsi, 0)
|
||||||
|
|
||||||
def test_call(self):
|
def test_call(self):
|
||||||
jsi = JSInterpreter('''
|
jsi = JSInterpreter('''
|
||||||
@ -463,6 +492,14 @@ class TestJSInterpreter(unittest.TestCase):
|
|||||||
self._test('function f(){return NaN << 42}', 0)
|
self._test('function f(){return NaN << 42}', 0)
|
||||||
self._test('function f(){return "21.9" << 1}', 42)
|
self._test('function f(){return "21.9" << 1}', 42)
|
||||||
self._test('function f(){return 21 << 4294967297}', 42)
|
self._test('function f(){return 21 << 4294967297}', 42)
|
||||||
|
self._test('function f(){return true << "5";}', 32)
|
||||||
|
self._test('function f(){return true << true;}', 2)
|
||||||
|
self._test('function f(){return "19" & "21.9";}', 17)
|
||||||
|
self._test('function f(){return "19" & false;}', 0)
|
||||||
|
self._test('function f(){return "11.0" >> "2.1";}', 2)
|
||||||
|
self._test('function f(){return 5 ^ 9;}', 12)
|
||||||
|
self._test('function f(){return 0.0 << NaN}', 0)
|
||||||
|
self._test('function f(){return null << undefined}', 0)
|
||||||
|
|
||||||
def test_negative(self):
|
def test_negative(self):
|
||||||
self._test('function f(){return 2 * -2.0 ;}', -4)
|
self._test('function f(){return 2 * -2.0 ;}', -4)
|
||||||
|
@ -223,6 +223,42 @@ _NSIG_TESTS = [
|
|||||||
'https://www.youtube.com/s/player/9c6dfc4a/player_ias.vflset/en_US/base.js',
|
'https://www.youtube.com/s/player/9c6dfc4a/player_ias.vflset/en_US/base.js',
|
||||||
'jbu7ylIosQHyJyJV', 'uwI0ESiynAmhNg',
|
'jbu7ylIosQHyJyJV', 'uwI0ESiynAmhNg',
|
||||||
),
|
),
|
||||||
|
(
|
||||||
|
'https://www.youtube.com/s/player/f6e09c70/player_ias.vflset/en_US/base.js',
|
||||||
|
'W9HJZKktxuYoDTqW', 'jHbbkcaxm54',
|
||||||
|
),
|
||||||
|
(
|
||||||
|
'https://www.youtube.com/s/player/f6e09c70/player_ias_tce.vflset/en_US/base.js',
|
||||||
|
'W9HJZKktxuYoDTqW', 'jHbbkcaxm54',
|
||||||
|
),
|
||||||
|
(
|
||||||
|
'https://www.youtube.com/s/player/643afba4/player_ias.vflset/en_US/base.js',
|
||||||
|
'W9HJZKktxuYoDTqW', 'larxUlagTRAcSw',
|
||||||
|
),
|
||||||
|
(
|
||||||
|
'https://www.youtube.com/s/player/e7567ecf/player_ias_tce.vflset/en_US/base.js',
|
||||||
|
'Sy4aDGc0VpYRR9ew_', '5UPOT1VhoZxNLQ',
|
||||||
|
),
|
||||||
|
(
|
||||||
|
'https://www.youtube.com/s/player/d50f54ef/player_ias_tce.vflset/en_US/base.js',
|
||||||
|
'Ha7507LzRmH3Utygtj', 'XFTb2HoeOE5MHg',
|
||||||
|
),
|
||||||
|
(
|
||||||
|
'https://www.youtube.com/s/player/074a8365/player_ias_tce.vflset/en_US/base.js',
|
||||||
|
'Ha7507LzRmH3Utygtj', 'ufTsrE0IVYrkl8v',
|
||||||
|
),
|
||||||
|
(
|
||||||
|
'https://www.youtube.com/s/player/643afba4/player_ias.vflset/en_US/base.js',
|
||||||
|
'N5uAlLqm0eg1GyHO', 'dCBQOejdq5s-ww',
|
||||||
|
),
|
||||||
|
(
|
||||||
|
'https://www.youtube.com/s/player/69f581a5/tv-player-ias.vflset/tv-player-ias.js',
|
||||||
|
'-qIP447rVlTTwaZjY', 'KNcGOksBAvwqQg',
|
||||||
|
),
|
||||||
|
(
|
||||||
|
'https://www.youtube.com/s/player/643afba4/tv-player-ias.vflset/tv-player-ias.js',
|
||||||
|
'ir9-V6cdbCiyKxhr', '2PL7ZDYAALMfmA',
|
||||||
|
),
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
@ -284,7 +320,7 @@ def t_factory(name, sig_func, url_pattern):
|
|||||||
|
|
||||||
|
|
||||||
def signature(jscode, sig_input):
|
def signature(jscode, sig_input):
|
||||||
func = YoutubeIE(FakeYDL())._parse_sig_js(jscode)
|
func = YoutubeIE(FakeYDL({'cachedir': False}))._parse_sig_js(jscode)
|
||||||
src_sig = (
|
src_sig = (
|
||||||
compat_str(string.printable[:sig_input])
|
compat_str(string.printable[:sig_input])
|
||||||
if isinstance(sig_input, int) else sig_input)
|
if isinstance(sig_input, int) else sig_input)
|
||||||
@ -292,9 +328,10 @@ def signature(jscode, sig_input):
|
|||||||
|
|
||||||
|
|
||||||
def n_sig(jscode, sig_input):
|
def n_sig(jscode, sig_input):
|
||||||
funcname = YoutubeIE(FakeYDL())._extract_n_function_name(jscode)
|
ie = YoutubeIE(FakeYDL({'cachedir': False}))
|
||||||
return JSInterpreter(jscode).call_function(
|
jsi = JSInterpreter(jscode)
|
||||||
funcname, sig_input, _ytdl_do_not_return=sig_input)
|
jsi, _, func_code = ie._extract_n_function_code_jsi(sig_input, jsi)
|
||||||
|
return ie._extract_n_function_from_code(jsi, func_code)(sig_input)
|
||||||
|
|
||||||
|
|
||||||
make_sig_test = t_factory(
|
make_sig_test = t_factory(
|
||||||
|
@ -18,7 +18,7 @@ from .compat import (
|
|||||||
compat_getpass,
|
compat_getpass,
|
||||||
compat_register_utf8,
|
compat_register_utf8,
|
||||||
compat_shlex_split,
|
compat_shlex_split,
|
||||||
workaround_optparse_bug9161,
|
_workaround_optparse_bug9161,
|
||||||
)
|
)
|
||||||
from .utils import (
|
from .utils import (
|
||||||
_UnsafeExtensionError,
|
_UnsafeExtensionError,
|
||||||
@ -50,7 +50,7 @@ def _real_main(argv=None):
|
|||||||
# Compatibility fix for Windows
|
# Compatibility fix for Windows
|
||||||
compat_register_utf8()
|
compat_register_utf8()
|
||||||
|
|
||||||
workaround_optparse_bug9161()
|
_workaround_optparse_bug9161()
|
||||||
|
|
||||||
setproctitle('youtube-dl')
|
setproctitle('youtube-dl')
|
||||||
|
|
||||||
|
@ -16,7 +16,6 @@ import os
|
|||||||
import platform
|
import platform
|
||||||
import re
|
import re
|
||||||
import shlex
|
import shlex
|
||||||
import shutil
|
|
||||||
import socket
|
import socket
|
||||||
import struct
|
import struct
|
||||||
import subprocess
|
import subprocess
|
||||||
@ -24,11 +23,15 @@ import sys
|
|||||||
import types
|
import types
|
||||||
import xml.etree.ElementTree
|
import xml.etree.ElementTree
|
||||||
|
|
||||||
|
_IDENTITY = lambda x: x
|
||||||
|
|
||||||
# naming convention
|
# naming convention
|
||||||
# 'compat_' + Python3_name.replace('.', '_')
|
# 'compat_' + Python3_name.replace('.', '_')
|
||||||
# other aliases exist for convenience and/or legacy
|
# other aliases exist for convenience and/or legacy
|
||||||
|
# wrap disposable test values in type() to reclaim storage
|
||||||
|
|
||||||
# deal with critical unicode/str things first
|
# deal with critical unicode/str things first:
|
||||||
|
# compat_str, compat_basestring, compat_chr
|
||||||
try:
|
try:
|
||||||
# Python 2
|
# Python 2
|
||||||
compat_str, compat_basestring, compat_chr = (
|
compat_str, compat_basestring, compat_chr = (
|
||||||
@ -39,18 +42,23 @@ except NameError:
|
|||||||
str, (str, bytes), chr
|
str, (str, bytes), chr
|
||||||
)
|
)
|
||||||
|
|
||||||
# casefold
|
|
||||||
|
# compat_casefold
|
||||||
try:
|
try:
|
||||||
compat_str.casefold
|
compat_str.casefold
|
||||||
compat_casefold = lambda s: s.casefold()
|
compat_casefold = lambda s: s.casefold()
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
from .casefold import _casefold as compat_casefold
|
from .casefold import _casefold as compat_casefold
|
||||||
|
|
||||||
|
|
||||||
|
# compat_collections_abc
|
||||||
try:
|
try:
|
||||||
import collections.abc as compat_collections_abc
|
import collections.abc as compat_collections_abc
|
||||||
except ImportError:
|
except ImportError:
|
||||||
import collections as compat_collections_abc
|
import collections as compat_collections_abc
|
||||||
|
|
||||||
|
|
||||||
|
# compat_urllib_request
|
||||||
try:
|
try:
|
||||||
import urllib.request as compat_urllib_request
|
import urllib.request as compat_urllib_request
|
||||||
except ImportError: # Python 2
|
except ImportError: # Python 2
|
||||||
@ -79,11 +87,15 @@ except TypeError:
|
|||||||
_add_init_method_arg(compat_urllib_request.Request)
|
_add_init_method_arg(compat_urllib_request.Request)
|
||||||
del _add_init_method_arg
|
del _add_init_method_arg
|
||||||
|
|
||||||
|
|
||||||
|
# compat_urllib_error
|
||||||
try:
|
try:
|
||||||
import urllib.error as compat_urllib_error
|
import urllib.error as compat_urllib_error
|
||||||
except ImportError: # Python 2
|
except ImportError: # Python 2
|
||||||
import urllib2 as compat_urllib_error
|
import urllib2 as compat_urllib_error
|
||||||
|
|
||||||
|
|
||||||
|
# compat_urllib_parse
|
||||||
try:
|
try:
|
||||||
import urllib.parse as compat_urllib_parse
|
import urllib.parse as compat_urllib_parse
|
||||||
except ImportError: # Python 2
|
except ImportError: # Python 2
|
||||||
@ -98,17 +110,23 @@ except ImportError: # Python 2
|
|||||||
compat_urlparse = compat_urllib_parse
|
compat_urlparse = compat_urllib_parse
|
||||||
compat_urllib_parse_urlparse = compat_urllib_parse.urlparse
|
compat_urllib_parse_urlparse = compat_urllib_parse.urlparse
|
||||||
|
|
||||||
|
|
||||||
|
# compat_urllib_response
|
||||||
try:
|
try:
|
||||||
import urllib.response as compat_urllib_response
|
import urllib.response as compat_urllib_response
|
||||||
except ImportError: # Python 2
|
except ImportError: # Python 2
|
||||||
import urllib as compat_urllib_response
|
import urllib as compat_urllib_response
|
||||||
|
|
||||||
|
|
||||||
|
# compat_urllib_response.addinfourl
|
||||||
try:
|
try:
|
||||||
compat_urllib_response.addinfourl.status
|
compat_urllib_response.addinfourl.status
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
# .getcode() is deprecated in Py 3.
|
# .getcode() is deprecated in Py 3.
|
||||||
compat_urllib_response.addinfourl.status = property(lambda self: self.getcode())
|
compat_urllib_response.addinfourl.status = property(lambda self: self.getcode())
|
||||||
|
|
||||||
|
|
||||||
|
# compat_http_cookiejar
|
||||||
try:
|
try:
|
||||||
import http.cookiejar as compat_cookiejar
|
import http.cookiejar as compat_cookiejar
|
||||||
except ImportError: # Python 2
|
except ImportError: # Python 2
|
||||||
@ -127,12 +145,16 @@ else:
|
|||||||
compat_cookiejar_Cookie = compat_cookiejar.Cookie
|
compat_cookiejar_Cookie = compat_cookiejar.Cookie
|
||||||
compat_http_cookiejar_Cookie = compat_cookiejar_Cookie
|
compat_http_cookiejar_Cookie = compat_cookiejar_Cookie
|
||||||
|
|
||||||
|
|
||||||
|
# compat_http_cookies
|
||||||
try:
|
try:
|
||||||
import http.cookies as compat_cookies
|
import http.cookies as compat_cookies
|
||||||
except ImportError: # Python 2
|
except ImportError: # Python 2
|
||||||
import Cookie as compat_cookies
|
import Cookie as compat_cookies
|
||||||
compat_http_cookies = compat_cookies
|
compat_http_cookies = compat_cookies
|
||||||
|
|
||||||
|
|
||||||
|
# compat_http_cookies_SimpleCookie
|
||||||
if sys.version_info[0] == 2 or sys.version_info < (3, 3):
|
if sys.version_info[0] == 2 or sys.version_info < (3, 3):
|
||||||
class compat_cookies_SimpleCookie(compat_cookies.SimpleCookie):
|
class compat_cookies_SimpleCookie(compat_cookies.SimpleCookie):
|
||||||
def load(self, rawdata):
|
def load(self, rawdata):
|
||||||
@ -155,11 +177,15 @@ else:
|
|||||||
compat_cookies_SimpleCookie = compat_cookies.SimpleCookie
|
compat_cookies_SimpleCookie = compat_cookies.SimpleCookie
|
||||||
compat_http_cookies_SimpleCookie = compat_cookies_SimpleCookie
|
compat_http_cookies_SimpleCookie = compat_cookies_SimpleCookie
|
||||||
|
|
||||||
|
|
||||||
|
# compat_html_entities, probably useless now
|
||||||
try:
|
try:
|
||||||
import html.entities as compat_html_entities
|
import html.entities as compat_html_entities
|
||||||
except ImportError: # Python 2
|
except ImportError: # Python 2
|
||||||
import htmlentitydefs as compat_html_entities
|
import htmlentitydefs as compat_html_entities
|
||||||
|
|
||||||
|
|
||||||
|
# compat_html_entities_html5
|
||||||
try: # Python >= 3.3
|
try: # Python >= 3.3
|
||||||
compat_html_entities_html5 = compat_html_entities.html5
|
compat_html_entities_html5 = compat_html_entities.html5
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
@ -2408,18 +2434,24 @@ except AttributeError:
|
|||||||
# Py < 3.1
|
# Py < 3.1
|
||||||
compat_http_client.HTTPResponse.getcode = lambda self: self.status
|
compat_http_client.HTTPResponse.getcode = lambda self: self.status
|
||||||
|
|
||||||
|
|
||||||
|
# compat_urllib_HTTPError
|
||||||
try:
|
try:
|
||||||
from urllib.error import HTTPError as compat_HTTPError
|
from urllib.error import HTTPError as compat_HTTPError
|
||||||
except ImportError: # Python 2
|
except ImportError: # Python 2
|
||||||
from urllib2 import HTTPError as compat_HTTPError
|
from urllib2 import HTTPError as compat_HTTPError
|
||||||
compat_urllib_HTTPError = compat_HTTPError
|
compat_urllib_HTTPError = compat_HTTPError
|
||||||
|
|
||||||
|
|
||||||
|
# compat_urllib_request_urlretrieve
|
||||||
try:
|
try:
|
||||||
from urllib.request import urlretrieve as compat_urlretrieve
|
from urllib.request import urlretrieve as compat_urlretrieve
|
||||||
except ImportError: # Python 2
|
except ImportError: # Python 2
|
||||||
from urllib import urlretrieve as compat_urlretrieve
|
from urllib import urlretrieve as compat_urlretrieve
|
||||||
compat_urllib_request_urlretrieve = compat_urlretrieve
|
compat_urllib_request_urlretrieve = compat_urlretrieve
|
||||||
|
|
||||||
|
|
||||||
|
# compat_html_parser_HTMLParser, compat_html_parser_HTMLParseError
|
||||||
try:
|
try:
|
||||||
from HTMLParser import (
|
from HTMLParser import (
|
||||||
HTMLParser as compat_HTMLParser,
|
HTMLParser as compat_HTMLParser,
|
||||||
@ -2432,22 +2464,33 @@ except ImportError: # Python 3
|
|||||||
# HTMLParseError was deprecated in Python 3.3 and removed in
|
# HTMLParseError was deprecated in Python 3.3 and removed in
|
||||||
# Python 3.5. Introducing dummy exception for Python >3.5 for compatible
|
# Python 3.5. Introducing dummy exception for Python >3.5 for compatible
|
||||||
# and uniform cross-version exception handling
|
# and uniform cross-version exception handling
|
||||||
|
|
||||||
class compat_HTMLParseError(Exception):
|
class compat_HTMLParseError(Exception):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
compat_html_parser_HTMLParser = compat_HTMLParser
|
compat_html_parser_HTMLParser = compat_HTMLParser
|
||||||
compat_html_parser_HTMLParseError = compat_HTMLParseError
|
compat_html_parser_HTMLParseError = compat_HTMLParseError
|
||||||
|
|
||||||
|
|
||||||
|
# compat_subprocess_get_DEVNULL
|
||||||
try:
|
try:
|
||||||
_DEVNULL = subprocess.DEVNULL
|
_DEVNULL = subprocess.DEVNULL
|
||||||
compat_subprocess_get_DEVNULL = lambda: _DEVNULL
|
compat_subprocess_get_DEVNULL = lambda: _DEVNULL
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
compat_subprocess_get_DEVNULL = lambda: open(os.path.devnull, 'w')
|
compat_subprocess_get_DEVNULL = lambda: open(os.path.devnull, 'w')
|
||||||
|
|
||||||
|
|
||||||
|
# compat_http_server
|
||||||
try:
|
try:
|
||||||
import http.server as compat_http_server
|
import http.server as compat_http_server
|
||||||
except ImportError:
|
except ImportError:
|
||||||
import BaseHTTPServer as compat_http_server
|
import BaseHTTPServer as compat_http_server
|
||||||
|
|
||||||
|
|
||||||
|
# compat_urllib_parse_unquote_to_bytes,
|
||||||
|
# compat_urllib_parse_unquote, compat_urllib_parse_unquote_plus,
|
||||||
|
# compat_urllib_parse_urlencode,
|
||||||
|
# compat_urllib_parse_parse_qs
|
||||||
try:
|
try:
|
||||||
from urllib.parse import unquote_to_bytes as compat_urllib_parse_unquote_to_bytes
|
from urllib.parse import unquote_to_bytes as compat_urllib_parse_unquote_to_bytes
|
||||||
from urllib.parse import unquote as compat_urllib_parse_unquote
|
from urllib.parse import unquote as compat_urllib_parse_unquote
|
||||||
@ -2598,6 +2641,8 @@ except ImportError: # Python 2
|
|||||||
|
|
||||||
compat_urllib_parse_parse_qs = compat_parse_qs
|
compat_urllib_parse_parse_qs = compat_parse_qs
|
||||||
|
|
||||||
|
|
||||||
|
# compat_urllib_request_DataHandler
|
||||||
try:
|
try:
|
||||||
from urllib.request import DataHandler as compat_urllib_request_DataHandler
|
from urllib.request import DataHandler as compat_urllib_request_DataHandler
|
||||||
except ImportError: # Python < 3.4
|
except ImportError: # Python < 3.4
|
||||||
@ -2632,16 +2677,20 @@ except ImportError: # Python < 3.4
|
|||||||
|
|
||||||
return compat_urllib_response.addinfourl(io.BytesIO(data), headers, url)
|
return compat_urllib_response.addinfourl(io.BytesIO(data), headers, url)
|
||||||
|
|
||||||
|
|
||||||
|
# compat_xml_etree_ElementTree_ParseError
|
||||||
try:
|
try:
|
||||||
from xml.etree.ElementTree import ParseError as compat_xml_parse_error
|
from xml.etree.ElementTree import ParseError as compat_xml_parse_error
|
||||||
except ImportError: # Python 2.6
|
except ImportError: # Python 2.6
|
||||||
from xml.parsers.expat import ExpatError as compat_xml_parse_error
|
from xml.parsers.expat import ExpatError as compat_xml_parse_error
|
||||||
compat_xml_etree_ElementTree_ParseError = compat_xml_parse_error
|
compat_xml_etree_ElementTree_ParseError = compat_xml_parse_error
|
||||||
|
|
||||||
etree = xml.etree.ElementTree
|
|
||||||
|
# compat_xml_etree_ElementTree_Element
|
||||||
|
_etree = xml.etree.ElementTree
|
||||||
|
|
||||||
|
|
||||||
class _TreeBuilder(etree.TreeBuilder):
|
class _TreeBuilder(_etree.TreeBuilder):
|
||||||
def doctype(self, name, pubid, system):
|
def doctype(self, name, pubid, system):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@ -2650,7 +2699,7 @@ try:
|
|||||||
# xml.etree.ElementTree.Element is a method in Python <=2.6 and
|
# xml.etree.ElementTree.Element is a method in Python <=2.6 and
|
||||||
# the following will crash with:
|
# the following will crash with:
|
||||||
# TypeError: isinstance() arg 2 must be a class, type, or tuple of classes and types
|
# TypeError: isinstance() arg 2 must be a class, type, or tuple of classes and types
|
||||||
isinstance(None, etree.Element)
|
isinstance(None, _etree.Element)
|
||||||
from xml.etree.ElementTree import Element as compat_etree_Element
|
from xml.etree.ElementTree import Element as compat_etree_Element
|
||||||
except TypeError: # Python <=2.6
|
except TypeError: # Python <=2.6
|
||||||
from xml.etree.ElementTree import _ElementInterface as compat_etree_Element
|
from xml.etree.ElementTree import _ElementInterface as compat_etree_Element
|
||||||
@ -2658,12 +2707,12 @@ compat_xml_etree_ElementTree_Element = compat_etree_Element
|
|||||||
|
|
||||||
if sys.version_info[0] >= 3:
|
if sys.version_info[0] >= 3:
|
||||||
def compat_etree_fromstring(text):
|
def compat_etree_fromstring(text):
|
||||||
return etree.XML(text, parser=etree.XMLParser(target=_TreeBuilder()))
|
return _etree.XML(text, parser=_etree.XMLParser(target=_TreeBuilder()))
|
||||||
else:
|
else:
|
||||||
# python 2.x tries to encode unicode strings with ascii (see the
|
# python 2.x tries to encode unicode strings with ascii (see the
|
||||||
# XMLParser._fixtext method)
|
# XMLParser._fixtext method)
|
||||||
try:
|
try:
|
||||||
_etree_iter = etree.Element.iter
|
_etree_iter = _etree.Element.iter
|
||||||
except AttributeError: # Python <=2.6
|
except AttributeError: # Python <=2.6
|
||||||
def _etree_iter(root):
|
def _etree_iter(root):
|
||||||
for el in root.findall('*'):
|
for el in root.findall('*'):
|
||||||
@ -2675,27 +2724,29 @@ else:
|
|||||||
# 2.7 source
|
# 2.7 source
|
||||||
def _XML(text, parser=None):
|
def _XML(text, parser=None):
|
||||||
if not parser:
|
if not parser:
|
||||||
parser = etree.XMLParser(target=_TreeBuilder())
|
parser = _etree.XMLParser(target=_TreeBuilder())
|
||||||
parser.feed(text)
|
parser.feed(text)
|
||||||
return parser.close()
|
return parser.close()
|
||||||
|
|
||||||
def _element_factory(*args, **kwargs):
|
def _element_factory(*args, **kwargs):
|
||||||
el = etree.Element(*args, **kwargs)
|
el = _etree.Element(*args, **kwargs)
|
||||||
for k, v in el.items():
|
for k, v in el.items():
|
||||||
if isinstance(v, bytes):
|
if isinstance(v, bytes):
|
||||||
el.set(k, v.decode('utf-8'))
|
el.set(k, v.decode('utf-8'))
|
||||||
return el
|
return el
|
||||||
|
|
||||||
def compat_etree_fromstring(text):
|
def compat_etree_fromstring(text):
|
||||||
doc = _XML(text, parser=etree.XMLParser(target=_TreeBuilder(element_factory=_element_factory)))
|
doc = _XML(text, parser=_etree.XMLParser(target=_TreeBuilder(element_factory=_element_factory)))
|
||||||
for el in _etree_iter(doc):
|
for el in _etree_iter(doc):
|
||||||
if el.text is not None and isinstance(el.text, bytes):
|
if el.text is not None and isinstance(el.text, bytes):
|
||||||
el.text = el.text.decode('utf-8')
|
el.text = el.text.decode('utf-8')
|
||||||
return doc
|
return doc
|
||||||
|
|
||||||
if hasattr(etree, 'register_namespace'):
|
|
||||||
compat_etree_register_namespace = etree.register_namespace
|
# compat_xml_etree_register_namespace
|
||||||
else:
|
try:
|
||||||
|
compat_etree_register_namespace = _etree.register_namespace
|
||||||
|
except AttributeError:
|
||||||
def compat_etree_register_namespace(prefix, uri):
|
def compat_etree_register_namespace(prefix, uri):
|
||||||
"""Register a namespace prefix.
|
"""Register a namespace prefix.
|
||||||
The registry is global, and any existing mapping for either the
|
The registry is global, and any existing mapping for either the
|
||||||
@ -2704,14 +2755,16 @@ else:
|
|||||||
attributes in this namespace will be serialized with prefix if possible.
|
attributes in this namespace will be serialized with prefix if possible.
|
||||||
ValueError is raised if prefix is reserved or is invalid.
|
ValueError is raised if prefix is reserved or is invalid.
|
||||||
"""
|
"""
|
||||||
if re.match(r"ns\d+$", prefix):
|
if re.match(r'ns\d+$', prefix):
|
||||||
raise ValueError("Prefix format reserved for internal use")
|
raise ValueError('Prefix format reserved for internal use')
|
||||||
for k, v in list(etree._namespace_map.items()):
|
for k, v in list(_etree._namespace_map.items()):
|
||||||
if k == uri or v == prefix:
|
if k == uri or v == prefix:
|
||||||
del etree._namespace_map[k]
|
del _etree._namespace_map[k]
|
||||||
etree._namespace_map[uri] = prefix
|
_etree._namespace_map[uri] = prefix
|
||||||
compat_xml_etree_register_namespace = compat_etree_register_namespace
|
compat_xml_etree_register_namespace = compat_etree_register_namespace
|
||||||
|
|
||||||
|
|
||||||
|
# compat_xpath, compat_etree_iterfind
|
||||||
if sys.version_info < (2, 7):
|
if sys.version_info < (2, 7):
|
||||||
# Here comes the crazy part: In 2.6, if the xpath is a unicode,
|
# Here comes the crazy part: In 2.6, if the xpath is a unicode,
|
||||||
# .//node does not match if a node is a direct child of . !
|
# .//node does not match if a node is a direct child of . !
|
||||||
@ -2898,7 +2951,6 @@ if sys.version_info < (2, 7):
|
|||||||
def __init__(self, root):
|
def __init__(self, root):
|
||||||
self.root = root
|
self.root = root
|
||||||
|
|
||||||
##
|
|
||||||
# Generate all matching objects.
|
# Generate all matching objects.
|
||||||
|
|
||||||
def compat_etree_iterfind(elem, path, namespaces=None):
|
def compat_etree_iterfind(elem, path, namespaces=None):
|
||||||
@ -2933,13 +2985,15 @@ if sys.version_info < (2, 7):
|
|||||||
|
|
||||||
|
|
||||||
else:
|
else:
|
||||||
compat_xpath = lambda xpath: xpath
|
|
||||||
compat_etree_iterfind = lambda element, match: element.iterfind(match)
|
compat_etree_iterfind = lambda element, match: element.iterfind(match)
|
||||||
|
compat_xpath = _IDENTITY
|
||||||
|
|
||||||
|
|
||||||
|
# compat_os_name
|
||||||
compat_os_name = os._name if os.name == 'java' else os.name
|
compat_os_name = os._name if os.name == 'java' else os.name
|
||||||
|
|
||||||
|
|
||||||
|
# compat_shlex_quote
|
||||||
if compat_os_name == 'nt':
|
if compat_os_name == 'nt':
|
||||||
def compat_shlex_quote(s):
|
def compat_shlex_quote(s):
|
||||||
return s if re.match(r'^[-_\w./]+$', s) else '"%s"' % s.replace('"', '\\"')
|
return s if re.match(r'^[-_\w./]+$', s) else '"%s"' % s.replace('"', '\\"')
|
||||||
@ -2954,6 +3008,7 @@ else:
|
|||||||
return "'" + s.replace("'", "'\"'\"'") + "'"
|
return "'" + s.replace("'", "'\"'\"'") + "'"
|
||||||
|
|
||||||
|
|
||||||
|
# compat_shlex.split
|
||||||
try:
|
try:
|
||||||
args = shlex.split('中文')
|
args = shlex.split('中文')
|
||||||
assert (isinstance(args, list)
|
assert (isinstance(args, list)
|
||||||
@ -2969,6 +3024,7 @@ except (AssertionError, UnicodeEncodeError):
|
|||||||
return list(map(lambda s: s.decode('utf-8'), shlex.split(s, comments, posix)))
|
return list(map(lambda s: s.decode('utf-8'), shlex.split(s, comments, posix)))
|
||||||
|
|
||||||
|
|
||||||
|
# compat_ord
|
||||||
def compat_ord(c):
|
def compat_ord(c):
|
||||||
if isinstance(c, int):
|
if isinstance(c, int):
|
||||||
return c
|
return c
|
||||||
@ -2976,6 +3032,7 @@ def compat_ord(c):
|
|||||||
return ord(c)
|
return ord(c)
|
||||||
|
|
||||||
|
|
||||||
|
# compat_getenv, compat_os_path_expanduser, compat_setenv
|
||||||
if sys.version_info >= (3, 0):
|
if sys.version_info >= (3, 0):
|
||||||
compat_getenv = os.getenv
|
compat_getenv = os.getenv
|
||||||
compat_expanduser = os.path.expanduser
|
compat_expanduser = os.path.expanduser
|
||||||
@ -3063,6 +3120,7 @@ else:
|
|||||||
compat_os_path_expanduser = compat_expanduser
|
compat_os_path_expanduser = compat_expanduser
|
||||||
|
|
||||||
|
|
||||||
|
# compat_os_path_realpath
|
||||||
if compat_os_name == 'nt' and sys.version_info < (3, 8):
|
if compat_os_name == 'nt' and sys.version_info < (3, 8):
|
||||||
# os.path.realpath on Windows does not follow symbolic links
|
# os.path.realpath on Windows does not follow symbolic links
|
||||||
# prior to Python 3.8 (see https://bugs.python.org/issue9949)
|
# prior to Python 3.8 (see https://bugs.python.org/issue9949)
|
||||||
@ -3076,6 +3134,7 @@ else:
|
|||||||
compat_os_path_realpath = compat_realpath
|
compat_os_path_realpath = compat_realpath
|
||||||
|
|
||||||
|
|
||||||
|
# compat_print
|
||||||
if sys.version_info < (3, 0):
|
if sys.version_info < (3, 0):
|
||||||
def compat_print(s):
|
def compat_print(s):
|
||||||
from .utils import preferredencoding
|
from .utils import preferredencoding
|
||||||
@ -3086,6 +3145,7 @@ else:
|
|||||||
print(s)
|
print(s)
|
||||||
|
|
||||||
|
|
||||||
|
# compat_getpass_getpass
|
||||||
if sys.version_info < (3, 0) and sys.platform == 'win32':
|
if sys.version_info < (3, 0) and sys.platform == 'win32':
|
||||||
def compat_getpass(prompt, *args, **kwargs):
|
def compat_getpass(prompt, *args, **kwargs):
|
||||||
if isinstance(prompt, compat_str):
|
if isinstance(prompt, compat_str):
|
||||||
@ -3098,22 +3158,22 @@ else:
|
|||||||
compat_getpass_getpass = compat_getpass
|
compat_getpass_getpass = compat_getpass
|
||||||
|
|
||||||
|
|
||||||
|
# compat_input
|
||||||
try:
|
try:
|
||||||
compat_input = raw_input
|
compat_input = raw_input
|
||||||
except NameError: # Python 3
|
except NameError: # Python 3
|
||||||
compat_input = input
|
compat_input = input
|
||||||
|
|
||||||
|
|
||||||
|
# compat_kwargs
|
||||||
# Python < 2.6.5 require kwargs to be bytes
|
# Python < 2.6.5 require kwargs to be bytes
|
||||||
try:
|
try:
|
||||||
def _testfunc(x):
|
(lambda x: x)(**{'x': 0})
|
||||||
pass
|
|
||||||
_testfunc(**{'x': 0})
|
|
||||||
except TypeError:
|
except TypeError:
|
||||||
def compat_kwargs(kwargs):
|
def compat_kwargs(kwargs):
|
||||||
return dict((bytes(k), v) for k, v in kwargs.items())
|
return dict((bytes(k), v) for k, v in kwargs.items())
|
||||||
else:
|
else:
|
||||||
compat_kwargs = lambda kwargs: kwargs
|
compat_kwargs = _IDENTITY
|
||||||
|
|
||||||
|
|
||||||
# compat_numeric_types
|
# compat_numeric_types
|
||||||
@ -3132,6 +3192,8 @@ except NameError: # Python 3
|
|||||||
# compat_int
|
# compat_int
|
||||||
compat_int = compat_integer_types[-1]
|
compat_int = compat_integer_types[-1]
|
||||||
|
|
||||||
|
|
||||||
|
# compat_socket_create_connection
|
||||||
if sys.version_info < (2, 7):
|
if sys.version_info < (2, 7):
|
||||||
def compat_socket_create_connection(address, timeout, source_address=None):
|
def compat_socket_create_connection(address, timeout, source_address=None):
|
||||||
host, port = address
|
host, port = address
|
||||||
@ -3158,6 +3220,7 @@ else:
|
|||||||
compat_socket_create_connection = socket.create_connection
|
compat_socket_create_connection = socket.create_connection
|
||||||
|
|
||||||
|
|
||||||
|
# compat_contextlib_suppress
|
||||||
try:
|
try:
|
||||||
from contextlib import suppress as compat_contextlib_suppress
|
from contextlib import suppress as compat_contextlib_suppress
|
||||||
except ImportError:
|
except ImportError:
|
||||||
@ -3205,7 +3268,7 @@ except AttributeError:
|
|||||||
|
|
||||||
# Fix https://github.com/ytdl-org/youtube-dl/issues/4223
|
# Fix https://github.com/ytdl-org/youtube-dl/issues/4223
|
||||||
# See http://bugs.python.org/issue9161 for what is broken
|
# See http://bugs.python.org/issue9161 for what is broken
|
||||||
def workaround_optparse_bug9161():
|
def _workaround_optparse_bug9161():
|
||||||
op = optparse.OptionParser()
|
op = optparse.OptionParser()
|
||||||
og = optparse.OptionGroup(op, 'foo')
|
og = optparse.OptionGroup(op, 'foo')
|
||||||
try:
|
try:
|
||||||
@ -3224,9 +3287,10 @@ def workaround_optparse_bug9161():
|
|||||||
optparse.OptionGroup.add_option = _compat_add_option
|
optparse.OptionGroup.add_option = _compat_add_option
|
||||||
|
|
||||||
|
|
||||||
if hasattr(shutil, 'get_terminal_size'): # Python >= 3.3
|
# compat_shutil_get_terminal_size
|
||||||
compat_get_terminal_size = shutil.get_terminal_size
|
try:
|
||||||
else:
|
from shutil import get_terminal_size as compat_get_terminal_size # Python >= 3.3
|
||||||
|
except ImportError:
|
||||||
_terminal_size = collections.namedtuple('terminal_size', ['columns', 'lines'])
|
_terminal_size = collections.namedtuple('terminal_size', ['columns', 'lines'])
|
||||||
|
|
||||||
def compat_get_terminal_size(fallback=(80, 24)):
|
def compat_get_terminal_size(fallback=(80, 24)):
|
||||||
@ -3256,27 +3320,33 @@ else:
|
|||||||
columns = _columns
|
columns = _columns
|
||||||
if lines is None or lines <= 0:
|
if lines is None or lines <= 0:
|
||||||
lines = _lines
|
lines = _lines
|
||||||
|
|
||||||
return _terminal_size(columns, lines)
|
return _terminal_size(columns, lines)
|
||||||
|
|
||||||
|
compat_shutil_get_terminal_size = compat_get_terminal_size
|
||||||
|
|
||||||
|
|
||||||
|
# compat_itertools_count
|
||||||
try:
|
try:
|
||||||
itertools.count(start=0, step=1)
|
type(itertools.count(start=0, step=1))
|
||||||
compat_itertools_count = itertools.count
|
compat_itertools_count = itertools.count
|
||||||
except TypeError: # Python 2.6
|
except TypeError: # Python 2.6 lacks step
|
||||||
def compat_itertools_count(start=0, step=1):
|
def compat_itertools_count(start=0, step=1):
|
||||||
while True:
|
while True:
|
||||||
yield start
|
yield start
|
||||||
start += step
|
start += step
|
||||||
|
|
||||||
|
|
||||||
|
# compat_tokenize_tokenize
|
||||||
if sys.version_info >= (3, 0):
|
if sys.version_info >= (3, 0):
|
||||||
from tokenize import tokenize as compat_tokenize_tokenize
|
from tokenize import tokenize as compat_tokenize_tokenize
|
||||||
else:
|
else:
|
||||||
from tokenize import generate_tokens as compat_tokenize_tokenize
|
from tokenize import generate_tokens as compat_tokenize_tokenize
|
||||||
|
|
||||||
|
|
||||||
|
# compat_struct_pack, compat_struct_unpack, compat_Struct
|
||||||
try:
|
try:
|
||||||
struct.pack('!I', 0)
|
type(struct.pack('!I', 0))
|
||||||
except TypeError:
|
except TypeError:
|
||||||
# In Python 2.6 and 2.7.x < 2.7.7, struct requires a bytes argument
|
# In Python 2.6 and 2.7.x < 2.7.7, struct requires a bytes argument
|
||||||
# See https://bugs.python.org/issue19099
|
# See https://bugs.python.org/issue19099
|
||||||
@ -3308,8 +3378,10 @@ else:
|
|||||||
compat_Struct = struct.Struct
|
compat_Struct = struct.Struct
|
||||||
|
|
||||||
|
|
||||||
# compat_map/filter() returning an iterator, supposedly the
|
# builtins returning an iterator
|
||||||
# same versioning as for zip below
|
|
||||||
|
# compat_map, compat_filter
|
||||||
|
# supposedly the same versioning as for zip below
|
||||||
try:
|
try:
|
||||||
from future_builtins import map as compat_map
|
from future_builtins import map as compat_map
|
||||||
except ImportError:
|
except ImportError:
|
||||||
@ -3326,6 +3398,7 @@ except ImportError:
|
|||||||
except ImportError:
|
except ImportError:
|
||||||
compat_filter = filter
|
compat_filter = filter
|
||||||
|
|
||||||
|
# compat_zip
|
||||||
try:
|
try:
|
||||||
from future_builtins import zip as compat_zip
|
from future_builtins import zip as compat_zip
|
||||||
except ImportError: # not 2.6+ or is 3.x
|
except ImportError: # not 2.6+ or is 3.x
|
||||||
@ -3335,6 +3408,7 @@ except ImportError: # not 2.6+ or is 3.x
|
|||||||
compat_zip = zip
|
compat_zip = zip
|
||||||
|
|
||||||
|
|
||||||
|
# compat_itertools_zip_longest
|
||||||
# method renamed between Py2/3
|
# method renamed between Py2/3
|
||||||
try:
|
try:
|
||||||
from itertools import zip_longest as compat_itertools_zip_longest
|
from itertools import zip_longest as compat_itertools_zip_longest
|
||||||
@ -3342,7 +3416,8 @@ except ImportError:
|
|||||||
from itertools import izip_longest as compat_itertools_zip_longest
|
from itertools import izip_longest as compat_itertools_zip_longest
|
||||||
|
|
||||||
|
|
||||||
# new class in collections
|
# compat_collections_chain_map
|
||||||
|
# collections.ChainMap: new class
|
||||||
try:
|
try:
|
||||||
from collections import ChainMap as compat_collections_chain_map
|
from collections import ChainMap as compat_collections_chain_map
|
||||||
# Py3.3's ChainMap is deficient
|
# Py3.3's ChainMap is deficient
|
||||||
@ -3398,19 +3473,22 @@ except ImportError:
|
|||||||
def new_child(self, m=None, **kwargs):
|
def new_child(self, m=None, **kwargs):
|
||||||
m = m or {}
|
m = m or {}
|
||||||
m.update(kwargs)
|
m.update(kwargs)
|
||||||
return compat_collections_chain_map(m, *self.maps)
|
# support inheritance !
|
||||||
|
return type(self)(m, *self.maps)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def parents(self):
|
def parents(self):
|
||||||
return compat_collections_chain_map(*(self.maps[1:]))
|
return type(self)(*(self.maps[1:]))
|
||||||
|
|
||||||
|
|
||||||
|
# compat_re_Pattern, compat_re_Match
|
||||||
# Pythons disagree on the type of a pattern (RegexObject, _sre.SRE_Pattern, Pattern, ...?)
|
# Pythons disagree on the type of a pattern (RegexObject, _sre.SRE_Pattern, Pattern, ...?)
|
||||||
compat_re_Pattern = type(re.compile(''))
|
compat_re_Pattern = type(re.compile(''))
|
||||||
# and on the type of a match
|
# and on the type of a match
|
||||||
compat_re_Match = type(re.match('a', 'a'))
|
compat_re_Match = type(re.match('a', 'a'))
|
||||||
|
|
||||||
|
|
||||||
|
# compat_base64_b64decode
|
||||||
if sys.version_info < (3, 3):
|
if sys.version_info < (3, 3):
|
||||||
def compat_b64decode(s, *args, **kwargs):
|
def compat_b64decode(s, *args, **kwargs):
|
||||||
if isinstance(s, compat_str):
|
if isinstance(s, compat_str):
|
||||||
@ -3422,6 +3500,7 @@ else:
|
|||||||
compat_base64_b64decode = compat_b64decode
|
compat_base64_b64decode = compat_b64decode
|
||||||
|
|
||||||
|
|
||||||
|
# compat_ctypes_WINFUNCTYPE
|
||||||
if platform.python_implementation() == 'PyPy' and sys.pypy_version_info < (5, 4, 0):
|
if platform.python_implementation() == 'PyPy' and sys.pypy_version_info < (5, 4, 0):
|
||||||
# PyPy2 prior to version 5.4.0 expects byte strings as Windows function
|
# PyPy2 prior to version 5.4.0 expects byte strings as Windows function
|
||||||
# names, see the original PyPy issue [1] and the youtube-dl one [2].
|
# names, see the original PyPy issue [1] and the youtube-dl one [2].
|
||||||
@ -3440,6 +3519,7 @@ else:
|
|||||||
return ctypes.WINFUNCTYPE(*args, **kwargs)
|
return ctypes.WINFUNCTYPE(*args, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
|
# compat_open
|
||||||
if sys.version_info < (3, 0):
|
if sys.version_info < (3, 0):
|
||||||
# open(file, mode='r', buffering=- 1, encoding=None, errors=None, newline=None, closefd=True) not: opener=None
|
# open(file, mode='r', buffering=- 1, encoding=None, errors=None, newline=None, closefd=True) not: opener=None
|
||||||
def compat_open(file_, *args, **kwargs):
|
def compat_open(file_, *args, **kwargs):
|
||||||
@ -3467,18 +3547,28 @@ except AttributeError:
|
|||||||
def compat_datetime_timedelta_total_seconds(td):
|
def compat_datetime_timedelta_total_seconds(td):
|
||||||
return (td.microseconds + (td.seconds + td.days * 24 * 3600) * 10**6) / 10**6
|
return (td.microseconds + (td.seconds + td.days * 24 * 3600) * 10**6) / 10**6
|
||||||
|
|
||||||
|
|
||||||
# optional decompression packages
|
# optional decompression packages
|
||||||
|
# compat_brotli
|
||||||
# PyPi brotli package implements 'br' Content-Encoding
|
# PyPi brotli package implements 'br' Content-Encoding
|
||||||
try:
|
try:
|
||||||
import brotli as compat_brotli
|
import brotli as compat_brotli
|
||||||
except ImportError:
|
except ImportError:
|
||||||
compat_brotli = None
|
compat_brotli = None
|
||||||
|
# compat_ncompress
|
||||||
# PyPi ncompress package implements 'compress' Content-Encoding
|
# PyPi ncompress package implements 'compress' Content-Encoding
|
||||||
try:
|
try:
|
||||||
import ncompress as compat_ncompress
|
import ncompress as compat_ncompress
|
||||||
except ImportError:
|
except ImportError:
|
||||||
compat_ncompress = None
|
compat_ncompress = None
|
||||||
|
|
||||||
|
# compat_zstandard
|
||||||
|
# PyPi zstandard package implements 'zstd' Content-Encoding (RFC 8878 7.2)
|
||||||
|
try:
|
||||||
|
import zstandard as compat_zstandard
|
||||||
|
except ImportError:
|
||||||
|
compat_zstandard = None
|
||||||
|
|
||||||
|
|
||||||
legacy = [
|
legacy = [
|
||||||
'compat_HTMLParseError',
|
'compat_HTMLParseError',
|
||||||
@ -3495,6 +3585,7 @@ legacy = [
|
|||||||
'compat_getpass',
|
'compat_getpass',
|
||||||
'compat_parse_qs',
|
'compat_parse_qs',
|
||||||
'compat_realpath',
|
'compat_realpath',
|
||||||
|
'compat_shlex_split',
|
||||||
'compat_urllib_parse_parse_qs',
|
'compat_urllib_parse_parse_qs',
|
||||||
'compat_urllib_parse_unquote',
|
'compat_urllib_parse_unquote',
|
||||||
'compat_urllib_parse_unquote_plus',
|
'compat_urllib_parse_unquote_plus',
|
||||||
@ -3508,8 +3599,6 @@ legacy = [
|
|||||||
|
|
||||||
|
|
||||||
__all__ = [
|
__all__ = [
|
||||||
'compat_html_parser_HTMLParseError',
|
|
||||||
'compat_html_parser_HTMLParser',
|
|
||||||
'compat_Struct',
|
'compat_Struct',
|
||||||
'compat_base64_b64decode',
|
'compat_base64_b64decode',
|
||||||
'compat_basestring',
|
'compat_basestring',
|
||||||
@ -3518,13 +3607,9 @@ __all__ = [
|
|||||||
'compat_chr',
|
'compat_chr',
|
||||||
'compat_collections_abc',
|
'compat_collections_abc',
|
||||||
'compat_collections_chain_map',
|
'compat_collections_chain_map',
|
||||||
'compat_datetime_timedelta_total_seconds',
|
|
||||||
'compat_http_cookiejar',
|
|
||||||
'compat_http_cookiejar_Cookie',
|
|
||||||
'compat_http_cookies',
|
|
||||||
'compat_http_cookies_SimpleCookie',
|
|
||||||
'compat_contextlib_suppress',
|
'compat_contextlib_suppress',
|
||||||
'compat_ctypes_WINFUNCTYPE',
|
'compat_ctypes_WINFUNCTYPE',
|
||||||
|
'compat_datetime_timedelta_total_seconds',
|
||||||
'compat_etree_fromstring',
|
'compat_etree_fromstring',
|
||||||
'compat_etree_iterfind',
|
'compat_etree_iterfind',
|
||||||
'compat_filter',
|
'compat_filter',
|
||||||
@ -3533,6 +3618,12 @@ __all__ = [
|
|||||||
'compat_getpass_getpass',
|
'compat_getpass_getpass',
|
||||||
'compat_html_entities',
|
'compat_html_entities',
|
||||||
'compat_html_entities_html5',
|
'compat_html_entities_html5',
|
||||||
|
'compat_html_parser_HTMLParseError',
|
||||||
|
'compat_html_parser_HTMLParser',
|
||||||
|
'compat_http_cookiejar',
|
||||||
|
'compat_http_cookiejar_Cookie',
|
||||||
|
'compat_http_cookies',
|
||||||
|
'compat_http_cookies_SimpleCookie',
|
||||||
'compat_http_client',
|
'compat_http_client',
|
||||||
'compat_http_server',
|
'compat_http_server',
|
||||||
'compat_input',
|
'compat_input',
|
||||||
@ -3555,7 +3646,7 @@ __all__ = [
|
|||||||
'compat_register_utf8',
|
'compat_register_utf8',
|
||||||
'compat_setenv',
|
'compat_setenv',
|
||||||
'compat_shlex_quote',
|
'compat_shlex_quote',
|
||||||
'compat_shlex_split',
|
'compat_shutil_get_terminal_size',
|
||||||
'compat_socket_create_connection',
|
'compat_socket_create_connection',
|
||||||
'compat_str',
|
'compat_str',
|
||||||
'compat_struct_pack',
|
'compat_struct_pack',
|
||||||
@ -3575,5 +3666,5 @@ __all__ = [
|
|||||||
'compat_xml_etree_register_namespace',
|
'compat_xml_etree_register_namespace',
|
||||||
'compat_xpath',
|
'compat_xpath',
|
||||||
'compat_zip',
|
'compat_zip',
|
||||||
'workaround_optparse_bug9161',
|
'compat_zstandard',
|
||||||
]
|
]
|
||||||
|
@ -413,8 +413,6 @@ from .foxnews import (
|
|||||||
FoxNewsArticleIE,
|
FoxNewsArticleIE,
|
||||||
)
|
)
|
||||||
from .foxsports import FoxSportsIE
|
from .foxsports import FoxSportsIE
|
||||||
from .franceculture import FranceCultureIE
|
|
||||||
from .franceinter import FranceInterIE
|
|
||||||
from .francetv import (
|
from .francetv import (
|
||||||
FranceTVIE,
|
FranceTVIE,
|
||||||
FranceTVSiteIE,
|
FranceTVSiteIE,
|
||||||
@ -1011,7 +1009,11 @@ from .radiocanada import (
|
|||||||
from .radiode import RadioDeIE
|
from .radiode import RadioDeIE
|
||||||
from .radiojavan import RadioJavanIE
|
from .radiojavan import RadioJavanIE
|
||||||
from .radiobremen import RadioBremenIE
|
from .radiobremen import RadioBremenIE
|
||||||
from .radiofrance import RadioFranceIE
|
from .radiofrance import (
|
||||||
|
RadioFrancePodcastEpisodeIE,
|
||||||
|
RadioFrancePodcastPlaylistIE,
|
||||||
|
RadioFranceWebradioIE,
|
||||||
|
)
|
||||||
from .rai import (
|
from .rai import (
|
||||||
RaiPlayIE,
|
RaiPlayIE,
|
||||||
RaiPlayLiveIE,
|
RaiPlayLiveIE,
|
||||||
|
@ -1,73 +0,0 @@
|
|||||||
# coding: utf-8
|
|
||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
from .common import InfoExtractor
|
|
||||||
from ..utils import (
|
|
||||||
determine_ext,
|
|
||||||
extract_attributes,
|
|
||||||
int_or_none,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class FranceCultureIE(InfoExtractor):
|
|
||||||
_VALID_URL = r'https?://(?:www\.)?franceculture\.fr/emissions/(?:[^/]+/)*(?P<id>[^/?#&]+)'
|
|
||||||
_TESTS = [{
|
|
||||||
'url': 'http://www.franceculture.fr/emissions/carnet-nomade/rendez-vous-au-pays-des-geeks',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'rendez-vous-au-pays-des-geeks',
|
|
||||||
'display_id': 'rendez-vous-au-pays-des-geeks',
|
|
||||||
'ext': 'mp3',
|
|
||||||
'title': 'Rendez-vous au pays des geeks',
|
|
||||||
'thumbnail': r're:^https?://.*\.jpg$',
|
|
||||||
'upload_date': '20140301',
|
|
||||||
'timestamp': 1393700400,
|
|
||||||
'vcodec': 'none',
|
|
||||||
}
|
|
||||||
}, {
|
|
||||||
# no thumbnail
|
|
||||||
'url': 'https://www.franceculture.fr/emissions/la-recherche-montre-en-main/la-recherche-montre-en-main-du-mercredi-10-octobre-2018',
|
|
||||||
'only_matching': True,
|
|
||||||
}]
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
display_id = self._match_id(url)
|
|
||||||
|
|
||||||
webpage = self._download_webpage(url, display_id)
|
|
||||||
|
|
||||||
video_data = extract_attributes(self._search_regex(
|
|
||||||
r'''(?sx)
|
|
||||||
(?:
|
|
||||||
</h1>|
|
|
||||||
<div[^>]+class="[^"]*?(?:title-zone-diffusion|heading-zone-(?:wrapper|player-button))[^"]*?"[^>]*>
|
|
||||||
).*?
|
|
||||||
(<button[^>]+data-(?:url|asset-source)="[^"]+"[^>]+>)
|
|
||||||
''',
|
|
||||||
webpage, 'video data'))
|
|
||||||
|
|
||||||
video_url = video_data.get('data-url') or video_data['data-asset-source']
|
|
||||||
title = video_data.get('data-asset-title') or video_data.get('data-diffusion-title') or self._og_search_title(webpage)
|
|
||||||
|
|
||||||
description = self._html_search_regex(
|
|
||||||
r'(?s)<div[^>]+class="intro"[^>]*>.*?<h2>(.+?)</h2>',
|
|
||||||
webpage, 'description', default=None)
|
|
||||||
thumbnail = self._search_regex(
|
|
||||||
r'(?s)<figure[^>]+itemtype="https://schema.org/ImageObject"[^>]*>.*?<img[^>]+(?:data-dejavu-)?src="([^"]+)"',
|
|
||||||
webpage, 'thumbnail', default=None)
|
|
||||||
uploader = self._html_search_regex(
|
|
||||||
r'(?s)<span class="author">(.*?)</span>',
|
|
||||||
webpage, 'uploader', default=None)
|
|
||||||
ext = determine_ext(video_url.lower())
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': display_id,
|
|
||||||
'display_id': display_id,
|
|
||||||
'url': video_url,
|
|
||||||
'title': title,
|
|
||||||
'description': description,
|
|
||||||
'thumbnail': thumbnail,
|
|
||||||
'ext': ext,
|
|
||||||
'vcodec': 'none' if ext == 'mp3' else None,
|
|
||||||
'uploader': uploader,
|
|
||||||
'timestamp': int_or_none(video_data.get('data-start-time')) or int_or_none(video_data.get('data-asset-created-date')),
|
|
||||||
'duration': int_or_none(video_data.get('data-duration')),
|
|
||||||
}
|
|
@ -1,59 +0,0 @@
|
|||||||
# coding: utf-8
|
|
||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
from .common import InfoExtractor
|
|
||||||
from ..utils import month_by_name
|
|
||||||
|
|
||||||
|
|
||||||
class FranceInterIE(InfoExtractor):
|
|
||||||
_VALID_URL = r'https?://(?:www\.)?franceinter\.fr/emissions/(?P<id>[^?#]+)'
|
|
||||||
|
|
||||||
_TEST = {
|
|
||||||
'url': 'https://www.franceinter.fr/emissions/affaires-sensibles/affaires-sensibles-07-septembre-2016',
|
|
||||||
'md5': '9e54d7bdb6fdc02a841007f8a975c094',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'affaires-sensibles/affaires-sensibles-07-septembre-2016',
|
|
||||||
'ext': 'mp3',
|
|
||||||
'title': 'Affaire Cahuzac : le contentieux du compte en Suisse',
|
|
||||||
'description': 'md5:401969c5d318c061f86bda1fa359292b',
|
|
||||||
'thumbnail': r're:^https?://.*\.jpg',
|
|
||||||
'upload_date': '20160907',
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
video_id = self._match_id(url)
|
|
||||||
|
|
||||||
webpage = self._download_webpage(url, video_id)
|
|
||||||
|
|
||||||
video_url = self._search_regex(
|
|
||||||
r'(?s)<div[^>]+class=["\']page-diffusion["\'][^>]*>.*?<button[^>]+data-url=(["\'])(?P<url>(?:(?!\1).)+)\1',
|
|
||||||
webpage, 'video url', group='url')
|
|
||||||
|
|
||||||
title = self._og_search_title(webpage)
|
|
||||||
description = self._og_search_description(webpage)
|
|
||||||
thumbnail = self._html_search_meta(['og:image', 'twitter:image'], webpage)
|
|
||||||
|
|
||||||
upload_date_str = self._search_regex(
|
|
||||||
r'class=["\']\s*cover-emission-period\s*["\'][^>]*>[^<]+\s+(\d{1,2}\s+[^\s]+\s+\d{4})<',
|
|
||||||
webpage, 'upload date', fatal=False)
|
|
||||||
if upload_date_str:
|
|
||||||
upload_date_list = upload_date_str.split()
|
|
||||||
upload_date_list.reverse()
|
|
||||||
upload_date_list[1] = '%02d' % (month_by_name(upload_date_list[1], lang='fr') or 0)
|
|
||||||
upload_date_list[2] = '%02d' % int(upload_date_list[2])
|
|
||||||
upload_date = ''.join(upload_date_list)
|
|
||||||
else:
|
|
||||||
upload_date = None
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': video_id,
|
|
||||||
'title': title,
|
|
||||||
'description': description,
|
|
||||||
'thumbnail': thumbnail,
|
|
||||||
'upload_date': upload_date,
|
|
||||||
'formats': [{
|
|
||||||
'url': video_url,
|
|
||||||
'vcodec': 'none',
|
|
||||||
}],
|
|
||||||
}
|
|
@ -4,56 +4,284 @@ from __future__ import unicode_literals
|
|||||||
import re
|
import re
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
|
from ..utils import (
|
||||||
|
ExtractorError,
|
||||||
|
get_element_by_attribute,
|
||||||
|
int_or_none,
|
||||||
|
parse_iso8601,
|
||||||
|
strip_or_none,
|
||||||
|
url_or_none
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class RadioFranceIE(InfoExtractor):
|
class RadioFranceBaseIE(InfoExtractor):
|
||||||
_VALID_URL = r'^https?://maison\.radiofrance\.fr/radiovisions/(?P<id>[^?#]+)'
|
_BASE_URL = r'https://www.radiofrance.fr/'
|
||||||
IE_NAME = 'radiofrance'
|
|
||||||
|
|
||||||
_TEST = {
|
def extract_api_data(self, api_path, id, html):
|
||||||
'url': 'http://maison.radiofrance.fr/radiovisions/one-one',
|
pattern = r'<script [^>]*sveltekit:data-url="https://www\.radiofrance\.fr/api/v[\d.]+/%s[^>]*>(?P<json>.*)</script>' % api_path
|
||||||
'md5': 'bdbb28ace95ed0e04faab32ba3160daf',
|
json = self._search_regex(pattern, html, 'API data', flags=re.DOTALL, group='json')
|
||||||
'info_dict': {
|
|
||||||
'id': 'one-one',
|
if not json:
|
||||||
'ext': 'ogg',
|
raise ExtractorError('%s: JSON data not found' % id)
|
||||||
'title': 'One to one',
|
|
||||||
'description': "Plutôt que d'imaginer la radio de demain comme technologie ou comme création de contenu, je veux montrer que quelles que soient ses évolutions, j'ai l'intime conviction que la radio continuera d'être un grand média de proximité pour les auditeurs.",
|
try:
|
||||||
'uploader': 'Thomas Hercouët',
|
json = self._parse_json(json, id)
|
||||||
},
|
json = self._parse_json(json['body'], id)
|
||||||
|
|
||||||
|
if api_path == 'path':
|
||||||
|
return json['content']
|
||||||
|
elif api_path == 'stations':
|
||||||
|
return json
|
||||||
|
else:
|
||||||
|
raise ExtractorError('Coding error')
|
||||||
|
except KeyError:
|
||||||
|
raise ExtractorError('%s: Invalid JSON' % id)
|
||||||
|
|
||||||
|
def get_title(self, api_data, webpage=None):
|
||||||
|
title = strip_or_none(api_data.get('title'))
|
||||||
|
if not title and webpage:
|
||||||
|
title = strip_or_none(get_element_by_attribute('h1', None, webpage, False)) or strip_or_none(self._og_search_title(webpage))
|
||||||
|
return title
|
||||||
|
|
||||||
|
def get_description(self, api_data, webpage=None):
|
||||||
|
description = strip_or_none(api_data.get('standFirst'))
|
||||||
|
if not description and webpage:
|
||||||
|
description = strip_or_none(self._og_search_description(webpage))
|
||||||
|
return description
|
||||||
|
|
||||||
|
def get_thumbnail(self, api_data, webpage=None):
|
||||||
|
thumbnail = None
|
||||||
|
visual = api_data.get('visual')
|
||||||
|
if visual:
|
||||||
|
thumbnail = url_or_none(visual.get('src'))
|
||||||
|
if not thumbnail and webpage:
|
||||||
|
thumbnail = self._og_search_thumbnail(webpage)
|
||||||
|
return thumbnail
|
||||||
|
|
||||||
|
def get_timestamp(self, api_data, webpage=None):
|
||||||
|
timestamp = api_data.get('publishedDate')
|
||||||
|
if not timestamp and webpage:
|
||||||
|
timestamp = parse_iso8601(self._html_search_meta('article:published_time', webpage, 'publication time', ))
|
||||||
|
return timestamp
|
||||||
|
|
||||||
|
def get_brand(self, api_data, webpage=None):
|
||||||
|
brand = strip_or_none(api_data.get('brand'))
|
||||||
|
if not brand and webpage:
|
||||||
|
brand = self._og_search_property('site_name', webpage, 'Station name', fatal=False)
|
||||||
|
return brand
|
||||||
|
|
||||||
|
def extract_episode(self, episode_id, api_data):
|
||||||
|
manifestations = api_data.get('manifestations')
|
||||||
|
if manifestations is None or len(manifestations) == 0:
|
||||||
|
return None, None
|
||||||
|
|
||||||
|
url = url_or_none(manifestations[0]['url'])
|
||||||
|
duration = int_or_none(manifestations[0].get('duration'))
|
||||||
|
return url, duration
|
||||||
|
|
||||||
|
def get_playlist_entries(self, playlist_url, playlist_id, api_data, direction):
|
||||||
|
playlist_data = api_data['expressions']
|
||||||
|
|
||||||
|
entries = []
|
||||||
|
items = playlist_data.get('items')
|
||||||
|
for item in items:
|
||||||
|
episode_path = item.get('path')
|
||||||
|
if episode_path is None:
|
||||||
|
self.report_warning('No path found for episode "%s"', item.get('title'))
|
||||||
|
continue
|
||||||
|
episode_id = RadioFrancePodcastEpisodeIE._match_id(self._BASE_URL + episode_path)
|
||||||
|
if episode_id is None:
|
||||||
|
self.report_warning('Could not parse id of episode from path: "%s"' % episode_path)
|
||||||
|
continue
|
||||||
|
episode_url, duration = self.extract_episode(episode_id, item)
|
||||||
|
if episode_url is None:
|
||||||
|
self.to_screen('Episode "%s" is not available' % episode_path)
|
||||||
|
continue
|
||||||
|
entry = {
|
||||||
|
'id': episode_id,
|
||||||
|
'url': episode_url,
|
||||||
|
'title': self.get_title(item),
|
||||||
|
'description': self.get_description(item),
|
||||||
|
'timestamp': self.get_timestamp(item),
|
||||||
|
'thumbnail': self.get_thumbnail(item),
|
||||||
|
'duration': duration,
|
||||||
}
|
}
|
||||||
|
entries.append(entry)
|
||||||
|
|
||||||
|
page_number = int_or_none(playlist_data.get('pageNumber'))
|
||||||
|
if page_number:
|
||||||
|
if direction in ['both', 'prev'] and playlist_data.get('prev') is not None:
|
||||||
|
webpage, other_api_data = self.get_data(playlist_url, 'path', playlist_id, page=page_number - 1)
|
||||||
|
entries = self.get_playlist_entries(playlist_url, playlist_id, other_api_data, direction='prev') + entries
|
||||||
|
if direction in ['both', 'next'] and playlist_data.get('next') is not None:
|
||||||
|
webpage, other_api_data = self.get_data(playlist_url, 'path', playlist_id, page=page_number + 1)
|
||||||
|
entries = entries + self.get_playlist_entries(playlist_url, playlist_id, other_api_data, direction='next')
|
||||||
|
|
||||||
|
return entries
|
||||||
|
|
||||||
|
def get_data(self, url, api_path, id, page=None):
|
||||||
|
query = {}
|
||||||
|
note = None
|
||||||
|
if page:
|
||||||
|
query['p'] = page
|
||||||
|
note = "Downloading page %i" % page
|
||||||
|
webpage = self._download_webpage(url, id, query=query, note=note)
|
||||||
|
api_data = self.extract_api_data(api_path, id, webpage)
|
||||||
|
return webpage, api_data
|
||||||
|
|
||||||
|
|
||||||
|
class RadioFrancePodcastEpisodeIE(RadioFranceBaseIE):
|
||||||
|
_VALID_URL = r'https?://www\.radiofrance\.fr/(?:francemusique|franceinter|franceculture|franceinfo|mouv|fip)/podcasts/.+/.+-(?P<id>\d+)$'
|
||||||
|
|
||||||
|
_TESTS = [{
|
||||||
|
'note': 'Podcast episode with audio from France Info',
|
||||||
|
'url': 'https://www.radiofrance.fr/franceinfo/podcasts/le-brief-eco/le-brief-eco-du-lundi-05-septembre-2022-8310713',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '8310713',
|
||||||
|
'ext': 'mp3',
|
||||||
|
'url': r're:^https?://.*\.mp3$',
|
||||||
|
'title': 'Pour la première fois en vingt ans, l’euro passe sous les 0,99\u00a0dollar',
|
||||||
|
'description': str,
|
||||||
|
'thumbnail': r're:^https?://.*\.jpg$',
|
||||||
|
'timestamp': int,
|
||||||
|
'duration': int,
|
||||||
|
'upload_date': str
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
'note': 'Podcast episode from France Musique',
|
||||||
|
'url': 'https://www.radiofrance.fr/francemusique/podcasts/allegretto/lever-du-jour-9233228',
|
||||||
|
'only_matching': True
|
||||||
|
}, {
|
||||||
|
'note': 'Podcast episode from FranceInter',
|
||||||
|
'url': 'https://www.radiofrance.fr/franceinter/podcasts/rendez-vous-avec-x/un-mysterieux-echange-digne-de-la-guerre-froide-9343281',
|
||||||
|
'only_matching': True
|
||||||
|
}, {
|
||||||
|
'note': 'Podcast episode from France Culture',
|
||||||
|
'url': 'https://www.radiofrance.fr/franceculture/podcasts/la-science-cqfd/teotihuacan-la-plus-mysterieuse-des-cites-d-or-9224610',
|
||||||
|
'only_matching': True
|
||||||
|
}, {
|
||||||
|
'note': 'Podcast episode from Le Mouv',
|
||||||
|
'url': 'https://www.radiofrance.fr/mouv/podcasts/mouv-dj-la-caution/ncr2a-ne-cherche-rien-d-autre-ailleurs-1197950',
|
||||||
|
'only_matching': True
|
||||||
|
}, {
|
||||||
|
'note': 'Podcast episode from FIP',
|
||||||
|
'url': 'https://www.radiofrance.fr/fip/podcasts/certains-l-aiment-fip/hommage-au-cinema-de-vangelis-4734742',
|
||||||
|
'only_matching': True
|
||||||
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
m = re.match(self._VALID_URL, url)
|
id = self._match_id(url)
|
||||||
video_id = m.group('id')
|
webpage, api_data = self.get_data(url, 'path', id)
|
||||||
|
url, duration = self.extract_episode(id, api_data)
|
||||||
webpage = self._download_webpage(url, video_id)
|
if url is None:
|
||||||
title = self._html_search_regex(r'<h1>(.*?)</h1>', webpage, 'title')
|
msg = 'Podcast file is not available. If the show is too recent, the file may not have been uploaded yet: try again later.'
|
||||||
description = self._html_search_regex(
|
raise ExtractorError(msg, expected=True, video_id=id)
|
||||||
r'<div class="bloc_page_wrapper"><div class="text">(.*?)</div>',
|
|
||||||
webpage, 'description', fatal=False)
|
|
||||||
uploader = self._html_search_regex(
|
|
||||||
r'<div class="credit"> © (.*?)</div>',
|
|
||||||
webpage, 'uploader', fatal=False)
|
|
||||||
|
|
||||||
formats_str = self._html_search_regex(
|
|
||||||
r'class="jp-jplayer[^"]*" data-source="([^"]+)">',
|
|
||||||
webpage, 'audio URLs')
|
|
||||||
formats = [
|
|
||||||
{
|
|
||||||
'format_id': fm[0],
|
|
||||||
'url': fm[1],
|
|
||||||
'vcodec': 'none',
|
|
||||||
'preference': i,
|
|
||||||
}
|
|
||||||
for i, fm in
|
|
||||||
enumerate(re.findall(r"([a-z0-9]+)\s*:\s*'([^']+)'", formats_str))
|
|
||||||
]
|
|
||||||
self._sort_formats(formats)
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': id,
|
||||||
'title': title,
|
'url': url,
|
||||||
'formats': formats,
|
'title': self.get_title(api_data, webpage),
|
||||||
'description': description,
|
'description': self.get_description(api_data, webpage),
|
||||||
'uploader': uploader,
|
'timestamp': self.get_timestamp(api_data, webpage),
|
||||||
|
'thumbnail': self.get_thumbnail(api_data, webpage),
|
||||||
|
'channel_id': self.get_brand(api_data, webpage),
|
||||||
|
'duration': duration,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class RadioFrancePodcastPlaylistIE(RadioFranceBaseIE):
|
||||||
|
_VALID_URL = r'https?://www\.radiofrance\.fr/(?:francemusique|franceinter|franceculture|franceinfo|mouv|fip)/podcasts/(?P<id>[^/]+?)(?:[?#].*)?$'
|
||||||
|
|
||||||
|
_TESTS = [{
|
||||||
|
'note': 'Podcast show with multiple pages of episodes and some of them are missing',
|
||||||
|
'url': 'https://www.radiofrance.fr/franceculture/podcasts/une-semaine-dans-le-monde-10-11?p=2',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'une-semaine-dans-le-monde-10-11',
|
||||||
|
'title': 'Une semaine dans le monde | 10-11',
|
||||||
|
'description': str,
|
||||||
|
'timestamp': int
|
||||||
|
},
|
||||||
|
'playlist_count': 23,
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
id = self._match_id(url)
|
||||||
|
webpage, api_data = self.get_data(url, 'path', id)
|
||||||
|
|
||||||
|
entries = self.get_playlist_entries(url, id, api_data, direction='both')
|
||||||
|
entries.reverse()
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': id,
|
||||||
|
'_type': 'playlist',
|
||||||
|
'entries': entries,
|
||||||
|
'title': self.get_title(api_data, webpage),
|
||||||
|
'description': self.get_description(api_data, webpage),
|
||||||
|
'timestamp': self.get_timestamp(api_data, webpage),
|
||||||
|
'thumbnail': self.get_thumbnail(api_data, webpage),
|
||||||
|
'channel_id': self.get_brand(api_data, webpage),
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class RadioFranceWebradioIE(RadioFranceBaseIE):
|
||||||
|
_VALID_URL = r'https?://www\.radiofrance\.fr/(?:francemusique|franceinter|franceculture|franceinfo|mouv|fip)/(?P<id>radio-[^/]+)$'
|
||||||
|
|
||||||
|
_TESTS = [{
|
||||||
|
'note': 'Full list of webradios available at https://www.radiofrance.fr/ecouter-musique',
|
||||||
|
'url': 'https://www.radiofrance.fr/fip/radio-metal',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'radio-metal',
|
||||||
|
'ext': 'aac',
|
||||||
|
'title': str,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'format': 'aac',
|
||||||
|
'skip_download': True,
|
||||||
|
}
|
||||||
|
}]
|
||||||
|
|
||||||
|
def get_livestream_formats(self, id, api_data):
|
||||||
|
sources = api_data['media']['sources']
|
||||||
|
|
||||||
|
formats = []
|
||||||
|
for source in sources:
|
||||||
|
url = source.get('url')
|
||||||
|
if not url:
|
||||||
|
continue
|
||||||
|
|
||||||
|
format_id = source.get('format')
|
||||||
|
format = {
|
||||||
|
'url': url,
|
||||||
|
'format_id': format_id,
|
||||||
|
'asr': 48000,
|
||||||
|
'vcodec': 'none'
|
||||||
|
}
|
||||||
|
if format_id == 'mp3':
|
||||||
|
format['preference'] = 1
|
||||||
|
format['acodec'] = 'mp3'
|
||||||
|
format['abr'] = source.get('bitrate')
|
||||||
|
elif format_id == 'aac':
|
||||||
|
format['preference'] = 2
|
||||||
|
format['acodec'] = 'aac'
|
||||||
|
format['abr'] = source.get('bitrate')
|
||||||
|
elif format_id == 'hls':
|
||||||
|
format['preference'] = 0
|
||||||
|
format['manifest_url'] = url
|
||||||
|
formats.append(format)
|
||||||
|
|
||||||
|
if len(formats) == 0:
|
||||||
|
raise ExtractorError('No live streaming URL found')
|
||||||
|
return formats
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
id = self._match_id(url)
|
||||||
|
webpage, api_data = self.get_data(url, 'stations', id)
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': id,
|
||||||
|
'title': self.get_title(api_data, webpage),
|
||||||
|
'formats': self.get_livestream_formats(id, api_data),
|
||||||
|
'thumbnail': self.get_thumbnail(api_data, webpage),
|
||||||
|
'channel_id': self.get_brand(api_data, webpage),
|
||||||
|
'is_live': True
|
||||||
}
|
}
|
||||||
|
@ -27,6 +27,7 @@ from ..compat import (
|
|||||||
)
|
)
|
||||||
from ..jsinterp import JSInterpreter
|
from ..jsinterp import JSInterpreter
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
|
bug_reports_message,
|
||||||
clean_html,
|
clean_html,
|
||||||
dict_get,
|
dict_get,
|
||||||
error_to_compat_str,
|
error_to_compat_str,
|
||||||
@ -65,6 +66,7 @@ from ..utils import (
|
|||||||
url_or_none,
|
url_or_none,
|
||||||
urlencode_postdata,
|
urlencode_postdata,
|
||||||
urljoin,
|
urljoin,
|
||||||
|
variadic,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@ -89,12 +91,12 @@ class YoutubeBaseInfoExtractor(InfoExtractor):
|
|||||||
'INNERTUBE_CONTEXT': {
|
'INNERTUBE_CONTEXT': {
|
||||||
'client': {
|
'client': {
|
||||||
'clientName': 'IOS',
|
'clientName': 'IOS',
|
||||||
'clientVersion': '19.45.4',
|
'clientVersion': '20.10.4',
|
||||||
'deviceMake': 'Apple',
|
'deviceMake': 'Apple',
|
||||||
'deviceModel': 'iPhone16,2',
|
'deviceModel': 'iPhone16,2',
|
||||||
'userAgent': 'com.google.ios.youtube/19.45.4 (iPhone16,2; U; CPU iOS 18_1_0 like Mac OS X;)',
|
'userAgent': 'com.google.ios.youtube/20.10.4 (iPhone16,2; U; CPU iOS 18_3_2 like Mac OS X;)',
|
||||||
'osName': 'iPhone',
|
'osName': 'iPhone',
|
||||||
'osVersion': '18.1.0.22B83',
|
'osVersion': '18.3.2.22D82',
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
'INNERTUBE_CONTEXT_CLIENT_NAME': 5,
|
'INNERTUBE_CONTEXT_CLIENT_NAME': 5,
|
||||||
@ -107,7 +109,7 @@ class YoutubeBaseInfoExtractor(InfoExtractor):
|
|||||||
'INNERTUBE_CONTEXT': {
|
'INNERTUBE_CONTEXT': {
|
||||||
'client': {
|
'client': {
|
||||||
'clientName': 'MWEB',
|
'clientName': 'MWEB',
|
||||||
'clientVersion': '2.20241202.07.00',
|
'clientVersion': '2.20250311.03.00',
|
||||||
# mweb previously did not require PO Token with this UA
|
# mweb previously did not require PO Token with this UA
|
||||||
'userAgent': 'Mozilla/5.0 (iPad; CPU OS 16_7_10 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/16.6 Mobile/15E148 Safari/604.1,gzip(gfe)',
|
'userAgent': 'Mozilla/5.0 (iPad; CPU OS 16_7_10 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/16.6 Mobile/15E148 Safari/604.1,gzip(gfe)',
|
||||||
},
|
},
|
||||||
@ -120,7 +122,8 @@ class YoutubeBaseInfoExtractor(InfoExtractor):
|
|||||||
'INNERTUBE_CONTEXT': {
|
'INNERTUBE_CONTEXT': {
|
||||||
'client': {
|
'client': {
|
||||||
'clientName': 'TVHTML5',
|
'clientName': 'TVHTML5',
|
||||||
'clientVersion': '7.20241201.18.00',
|
'clientVersion': '7.20250312.16.00',
|
||||||
|
'userAgent': 'Mozilla/5.0 (ChromiumStylePlatform) Cobalt/Version',
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
'INNERTUBE_CONTEXT_CLIENT_NAME': 7,
|
'INNERTUBE_CONTEXT_CLIENT_NAME': 7,
|
||||||
@ -130,7 +133,7 @@ class YoutubeBaseInfoExtractor(InfoExtractor):
|
|||||||
'INNERTUBE_CONTEXT': {
|
'INNERTUBE_CONTEXT': {
|
||||||
'client': {
|
'client': {
|
||||||
'clientName': 'WEB',
|
'clientName': 'WEB',
|
||||||
'clientVersion': '2.20241126.01.00',
|
'clientVersion': '2.20250312.04.00',
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
'INNERTUBE_CONTEXT_CLIENT_NAME': 1,
|
'INNERTUBE_CONTEXT_CLIENT_NAME': 1,
|
||||||
@ -460,6 +463,26 @@ class YoutubeBaseInfoExtractor(InfoExtractor):
|
|||||||
'uploader': uploader,
|
'uploader': uploader,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _extract_thumbnails(data, *path_list, **kw_final_key):
|
||||||
|
"""
|
||||||
|
Extract thumbnails from thumbnails dict
|
||||||
|
@param path_list: path list to level that contains 'thumbnails' key
|
||||||
|
"""
|
||||||
|
final_key = kw_final_key.get('final_key', 'thumbnails')
|
||||||
|
|
||||||
|
return traverse_obj(data, ((
|
||||||
|
tuple(variadic(path) + (final_key, Ellipsis)
|
||||||
|
for path in path_list or [()])), {
|
||||||
|
'url': ('url', T(url_or_none),
|
||||||
|
# Sometimes youtube gives a wrong thumbnail URL. See:
|
||||||
|
# https://github.com/yt-dlp/yt-dlp/issues/233
|
||||||
|
# https://github.com/ytdl-org/youtube-dl/issues/28023
|
||||||
|
T(lambda u: update_url(u, query=None) if u and 'maxresdefault' in u else u)),
|
||||||
|
'height': ('height', T(int_or_none)),
|
||||||
|
'width': ('width', T(int_or_none)),
|
||||||
|
}, T(lambda t: t if t.get('url') else None)))
|
||||||
|
|
||||||
def _search_results(self, query, params):
|
def _search_results(self, query, params):
|
||||||
data = {
|
data = {
|
||||||
'context': {
|
'context': {
|
||||||
@ -669,7 +692,7 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
|||||||
'invidious': '|'.join(_INVIDIOUS_SITES),
|
'invidious': '|'.join(_INVIDIOUS_SITES),
|
||||||
}
|
}
|
||||||
_PLAYER_INFO_RE = (
|
_PLAYER_INFO_RE = (
|
||||||
r'/s/player/(?P<id>[a-zA-Z0-9_-]{8,})/player',
|
r'/s/player/(?P<id>[a-zA-Z0-9_-]{8,})//(?:tv-)?player',
|
||||||
r'/(?P<id>[a-zA-Z0-9_-]{8,})/player(?:_ias\.vflset(?:/[a-zA-Z]{2,3}_[a-zA-Z]{2,3})?|-plasma-ias-(?:phone|tablet)-[a-z]{2}_[A-Z]{2}\.vflset)/base\.js$',
|
r'/(?P<id>[a-zA-Z0-9_-]{8,})/player(?:_ias\.vflset(?:/[a-zA-Z]{2,3}_[a-zA-Z]{2,3})?|-plasma-ias-(?:phone|tablet)-[a-z]{2}_[A-Z]{2}\.vflset)/base\.js$',
|
||||||
r'\b(?P<id>vfl[a-zA-Z0-9_-]+)\b.*?\.js$',
|
r'\b(?P<id>vfl[a-zA-Z0-9_-]+)\b.*?\.js$',
|
||||||
)
|
)
|
||||||
@ -1829,11 +1852,21 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
|||||||
|
|
||||||
if func_code:
|
if func_code:
|
||||||
return jsi, player_id, func_code
|
return jsi, player_id, func_code
|
||||||
|
return self._extract_n_function_code_jsi(video_id, jsi, player_id)
|
||||||
|
|
||||||
func_name = self._extract_n_function_name(jscode)
|
def _extract_n_function_code_jsi(self, video_id, jsi, player_id=None):
|
||||||
|
|
||||||
|
var_ay = self._search_regex(
|
||||||
|
r'(?:[;\s]|^)\s*(var\s*[\w$]+\s*=\s*"(?:\\"|[^"])+"\s*\.\s*split\("\W+"\))(?=\s*[,;])',
|
||||||
|
jsi.code, 'useful values', default='')
|
||||||
|
|
||||||
|
func_name = self._extract_n_function_name(jsi.code)
|
||||||
|
|
||||||
func_code = jsi.extract_function_code(func_name)
|
func_code = jsi.extract_function_code(func_name)
|
||||||
|
if var_ay:
|
||||||
|
func_code = (func_code[0], ';\n'.join((var_ay, func_code[1])))
|
||||||
|
|
||||||
|
if player_id:
|
||||||
self.cache.store('youtube-nsig', player_id, func_code)
|
self.cache.store('youtube-nsig', player_id, func_code)
|
||||||
return jsi, player_id, func_code
|
return jsi, player_id, func_code
|
||||||
|
|
||||||
@ -3183,8 +3216,12 @@ class YoutubeTabIE(YoutubeBaseInfoExtractor):
|
|||||||
expected_type=txt_or_none)
|
expected_type=txt_or_none)
|
||||||
|
|
||||||
def _grid_entries(self, grid_renderer):
|
def _grid_entries(self, grid_renderer):
|
||||||
for item in grid_renderer['items']:
|
for item in traverse_obj(grid_renderer, ('items', Ellipsis, T(dict))):
|
||||||
if not isinstance(item, dict):
|
lockup_view_model = traverse_obj(item, ('lockupViewModel', T(dict)))
|
||||||
|
if lockup_view_model:
|
||||||
|
entry = self._extract_lockup_view_model(lockup_view_model)
|
||||||
|
if entry:
|
||||||
|
yield entry
|
||||||
continue
|
continue
|
||||||
renderer = self._extract_grid_item_renderer(item)
|
renderer = self._extract_grid_item_renderer(item)
|
||||||
if not isinstance(renderer, dict):
|
if not isinstance(renderer, dict):
|
||||||
@ -3268,6 +3305,25 @@ class YoutubeTabIE(YoutubeBaseInfoExtractor):
|
|||||||
continue
|
continue
|
||||||
yield self._extract_video(renderer)
|
yield self._extract_video(renderer)
|
||||||
|
|
||||||
|
def _extract_lockup_view_model(self, view_model):
|
||||||
|
content_id = view_model.get('contentId')
|
||||||
|
if not content_id:
|
||||||
|
return
|
||||||
|
content_type = view_model.get('contentType')
|
||||||
|
if content_type not in ('LOCKUP_CONTENT_TYPE_PLAYLIST', 'LOCKUP_CONTENT_TYPE_PODCAST'):
|
||||||
|
self.report_warning(
|
||||||
|
'Unsupported lockup view model content type "{0}"{1}'.format(content_type, bug_reports_message()), only_once=True)
|
||||||
|
return
|
||||||
|
return merge_dicts(self.url_result(
|
||||||
|
update_url_query('https://www.youtube.com/playlist', {'list': content_id}),
|
||||||
|
ie=YoutubeTabIE.ie_key(), video_id=content_id), {
|
||||||
|
'title': traverse_obj(view_model, (
|
||||||
|
'metadata', 'lockupMetadataViewModel', 'title', 'content', T(compat_str))),
|
||||||
|
'thumbnails': self._extract_thumbnails(view_model, (
|
||||||
|
'contentImage', 'collectionThumbnailViewModel', 'primaryThumbnail',
|
||||||
|
'thumbnailViewModel', 'image'), final_key='sources'),
|
||||||
|
})
|
||||||
|
|
||||||
def _video_entry(self, video_renderer):
|
def _video_entry(self, video_renderer):
|
||||||
video_id = video_renderer.get('videoId')
|
video_id = video_renderer.get('videoId')
|
||||||
if video_id:
|
if video_id:
|
||||||
|
@ -1,10 +1,12 @@
|
|||||||
# coding: utf-8
|
# coding: utf-8
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import calendar
|
||||||
import itertools
|
import itertools
|
||||||
import json
|
import json
|
||||||
import operator
|
import operator
|
||||||
import re
|
import re
|
||||||
|
import time
|
||||||
|
|
||||||
from functools import update_wrapper, wraps
|
from functools import update_wrapper, wraps
|
||||||
|
|
||||||
@ -12,8 +14,10 @@ from .utils import (
|
|||||||
error_to_compat_str,
|
error_to_compat_str,
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
float_or_none,
|
float_or_none,
|
||||||
|
int_or_none,
|
||||||
js_to_json,
|
js_to_json,
|
||||||
remove_quotes,
|
remove_quotes,
|
||||||
|
str_or_none,
|
||||||
unified_timestamp,
|
unified_timestamp,
|
||||||
variadic,
|
variadic,
|
||||||
write_string,
|
write_string,
|
||||||
@ -150,6 +154,7 @@ def _js_to_primitive(v):
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
# more exact: yt-dlp/yt-dlp#12110
|
||||||
def _js_toString(v):
|
def _js_toString(v):
|
||||||
return (
|
return (
|
||||||
'undefined' if v is JS_Undefined
|
'undefined' if v is JS_Undefined
|
||||||
@ -158,7 +163,7 @@ def _js_toString(v):
|
|||||||
else 'null' if v is None
|
else 'null' if v is None
|
||||||
# bool <= int: do this first
|
# bool <= int: do this first
|
||||||
else ('false', 'true')[v] if isinstance(v, bool)
|
else ('false', 'true')[v] if isinstance(v, bool)
|
||||||
else '{0:.7f}'.format(v).rstrip('.0') if isinstance(v, compat_numeric_types)
|
else re.sub(r'(?<=\d)\.?0*$', '', '{0:.7f}'.format(v)) if isinstance(v, compat_numeric_types)
|
||||||
else _js_to_primitive(v))
|
else _js_to_primitive(v))
|
||||||
|
|
||||||
|
|
||||||
@ -404,6 +409,7 @@ class JSInterpreter(object):
|
|||||||
class Exception(ExtractorError):
|
class Exception(ExtractorError):
|
||||||
def __init__(self, msg, *args, **kwargs):
|
def __init__(self, msg, *args, **kwargs):
|
||||||
expr = kwargs.pop('expr', None)
|
expr = kwargs.pop('expr', None)
|
||||||
|
msg = str_or_none(msg, default='"None"')
|
||||||
if expr is not None:
|
if expr is not None:
|
||||||
msg = '{0} in: {1!r:.100}'.format(msg.rstrip(), expr)
|
msg = '{0} in: {1!r:.100}'.format(msg.rstrip(), expr)
|
||||||
super(JSInterpreter.Exception, self).__init__(msg, *args, **kwargs)
|
super(JSInterpreter.Exception, self).__init__(msg, *args, **kwargs)
|
||||||
@ -431,6 +437,7 @@ class JSInterpreter(object):
|
|||||||
flags, _ = self.regex_flags(flags)
|
flags, _ = self.regex_flags(flags)
|
||||||
# First, avoid https://github.com/python/cpython/issues/74534
|
# First, avoid https://github.com/python/cpython/issues/74534
|
||||||
self.__self = None
|
self.__self = None
|
||||||
|
pattern_txt = str_or_none(pattern_txt) or '(?:)'
|
||||||
self.__pattern_txt = pattern_txt.replace('[[', r'[\[')
|
self.__pattern_txt = pattern_txt.replace('[[', r'[\[')
|
||||||
self.__flags = flags
|
self.__flags = flags
|
||||||
|
|
||||||
@ -475,6 +482,73 @@ class JSInterpreter(object):
|
|||||||
flags |= cls.RE_FLAGS[ch]
|
flags |= cls.RE_FLAGS[ch]
|
||||||
return flags, expr[idx + 1:]
|
return flags, expr[idx + 1:]
|
||||||
|
|
||||||
|
class JS_Date(object):
|
||||||
|
_t = None
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def __ymd_etc(*args, **kw_is_utc):
|
||||||
|
# args: year, monthIndex, day, hours, minutes, seconds, milliseconds
|
||||||
|
is_utc = kw_is_utc.get('is_utc', False)
|
||||||
|
|
||||||
|
args = list(args[:7])
|
||||||
|
args += [0] * (9 - len(args))
|
||||||
|
args[1] += 1 # month 0..11 -> 1..12
|
||||||
|
ms = args[6]
|
||||||
|
for i in range(6, 9):
|
||||||
|
args[i] = -1 # don't know
|
||||||
|
if is_utc:
|
||||||
|
args[-1] = 1
|
||||||
|
# TODO: [MDN] When a segment overflows or underflows its expected
|
||||||
|
# range, it usually "carries over to" or "borrows from" the higher segment.
|
||||||
|
try:
|
||||||
|
mktime = calendar.timegm if is_utc else time.mktime
|
||||||
|
return mktime(time.struct_time(args)) * 1000 + ms
|
||||||
|
except (OverflowError, ValueError):
|
||||||
|
return None
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def UTC(cls, *args):
|
||||||
|
t = cls.__ymd_etc(*args, is_utc=True)
|
||||||
|
return _NaN if t is None else t
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def parse(date_str, **kw_is_raw):
|
||||||
|
is_raw = kw_is_raw.get('is_raw', False)
|
||||||
|
|
||||||
|
t = unified_timestamp(str_or_none(date_str), False)
|
||||||
|
return int(t * 1000) if t is not None else t if is_raw else _NaN
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def now(**kw_is_raw):
|
||||||
|
is_raw = kw_is_raw.get('is_raw', False)
|
||||||
|
|
||||||
|
t = time.time()
|
||||||
|
return int(t * 1000) if t is not None else t if is_raw else _NaN
|
||||||
|
|
||||||
|
def __init__(self, *args):
|
||||||
|
if not args:
|
||||||
|
args = [self.now(is_raw=True)]
|
||||||
|
if len(args) == 1:
|
||||||
|
if isinstance(args[0], JSInterpreter.JS_Date):
|
||||||
|
self._t = int_or_none(args[0].valueOf(), default=None)
|
||||||
|
else:
|
||||||
|
arg_type = _js_typeof(args[0])
|
||||||
|
if arg_type == 'string':
|
||||||
|
self._t = self.parse(args[0], is_raw=True)
|
||||||
|
elif arg_type == 'number':
|
||||||
|
self._t = int(args[0])
|
||||||
|
else:
|
||||||
|
self._t = self.__ymd_etc(*args)
|
||||||
|
|
||||||
|
def toString(self):
|
||||||
|
try:
|
||||||
|
return time.strftime('%a %b %0d %Y %H:%M:%S %Z%z', self._t).rstrip()
|
||||||
|
except TypeError:
|
||||||
|
return "Invalid Date"
|
||||||
|
|
||||||
|
def valueOf(self):
|
||||||
|
return _NaN if self._t is None else self._t
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def __op_chars(cls):
|
def __op_chars(cls):
|
||||||
op_chars = set(';,[')
|
op_chars = set(';,[')
|
||||||
@ -599,14 +673,15 @@ class JSInterpreter(object):
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise self.Exception('Failed to evaluate {left_val!r:.50} {op} {right_val!r:.50}'.format(**locals()), expr, cause=e)
|
raise self.Exception('Failed to evaluate {left_val!r:.50} {op} {right_val!r:.50}'.format(**locals()), expr, cause=e)
|
||||||
|
|
||||||
def _index(self, obj, idx, allow_undefined=True):
|
def _index(self, obj, idx, allow_undefined=None):
|
||||||
if idx == 'length' and isinstance(obj, list):
|
if idx == 'length' and isinstance(obj, list):
|
||||||
return len(obj)
|
return len(obj)
|
||||||
try:
|
try:
|
||||||
return obj[int(idx)] if isinstance(obj, list) else obj[compat_str(idx)]
|
return obj[int(idx)] if isinstance(obj, list) else obj[compat_str(idx)]
|
||||||
except (TypeError, KeyError, IndexError) as e:
|
except (TypeError, KeyError, IndexError) as e:
|
||||||
if allow_undefined:
|
# allow_undefined is None gives correct behaviour
|
||||||
# when is not allowed?
|
if allow_undefined or (
|
||||||
|
allow_undefined is None and not isinstance(e, TypeError)):
|
||||||
return JS_Undefined
|
return JS_Undefined
|
||||||
raise self.Exception('Cannot get index {idx!r:.100}'.format(**locals()), expr=repr(obj), cause=e)
|
raise self.Exception('Cannot get index {idx!r:.100}'.format(**locals()), expr=repr(obj), cause=e)
|
||||||
|
|
||||||
@ -715,7 +790,7 @@ class JSInterpreter(object):
|
|||||||
|
|
||||||
new_kw, _, obj = expr.partition('new ')
|
new_kw, _, obj = expr.partition('new ')
|
||||||
if not new_kw:
|
if not new_kw:
|
||||||
for klass, konstr in (('Date', lambda x: int(unified_timestamp(x, False) * 1000)),
|
for klass, konstr in (('Date', lambda *x: self.JS_Date(*x).valueOf()),
|
||||||
('RegExp', self.JS_RegExp),
|
('RegExp', self.JS_RegExp),
|
||||||
('Error', self.Exception)):
|
('Error', self.Exception)):
|
||||||
if not obj.startswith(klass + '('):
|
if not obj.startswith(klass + '('):
|
||||||
@ -1034,6 +1109,7 @@ class JSInterpreter(object):
|
|||||||
'String': compat_str,
|
'String': compat_str,
|
||||||
'Math': float,
|
'Math': float,
|
||||||
'Array': list,
|
'Array': list,
|
||||||
|
'Date': self.JS_Date,
|
||||||
}
|
}
|
||||||
obj = local_vars.get(variable)
|
obj = local_vars.get(variable)
|
||||||
if obj in (JS_Undefined, None):
|
if obj in (JS_Undefined, None):
|
||||||
@ -1086,6 +1162,8 @@ class JSInterpreter(object):
|
|||||||
assertion(len(argvals) == 2, 'takes two arguments')
|
assertion(len(argvals) == 2, 'takes two arguments')
|
||||||
return argvals[0] ** argvals[1]
|
return argvals[0] ** argvals[1]
|
||||||
raise self.Exception('Unsupported Math method ' + member, expr=expr)
|
raise self.Exception('Unsupported Math method ' + member, expr=expr)
|
||||||
|
elif obj is self.JS_Date:
|
||||||
|
return getattr(obj, member)(*argvals)
|
||||||
|
|
||||||
if member == 'split':
|
if member == 'split':
|
||||||
assertion(len(argvals) <= 2, 'takes at most two arguments')
|
assertion(len(argvals) <= 2, 'takes at most two arguments')
|
||||||
|
Loading…
x
Reference in New Issue
Block a user