Compare commits
62 Commits
2014.04.04
...
2014.04.07
Author | SHA1 | Date | |
---|---|---|---|
![]() |
a5863bdf33 | ||
![]() |
b58ddb32ba | ||
![]() |
b9e12a8140 | ||
![]() |
104aa7388a | ||
![]() |
c3855d28b0 | ||
![]() |
734f90bb41 | ||
![]() |
91a6addeeb | ||
![]() |
9afb76c5ad | ||
![]() |
dfb2cb5cfd | ||
![]() |
650d688d10 | ||
![]() |
0ba77818f3 | ||
![]() |
09baa7da7e | ||
![]() |
85e787f51d | ||
![]() |
2a9e1e453a | ||
![]() |
ee1e199685 | ||
![]() |
17c5a00774 | ||
![]() |
15c0e8e7b2 | ||
![]() |
cca37fba48 | ||
![]() |
9d0993ec4a | ||
![]() |
342f33bf9e | ||
![]() |
7cd3bc5f99 | ||
![]() |
931055e6cb | ||
![]() |
d0e4cf82f1 | ||
![]() |
6f88df2c57 | ||
![]() |
4479bf2762 | ||
![]() |
1ff7c0f7d8 | ||
![]() |
610e47c87e | ||
![]() |
50f566076f | ||
![]() |
92810ff497 | ||
![]() |
60ccc59a1c | ||
![]() |
91745595d3 | ||
![]() |
d6e40507d0 | ||
![]() |
deed48b472 | ||
![]() |
e4d41bfca5 | ||
![]() |
a355b70f27 | ||
![]() |
f8514f6186 | ||
![]() |
e09b8fcd9d | ||
![]() |
7d1b527ff9 | ||
![]() |
f943c7b622 | ||
![]() |
676eb3f2dd | ||
![]() |
98b7cf1ace | ||
![]() |
c465afd736 | ||
![]() |
b84d6e7fc4 | ||
![]() |
2efd5d78c1 | ||
![]() |
c8edf47b3a | ||
![]() |
3b4c26a428 | ||
![]() |
1525148114 | ||
![]() |
9e0c5791c1 | ||
![]() |
29a1ab2afc | ||
![]() |
fa387d2d99 | ||
![]() |
6d0d573eca | ||
![]() |
bb799e811b | ||
![]() |
04ee53eca1 | ||
![]() |
659eb98a53 | ||
![]() |
ca6aada48e | ||
![]() |
43df5a7e71 | ||
![]() |
88f1c6de7b | ||
![]() |
65a40ab82b | ||
![]() |
4b9cced103 | ||
![]() |
5c38625259 | ||
![]() |
6344fa04bb | ||
![]() |
e3ced9ed61 |
@@ -157,5 +157,11 @@ class TestAllURLsMatching(unittest.TestCase):
|
|||||||
'http://thedailyshow.cc.com/guests/michael-lewis/3efna8/exclusive---michael-lewis-extended-interview-pt--3',
|
'http://thedailyshow.cc.com/guests/michael-lewis/3efna8/exclusive---michael-lewis-extended-interview-pt--3',
|
||||||
['ComedyCentralShows'])
|
['ComedyCentralShows'])
|
||||||
|
|
||||||
|
def test_yahoo_https(self):
|
||||||
|
# https://github.com/rg3/youtube-dl/issues/2701
|
||||||
|
self.assertMatch(
|
||||||
|
'https://screen.yahoo.com/smartwatches-latest-wearable-gadgets-163745379-cbs.html',
|
||||||
|
['Yahoo'])
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
@@ -324,7 +324,6 @@ class TestPlaylists(unittest.TestCase):
|
|||||||
self.assertEqual(result['id'], '342759')
|
self.assertEqual(result['id'], '342759')
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
result['title'], 'General Motors Ignition Switch Recall')
|
result['title'], 'General Motors Ignition Switch Recall')
|
||||||
self.assertEqual(len(result['entries']), 9)
|
|
||||||
whole_duration = sum(e['duration'] for e in result['entries'])
|
whole_duration = sum(e['duration'] for e in result['entries'])
|
||||||
self.assertEqual(whole_duration, 14855)
|
self.assertEqual(whole_duration, 14855)
|
||||||
|
|
||||||
|
@@ -38,6 +38,7 @@ from youtube_dl.utils import (
|
|||||||
xpath_with_ns,
|
xpath_with_ns,
|
||||||
parse_iso8601,
|
parse_iso8601,
|
||||||
strip_jsonp,
|
strip_jsonp,
|
||||||
|
uppercase_escape,
|
||||||
)
|
)
|
||||||
|
|
||||||
if sys.version_info < (3, 0):
|
if sys.version_info < (3, 0):
|
||||||
@@ -279,6 +280,9 @@ class TestUtil(unittest.TestCase):
|
|||||||
d = json.loads(stripped)
|
d = json.loads(stripped)
|
||||||
self.assertEqual(d, [{"id": "532cb", "x": 3}])
|
self.assertEqual(d, [{"id": "532cb", "x": 3}])
|
||||||
|
|
||||||
|
def test_uppercase_escpae(self):
|
||||||
|
self.assertEqual(uppercase_escape(u'aä'), u'aä')
|
||||||
|
self.assertEqual(uppercase_escape(u'\\U0001d550'), u'𝕐')
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
@@ -286,6 +286,9 @@ class YoutubeDL(object):
|
|||||||
"""Print message to stdout if not in quiet mode."""
|
"""Print message to stdout if not in quiet mode."""
|
||||||
return self.to_stdout(message, skip_eol, check_quiet=True)
|
return self.to_stdout(message, skip_eol, check_quiet=True)
|
||||||
|
|
||||||
|
def _write_string(self, s, out=None):
|
||||||
|
write_string(s, out=out, encoding=self.params.get('encoding'))
|
||||||
|
|
||||||
def to_stdout(self, message, skip_eol=False, check_quiet=False):
|
def to_stdout(self, message, skip_eol=False, check_quiet=False):
|
||||||
"""Print message to stdout if not in quiet mode."""
|
"""Print message to stdout if not in quiet mode."""
|
||||||
if self.params.get('logger'):
|
if self.params.get('logger'):
|
||||||
@@ -295,7 +298,7 @@ class YoutubeDL(object):
|
|||||||
terminator = ['\n', ''][skip_eol]
|
terminator = ['\n', ''][skip_eol]
|
||||||
output = message + terminator
|
output = message + terminator
|
||||||
|
|
||||||
write_string(output, self._screen_file)
|
self._write_string(output, self._screen_file)
|
||||||
|
|
||||||
def to_stderr(self, message):
|
def to_stderr(self, message):
|
||||||
"""Print message to stderr."""
|
"""Print message to stderr."""
|
||||||
@@ -305,7 +308,7 @@ class YoutubeDL(object):
|
|||||||
else:
|
else:
|
||||||
message = self._bidi_workaround(message)
|
message = self._bidi_workaround(message)
|
||||||
output = message + '\n'
|
output = message + '\n'
|
||||||
write_string(output, self._err_file)
|
self._write_string(output, self._err_file)
|
||||||
|
|
||||||
def to_console_title(self, message):
|
def to_console_title(self, message):
|
||||||
if not self.params.get('consoletitle', False):
|
if not self.params.get('consoletitle', False):
|
||||||
@@ -315,21 +318,21 @@ class YoutubeDL(object):
|
|||||||
# already of type unicode()
|
# already of type unicode()
|
||||||
ctypes.windll.kernel32.SetConsoleTitleW(ctypes.c_wchar_p(message))
|
ctypes.windll.kernel32.SetConsoleTitleW(ctypes.c_wchar_p(message))
|
||||||
elif 'TERM' in os.environ:
|
elif 'TERM' in os.environ:
|
||||||
write_string('\033]0;%s\007' % message, self._screen_file)
|
self._write_string('\033]0;%s\007' % message, self._screen_file)
|
||||||
|
|
||||||
def save_console_title(self):
|
def save_console_title(self):
|
||||||
if not self.params.get('consoletitle', False):
|
if not self.params.get('consoletitle', False):
|
||||||
return
|
return
|
||||||
if 'TERM' in os.environ:
|
if 'TERM' in os.environ:
|
||||||
# Save the title on stack
|
# Save the title on stack
|
||||||
write_string('\033[22;0t', self._screen_file)
|
self._write_string('\033[22;0t', self._screen_file)
|
||||||
|
|
||||||
def restore_console_title(self):
|
def restore_console_title(self):
|
||||||
if not self.params.get('consoletitle', False):
|
if not self.params.get('consoletitle', False):
|
||||||
return
|
return
|
||||||
if 'TERM' in os.environ:
|
if 'TERM' in os.environ:
|
||||||
# Restore the title from stack
|
# Restore the title from stack
|
||||||
write_string('\033[23;0t', self._screen_file)
|
self._write_string('\033[23;0t', self._screen_file)
|
||||||
|
|
||||||
def __enter__(self):
|
def __enter__(self):
|
||||||
self.save_console_title()
|
self.save_console_title()
|
||||||
@@ -1211,9 +1214,16 @@ class YoutubeDL(object):
|
|||||||
if not self.params.get('verbose'):
|
if not self.params.get('verbose'):
|
||||||
return
|
return
|
||||||
|
|
||||||
write_string('[debug] Encodings: locale %s, fs %s, out %s, pref %s\n' %
|
write_string(
|
||||||
(locale.getpreferredencoding(), sys.getfilesystemencoding(), sys.stdout.encoding, self.get_encoding()))
|
'[debug] Encodings: locale %s, fs %s, out %s, pref %s\n' % (
|
||||||
write_string('[debug] youtube-dl version ' + __version__ + '\n')
|
locale.getpreferredencoding(),
|
||||||
|
sys.getfilesystemencoding(),
|
||||||
|
sys.stdout.encoding,
|
||||||
|
self.get_encoding()),
|
||||||
|
encoding=None
|
||||||
|
)
|
||||||
|
|
||||||
|
self._write_string('[debug] youtube-dl version ' + __version__ + '\n')
|
||||||
try:
|
try:
|
||||||
sp = subprocess.Popen(
|
sp = subprocess.Popen(
|
||||||
['git', 'rev-parse', '--short', 'HEAD'],
|
['git', 'rev-parse', '--short', 'HEAD'],
|
||||||
@@ -1222,20 +1232,20 @@ class YoutubeDL(object):
|
|||||||
out, err = sp.communicate()
|
out, err = sp.communicate()
|
||||||
out = out.decode().strip()
|
out = out.decode().strip()
|
||||||
if re.match('[0-9a-f]+', out):
|
if re.match('[0-9a-f]+', out):
|
||||||
write_string('[debug] Git HEAD: ' + out + '\n')
|
self._write_string('[debug] Git HEAD: ' + out + '\n')
|
||||||
except:
|
except:
|
||||||
try:
|
try:
|
||||||
sys.exc_clear()
|
sys.exc_clear()
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
write_string('[debug] Python version %s - %s' %
|
self._write_string('[debug] Python version %s - %s' %
|
||||||
(platform.python_version(), platform_name()) + '\n')
|
(platform.python_version(), platform_name()) + '\n')
|
||||||
|
|
||||||
proxy_map = {}
|
proxy_map = {}
|
||||||
for handler in self._opener.handlers:
|
for handler in self._opener.handlers:
|
||||||
if hasattr(handler, 'proxies'):
|
if hasattr(handler, 'proxies'):
|
||||||
proxy_map.update(handler.proxies)
|
proxy_map.update(handler.proxies)
|
||||||
write_string('[debug] Proxy map: ' + compat_str(proxy_map) + '\n')
|
self._write_string('[debug] Proxy map: ' + compat_str(proxy_map) + '\n')
|
||||||
|
|
||||||
def _setup_opener(self):
|
def _setup_opener(self):
|
||||||
timeout_val = self.params.get('socket_timeout')
|
timeout_val = self.params.get('socket_timeout')
|
||||||
|
@@ -52,6 +52,7 @@ __authors__ = (
|
|||||||
'Juan C. Olivares',
|
'Juan C. Olivares',
|
||||||
'Mattias Harrysson',
|
'Mattias Harrysson',
|
||||||
'phaer',
|
'phaer',
|
||||||
|
'Sainyam Kapoor',
|
||||||
)
|
)
|
||||||
|
|
||||||
__license__ = 'Public Domain'
|
__license__ = 'Public Domain'
|
||||||
@@ -242,7 +243,7 @@ def parseOpts(overrideArguments=None):
|
|||||||
help='Use the specified HTTP/HTTPS proxy. Pass in an empty string (--proxy "") for direct connection')
|
help='Use the specified HTTP/HTTPS proxy. Pass in an empty string (--proxy "") for direct connection')
|
||||||
general.add_option('--no-check-certificate', action='store_true', dest='no_check_certificate', default=False, help='Suppress HTTPS certificate validation.')
|
general.add_option('--no-check-certificate', action='store_true', dest='no_check_certificate', default=False, help='Suppress HTTPS certificate validation.')
|
||||||
general.add_option(
|
general.add_option(
|
||||||
'--prefer-insecure', action='store_true', dest='prefer_insecure',
|
'--prefer-insecure', '--prefer-unsecure', action='store_true', dest='prefer_insecure',
|
||||||
help='Use an unencrypted connection to retrieve information about the video. (Currently supported only for YouTube)')
|
help='Use an unencrypted connection to retrieve information about the video. (Currently supported only for YouTube)')
|
||||||
general.add_option(
|
general.add_option(
|
||||||
'--cache-dir', dest='cachedir', default=get_cachedir(), metavar='DIR',
|
'--cache-dir', dest='cachedir', default=get_cachedir(), metavar='DIR',
|
||||||
|
@@ -4,9 +4,10 @@ import sys
|
|||||||
import time
|
import time
|
||||||
|
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
|
compat_str,
|
||||||
encodeFilename,
|
encodeFilename,
|
||||||
timeconvert,
|
|
||||||
format_bytes,
|
format_bytes,
|
||||||
|
timeconvert,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@@ -173,7 +174,7 @@ class FileDownloader(object):
|
|||||||
return
|
return
|
||||||
os.rename(encodeFilename(old_filename), encodeFilename(new_filename))
|
os.rename(encodeFilename(old_filename), encodeFilename(new_filename))
|
||||||
except (IOError, OSError) as err:
|
except (IOError, OSError) as err:
|
||||||
self.report_error(u'unable to rename file: %s' % str(err))
|
self.report_error(u'unable to rename file: %s' % compat_str(err))
|
||||||
|
|
||||||
def try_utime(self, filename, last_modified_hdr):
|
def try_utime(self, filename, last_modified_hdr):
|
||||||
"""Try to set the last-modified time of the given file."""
|
"""Try to set the last-modified time of the given file."""
|
||||||
|
@@ -32,6 +32,7 @@ from .canal13cl import Canal13clIE
|
|||||||
from .canalplus import CanalplusIE
|
from .canalplus import CanalplusIE
|
||||||
from .canalc2 import Canalc2IE
|
from .canalc2 import Canalc2IE
|
||||||
from .cbs import CBSIE
|
from .cbs import CBSIE
|
||||||
|
from .cbsnews import CBSNewsIE
|
||||||
from .ceskatelevize import CeskaTelevizeIE
|
from .ceskatelevize import CeskaTelevizeIE
|
||||||
from .channel9 import Channel9IE
|
from .channel9 import Channel9IE
|
||||||
from .chilloutzone import ChilloutzoneIE
|
from .chilloutzone import ChilloutzoneIE
|
||||||
@@ -62,6 +63,7 @@ from .dotsub import DotsubIE
|
|||||||
from .dreisat import DreiSatIE
|
from .dreisat import DreiSatIE
|
||||||
from .defense import DefenseGouvFrIE
|
from .defense import DefenseGouvFrIE
|
||||||
from .discovery import DiscoveryIE
|
from .discovery import DiscoveryIE
|
||||||
|
from .divxstage import DivxStageIE
|
||||||
from .dropbox import DropboxIE
|
from .dropbox import DropboxIE
|
||||||
from .ebaumsworld import EbaumsWorldIE
|
from .ebaumsworld import EbaumsWorldIE
|
||||||
from .ehow import EHowIE
|
from .ehow import EHowIE
|
||||||
@@ -156,6 +158,7 @@ from .mofosex import MofosexIE
|
|||||||
from .mooshare import MooshareIE
|
from .mooshare import MooshareIE
|
||||||
from .morningstar import MorningstarIE
|
from .morningstar import MorningstarIE
|
||||||
from .motorsport import MotorsportIE
|
from .motorsport import MotorsportIE
|
||||||
|
from .movshare import MovShareIE
|
||||||
from .mtv import (
|
from .mtv import (
|
||||||
MTVIE,
|
MTVIE,
|
||||||
MTVIggyIE,
|
MTVIggyIE,
|
||||||
@@ -205,6 +208,7 @@ from .rottentomatoes import RottenTomatoesIE
|
|||||||
from .roxwel import RoxwelIE
|
from .roxwel import RoxwelIE
|
||||||
from .rtlnow import RTLnowIE
|
from .rtlnow import RTLnowIE
|
||||||
from .rts import RTSIE
|
from .rts import RTSIE
|
||||||
|
from .rtve import RTVEALaCartaIE
|
||||||
from .rutube import (
|
from .rutube import (
|
||||||
RutubeIE,
|
RutubeIE,
|
||||||
RutubeChannelIE,
|
RutubeChannelIE,
|
||||||
@@ -276,6 +280,7 @@ from .videodetective import VideoDetectiveIE
|
|||||||
from .videolecturesnet import VideoLecturesNetIE
|
from .videolecturesnet import VideoLecturesNetIE
|
||||||
from .videofyme import VideofyMeIE
|
from .videofyme import VideofyMeIE
|
||||||
from .videopremium import VideoPremiumIE
|
from .videopremium import VideoPremiumIE
|
||||||
|
from .videoweed import VideoWeedIE
|
||||||
from .vimeo import (
|
from .vimeo import (
|
||||||
VimeoIE,
|
VimeoIE,
|
||||||
VimeoChannelIE,
|
VimeoChannelIE,
|
||||||
|
@@ -27,9 +27,10 @@ class BreakIE(InfoExtractor):
|
|||||||
webpage, 'info json', flags=re.DOTALL)
|
webpage, 'info json', flags=re.DOTALL)
|
||||||
info = json.loads(info_json)
|
info = json.loads(info_json)
|
||||||
video_url = info['videoUri']
|
video_url = info['videoUri']
|
||||||
m_youtube = re.search(r'(https?://www\.youtube\.com/watch\?v=.*)', video_url)
|
youtube_id = info.get('youtubeId')
|
||||||
if m_youtube is not None:
|
if youtube_id:
|
||||||
return self.url_result(m_youtube.group(1), 'Youtube')
|
return self.url_result(youtube_id, 'Youtube')
|
||||||
|
|
||||||
final_url = video_url + '?' + info['AuthToken']
|
final_url = video_url + '?' + info['AuthToken']
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
|
87
youtube_dl/extractor/cbsnews.py
Normal file
87
youtube_dl/extractor/cbsnews.py
Normal file
@@ -0,0 +1,87 @@
|
|||||||
|
# encoding: utf-8
|
||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import re
|
||||||
|
import json
|
||||||
|
|
||||||
|
from .common import InfoExtractor
|
||||||
|
|
||||||
|
|
||||||
|
class CBSNewsIE(InfoExtractor):
|
||||||
|
IE_DESC = 'CBS News'
|
||||||
|
_VALID_URL = r'http://(?:www\.)?cbsnews\.com/(?:[^/]+/)+(?P<id>[\da-z_-]+)'
|
||||||
|
|
||||||
|
_TESTS = [
|
||||||
|
{
|
||||||
|
'url': 'http://www.cbsnews.com/news/tesla-and-spacex-elon-musks-industrial-empire/',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'tesla-and-spacex-elon-musks-industrial-empire',
|
||||||
|
'ext': 'flv',
|
||||||
|
'title': 'Tesla and SpaceX: Elon Musk\'s industrial empire',
|
||||||
|
'thumbnail': 'http://beta.img.cbsnews.com/i/2014/03/30/60147937-2f53-4565-ad64-1bdd6eb64679/60-0330-pelley-640x360.jpg',
|
||||||
|
'duration': 791,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
# rtmp download
|
||||||
|
'skip_download': True,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
'url': 'http://www.cbsnews.com/videos/fort-hood-shooting-army-downplays-mental-illness-as-cause-of-attack/',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'fort-hood-shooting-army-downplays-mental-illness-as-cause-of-attack',
|
||||||
|
'ext': 'flv',
|
||||||
|
'title': 'Fort Hood shooting: Army downplays mental illness as cause of attack',
|
||||||
|
'thumbnail': 'http://cbsnews2.cbsistatic.com/hub/i/r/2014/04/04/0c9fbc66-576b-41ca-8069-02d122060dd2/thumbnail/140x90/6dad7a502f88875ceac38202984b6d58/en-0404-werner-replace-640x360.jpg',
|
||||||
|
'duration': 205,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
# rtmp download
|
||||||
|
'skip_download': True,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
mobj = re.match(self._VALID_URL, url)
|
||||||
|
video_id = mobj.group('id')
|
||||||
|
|
||||||
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
|
video_info = json.loads(self._html_search_regex(
|
||||||
|
r'(?:<ul class="media-list items" id="media-related-items"><li data-video-info|<div id="cbsNewsVideoPlayer" data-video-player-options)=\'({.+?})\'',
|
||||||
|
webpage, 'video JSON info'))
|
||||||
|
|
||||||
|
item = video_info['item'] if 'item' in video_info else video_info
|
||||||
|
title = item.get('articleTitle') or item.get('hed')
|
||||||
|
duration = item.get('duration')
|
||||||
|
thumbnail = item.get('mediaImage') or item.get('thumbnail')
|
||||||
|
|
||||||
|
formats = []
|
||||||
|
for format_id in ['RtmpMobileLow', 'RtmpMobileHigh', 'Hls', 'RtmpDesktop']:
|
||||||
|
uri = item.get('media' + format_id + 'URI')
|
||||||
|
if not uri:
|
||||||
|
continue
|
||||||
|
fmt = {
|
||||||
|
'url': uri,
|
||||||
|
'format_id': format_id,
|
||||||
|
}
|
||||||
|
if uri.startswith('rtmp'):
|
||||||
|
fmt.update({
|
||||||
|
'app': 'ondemand?auth=cbs',
|
||||||
|
'play_path': 'mp4:' + uri.split('<break>')[-1],
|
||||||
|
'player_url': 'http://www.cbsnews.com/[[IMPORT]]/vidtech.cbsinteractive.com/player/3_3_0/CBSI_PLAYER_HD.swf',
|
||||||
|
'page_url': 'http://www.cbsnews.com',
|
||||||
|
'ext': 'flv',
|
||||||
|
})
|
||||||
|
elif uri.endswith('.m3u8'):
|
||||||
|
fmt['ext'] = 'mp4'
|
||||||
|
formats.append(fmt)
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'title': title,
|
||||||
|
'thumbnail': thumbnail,
|
||||||
|
'duration': duration,
|
||||||
|
'formats': formats,
|
||||||
|
}
|
@@ -8,7 +8,6 @@ from .subtitles import SubtitlesInfoExtractor
|
|||||||
from ..utils import (
|
from ..utils import (
|
||||||
compat_urllib_request,
|
compat_urllib_request,
|
||||||
compat_str,
|
compat_str,
|
||||||
get_element_by_attribute,
|
|
||||||
get_element_by_id,
|
get_element_by_id,
|
||||||
orderedSet,
|
orderedSet,
|
||||||
str_to_int,
|
str_to_int,
|
||||||
|
27
youtube_dl/extractor/divxstage.py
Normal file
27
youtube_dl/extractor/divxstage.py
Normal file
@@ -0,0 +1,27 @@
|
|||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
from .novamov import NovaMovIE
|
||||||
|
|
||||||
|
|
||||||
|
class DivxStageIE(NovaMovIE):
|
||||||
|
IE_NAME = 'divxstage'
|
||||||
|
IE_DESC = 'DivxStage'
|
||||||
|
|
||||||
|
_VALID_URL = NovaMovIE._VALID_URL_TEMPLATE % {'host': 'divxstage\.(?:eu|net|ch|co|at|ag)'}
|
||||||
|
|
||||||
|
_HOST = 'www.divxstage.eu'
|
||||||
|
|
||||||
|
_FILE_DELETED_REGEX = r'>This file no longer exists on our servers.<'
|
||||||
|
_TITLE_REGEX = r'<div class="video_det">\s*<strong>([^<]+)</strong>'
|
||||||
|
_DESCRIPTION_REGEX = r'<div class="video_det">\s*<strong>[^<]+</strong>\s*<p>([^<]+)</p>'
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://www.divxstage.eu/video/57f238e2e5e01',
|
||||||
|
'md5': '63969f6eb26533a1968c4d325be63e72',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '57f238e2e5e01',
|
||||||
|
'ext': 'flv',
|
||||||
|
'title': 'youtubedl test video',
|
||||||
|
'description': 'This is a test video for youtubedl.',
|
||||||
|
}
|
||||||
|
}
|
@@ -184,6 +184,17 @@ class GenericIE(InfoExtractor):
|
|||||||
'description': 'md5:ddb2a40ecd6b6a147e400e535874947b',
|
'description': 'md5:ddb2a40ecd6b6a147e400e535874947b',
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
# Embeded Ustream video
|
||||||
|
{
|
||||||
|
'url': 'http://www.american.edu/spa/pti/nsa-privacy-janus-2014.cfm',
|
||||||
|
'md5': '27b99cdb639c9b12a79bca876a073417',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '45734260',
|
||||||
|
'ext': 'flv',
|
||||||
|
'uploader': 'AU SPA: The NSA and Privacy',
|
||||||
|
'title': 'NSA and Privacy Forum Debate featuring General Hayden and Barton Gellman'
|
||||||
|
}
|
||||||
|
},
|
||||||
# nowvideo embed hidden behind percent encoding
|
# nowvideo embed hidden behind percent encoding
|
||||||
{
|
{
|
||||||
'url': 'http://www.waoanime.tv/the-super-dimension-fortress-macross-episode-1/',
|
'url': 'http://www.waoanime.tv/the-super-dimension-fortress-macross-episode-1/',
|
||||||
@@ -500,17 +511,18 @@ class GenericIE(InfoExtractor):
|
|||||||
if mobj is not None:
|
if mobj is not None:
|
||||||
return self.url_result(mobj.group(1), 'Mpora')
|
return self.url_result(mobj.group(1), 'Mpora')
|
||||||
|
|
||||||
# Look for embedded NovaMov player
|
# Look for embedded NovaMov-based player
|
||||||
mobj = re.search(
|
mobj = re.search(
|
||||||
r'<iframe[^>]+?src=(["\'])(?P<url>http://(?:(?:embed|www)\.)?novamov\.com/embed\.php.+?)\1', webpage)
|
r'''(?x)<iframe[^>]+?src=(["\'])
|
||||||
|
(?P<url>http://(?:(?:embed|www)\.)?
|
||||||
|
(?:novamov\.com|
|
||||||
|
nowvideo\.(?:ch|sx|eu|at|ag|co)|
|
||||||
|
videoweed\.(?:es|com)|
|
||||||
|
movshare\.(?:net|sx|ag)|
|
||||||
|
divxstage\.(?:eu|net|ch|co|at|ag))
|
||||||
|
/embed\.php.+?)\1''', webpage)
|
||||||
if mobj is not None:
|
if mobj is not None:
|
||||||
return self.url_result(mobj.group('url'), 'NovaMov')
|
return self.url_result(mobj.group('url'))
|
||||||
|
|
||||||
# Look for embedded NowVideo player
|
|
||||||
mobj = re.search(
|
|
||||||
r'<iframe[^>]+?src=(["\'])(?P<url>http://(?:(?:embed|www)\.)?nowvideo\.(?:ch|sx|eu)/embed\.php.+?)\1', webpage)
|
|
||||||
if mobj is not None:
|
|
||||||
return self.url_result(mobj.group('url'), 'NowVideo')
|
|
||||||
|
|
||||||
# Look for embedded Facebook player
|
# Look for embedded Facebook player
|
||||||
mobj = re.search(
|
mobj = re.search(
|
||||||
@@ -556,6 +568,12 @@ class GenericIE(InfoExtractor):
|
|||||||
if mobj is not None:
|
if mobj is not None:
|
||||||
return self.url_result(mobj.group('url'), 'TED')
|
return self.url_result(mobj.group('url'), 'TED')
|
||||||
|
|
||||||
|
# Look for embedded Ustream videos
|
||||||
|
mobj = re.search(
|
||||||
|
r'<iframe[^>]+?src=(["\'])(?P<url>http://www\.ustream\.tv/embed/.+?)\1', webpage)
|
||||||
|
if mobj is not None:
|
||||||
|
return self.url_result(mobj.group('url'), 'Ustream')
|
||||||
|
|
||||||
# Look for embedded arte.tv player
|
# Look for embedded arte.tv player
|
||||||
mobj = re.search(
|
mobj = re.search(
|
||||||
r'<script [^>]*?src="(?P<url>http://www\.arte\.tv/playerv2/embed[^"]+)"',
|
r'<script [^>]*?src="(?P<url>http://www\.arte\.tv/playerv2/embed[^"]+)"',
|
||||||
|
@@ -1,9 +1,12 @@
|
|||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
|
compat_str,
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
formatSeconds,
|
formatSeconds,
|
||||||
)
|
)
|
||||||
@@ -24,34 +27,31 @@ class JustinTVIE(InfoExtractor):
|
|||||||
/?(?:\#.*)?$
|
/?(?:\#.*)?$
|
||||||
"""
|
"""
|
||||||
_JUSTIN_PAGE_LIMIT = 100
|
_JUSTIN_PAGE_LIMIT = 100
|
||||||
IE_NAME = u'justin.tv'
|
IE_NAME = 'justin.tv'
|
||||||
|
IE_DESC = 'justin.tv and twitch.tv'
|
||||||
_TEST = {
|
_TEST = {
|
||||||
u'url': u'http://www.twitch.tv/thegamedevhub/b/296128360',
|
'url': 'http://www.twitch.tv/thegamedevhub/b/296128360',
|
||||||
u'file': u'296128360.flv',
|
'md5': 'ecaa8a790c22a40770901460af191c9a',
|
||||||
u'md5': u'ecaa8a790c22a40770901460af191c9a',
|
'info_dict': {
|
||||||
u'info_dict': {
|
'id': '296128360',
|
||||||
u"upload_date": u"20110927",
|
'ext': 'flv',
|
||||||
u"uploader_id": 25114803,
|
'upload_date': '20110927',
|
||||||
u"uploader": u"thegamedevhub",
|
'uploader_id': 25114803,
|
||||||
u"title": u"Beginner Series - Scripting With Python Pt.1"
|
'uploader': 'thegamedevhub',
|
||||||
|
'title': 'Beginner Series - Scripting With Python Pt.1'
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
def report_download_page(self, channel, offset):
|
|
||||||
"""Report attempt to download a single page of videos."""
|
|
||||||
self.to_screen(u'%s: Downloading video information from %d to %d' %
|
|
||||||
(channel, offset, offset + self._JUSTIN_PAGE_LIMIT))
|
|
||||||
|
|
||||||
# Return count of items, list of *valid* items
|
# Return count of items, list of *valid* items
|
||||||
def _parse_page(self, url, video_id):
|
def _parse_page(self, url, video_id):
|
||||||
info_json = self._download_webpage(url, video_id,
|
info_json = self._download_webpage(url, video_id,
|
||||||
u'Downloading video info JSON',
|
'Downloading video info JSON',
|
||||||
u'unable to download video info JSON')
|
'unable to download video info JSON')
|
||||||
|
|
||||||
response = json.loads(info_json)
|
response = json.loads(info_json)
|
||||||
if type(response) != list:
|
if type(response) != list:
|
||||||
error_text = response.get('error', 'unknown error')
|
error_text = response.get('error', 'unknown error')
|
||||||
raise ExtractorError(u'Justin.tv API: %s' % error_text)
|
raise ExtractorError('Justin.tv API: %s' % error_text)
|
||||||
info = []
|
info = []
|
||||||
for clip in response:
|
for clip in response:
|
||||||
video_url = clip['video_file_url']
|
video_url = clip['video_file_url']
|
||||||
@@ -62,7 +62,7 @@ class JustinTVIE(InfoExtractor):
|
|||||||
video_id = clip['id']
|
video_id = clip['id']
|
||||||
video_title = clip.get('title', video_id)
|
video_title = clip.get('title', video_id)
|
||||||
info.append({
|
info.append({
|
||||||
'id': video_id,
|
'id': compat_str(video_id),
|
||||||
'url': video_url,
|
'url': video_url,
|
||||||
'title': video_title,
|
'title': video_title,
|
||||||
'uploader': clip.get('channel_name', video_uploader_id),
|
'uploader': clip.get('channel_name', video_uploader_id),
|
||||||
@@ -74,8 +74,6 @@ class JustinTVIE(InfoExtractor):
|
|||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
mobj = re.match(self._VALID_URL, url)
|
mobj = re.match(self._VALID_URL, url)
|
||||||
if mobj is None:
|
|
||||||
raise ExtractorError(u'invalid URL: %s' % url)
|
|
||||||
|
|
||||||
api_base = 'http://api.justin.tv'
|
api_base = 'http://api.justin.tv'
|
||||||
paged = False
|
paged = False
|
||||||
@@ -89,40 +87,41 @@ class JustinTVIE(InfoExtractor):
|
|||||||
webpage = self._download_webpage(url, chapter_id)
|
webpage = self._download_webpage(url, chapter_id)
|
||||||
m = re.search(r'PP\.archive_id = "([0-9]+)";', webpage)
|
m = re.search(r'PP\.archive_id = "([0-9]+)";', webpage)
|
||||||
if not m:
|
if not m:
|
||||||
raise ExtractorError(u'Cannot find archive of a chapter')
|
raise ExtractorError('Cannot find archive of a chapter')
|
||||||
archive_id = m.group(1)
|
archive_id = m.group(1)
|
||||||
|
|
||||||
api = api_base + '/broadcast/by_chapter/%s.xml' % chapter_id
|
api = api_base + '/broadcast/by_chapter/%s.xml' % chapter_id
|
||||||
doc = self._download_xml(api, chapter_id,
|
doc = self._download_xml(
|
||||||
note=u'Downloading chapter information',
|
api, chapter_id,
|
||||||
errnote=u'Chapter information download failed')
|
note='Downloading chapter information',
|
||||||
|
errnote='Chapter information download failed')
|
||||||
for a in doc.findall('.//archive'):
|
for a in doc.findall('.//archive'):
|
||||||
if archive_id == a.find('./id').text:
|
if archive_id == a.find('./id').text:
|
||||||
break
|
break
|
||||||
else:
|
else:
|
||||||
raise ExtractorError(u'Could not find chapter in chapter information')
|
raise ExtractorError('Could not find chapter in chapter information')
|
||||||
|
|
||||||
video_url = a.find('./video_file_url').text
|
video_url = a.find('./video_file_url').text
|
||||||
video_ext = video_url.rpartition('.')[2] or u'flv'
|
video_ext = video_url.rpartition('.')[2] or 'flv'
|
||||||
|
|
||||||
chapter_api_url = u'https://api.twitch.tv/kraken/videos/c' + chapter_id
|
chapter_api_url = 'https://api.twitch.tv/kraken/videos/c' + chapter_id
|
||||||
chapter_info_json = self._download_webpage(chapter_api_url, u'c' + chapter_id,
|
chapter_info = self._download_json(
|
||||||
note='Downloading chapter metadata',
|
chapter_api_url, 'c' + chapter_id,
|
||||||
errnote='Download of chapter metadata failed')
|
note='Downloading chapter metadata',
|
||||||
chapter_info = json.loads(chapter_info_json)
|
errnote='Download of chapter metadata failed')
|
||||||
|
|
||||||
bracket_start = int(doc.find('.//bracket_start').text)
|
bracket_start = int(doc.find('.//bracket_start').text)
|
||||||
bracket_end = int(doc.find('.//bracket_end').text)
|
bracket_end = int(doc.find('.//bracket_end').text)
|
||||||
|
|
||||||
# TODO determine start (and probably fix up file)
|
# TODO determine start (and probably fix up file)
|
||||||
# youtube-dl -v http://www.twitch.tv/firmbelief/c/1757457
|
# youtube-dl -v http://www.twitch.tv/firmbelief/c/1757457
|
||||||
#video_url += u'?start=' + TODO:start_timestamp
|
#video_url += '?start=' + TODO:start_timestamp
|
||||||
# bracket_start is 13290, but we want 51670615
|
# bracket_start is 13290, but we want 51670615
|
||||||
self._downloader.report_warning(u'Chapter detected, but we can just download the whole file. '
|
self._downloader.report_warning('Chapter detected, but we can just download the whole file. '
|
||||||
u'Chapter starts at %s and ends at %s' % (formatSeconds(bracket_start), formatSeconds(bracket_end)))
|
'Chapter starts at %s and ends at %s' % (formatSeconds(bracket_start), formatSeconds(bracket_end)))
|
||||||
|
|
||||||
info = {
|
info = {
|
||||||
'id': u'c' + chapter_id,
|
'id': 'c' + chapter_id,
|
||||||
'url': video_url,
|
'url': video_url,
|
||||||
'ext': video_ext,
|
'ext': video_ext,
|
||||||
'title': chapter_info['title'],
|
'title': chapter_info['title'],
|
||||||
@@ -131,14 +130,12 @@ class JustinTVIE(InfoExtractor):
|
|||||||
'uploader': chapter_info['channel']['display_name'],
|
'uploader': chapter_info['channel']['display_name'],
|
||||||
'uploader_id': chapter_info['channel']['name'],
|
'uploader_id': chapter_info['channel']['name'],
|
||||||
}
|
}
|
||||||
return [info]
|
return info
|
||||||
else:
|
else:
|
||||||
video_id = mobj.group('videoid')
|
video_id = mobj.group('videoid')
|
||||||
api = api_base + '/broadcast/by_archive/%s.json' % video_id
|
api = api_base + '/broadcast/by_archive/%s.json' % video_id
|
||||||
|
|
||||||
self.report_extraction(video_id)
|
entries = []
|
||||||
|
|
||||||
info = []
|
|
||||||
offset = 0
|
offset = 0
|
||||||
limit = self._JUSTIN_PAGE_LIMIT
|
limit = self._JUSTIN_PAGE_LIMIT
|
||||||
while True:
|
while True:
|
||||||
@@ -146,8 +143,12 @@ class JustinTVIE(InfoExtractor):
|
|||||||
self.report_download_page(video_id, offset)
|
self.report_download_page(video_id, offset)
|
||||||
page_url = api + ('?offset=%d&limit=%d' % (offset, limit))
|
page_url = api + ('?offset=%d&limit=%d' % (offset, limit))
|
||||||
page_count, page_info = self._parse_page(page_url, video_id)
|
page_count, page_info = self._parse_page(page_url, video_id)
|
||||||
info.extend(page_info)
|
entries.extend(page_info)
|
||||||
if not paged or page_count != limit:
|
if not paged or page_count != limit:
|
||||||
break
|
break
|
||||||
offset += limit
|
offset += limit
|
||||||
return info
|
return {
|
||||||
|
'_type': 'playlist',
|
||||||
|
'id': video_id,
|
||||||
|
'entries': entries,
|
||||||
|
}
|
||||||
|
@@ -1,3 +1,5 @@
|
|||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
|
|
||||||
@@ -11,22 +13,22 @@ from ..aes import (
|
|||||||
aes_decrypt_text
|
aes_decrypt_text
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class KeezMoviesIE(InfoExtractor):
|
class KeezMoviesIE(InfoExtractor):
|
||||||
_VALID_URL = r'^(?:https?://)?(?:www\.)?(?P<url>keezmovies\.com/video/.+?(?P<videoid>[0-9]+))(?:[/?&]|$)'
|
_VALID_URL = r'^https?://(?:www\.)?keezmovies\.com/video/.+?(?P<videoid>[0-9]+)(?:[/?&]|$)'
|
||||||
_TEST = {
|
_TEST = {
|
||||||
u'url': u'http://www.keezmovies.com/video/petite-asian-lady-mai-playing-in-bathtub-1214711',
|
'url': 'http://www.keezmovies.com/video/petite-asian-lady-mai-playing-in-bathtub-1214711',
|
||||||
u'file': u'1214711.mp4',
|
'file': '1214711.mp4',
|
||||||
u'md5': u'6e297b7e789329923fcf83abb67c9289',
|
'md5': '6e297b7e789329923fcf83abb67c9289',
|
||||||
u'info_dict': {
|
'info_dict': {
|
||||||
u"title": u"Petite Asian Lady Mai Playing In Bathtub",
|
'title': 'Petite Asian Lady Mai Playing In Bathtub',
|
||||||
u"age_limit": 18,
|
'age_limit': 18,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
mobj = re.match(self._VALID_URL, url)
|
mobj = re.match(self._VALID_URL, url)
|
||||||
video_id = mobj.group('videoid')
|
video_id = mobj.group('videoid')
|
||||||
url = 'http://www.' + mobj.group('url')
|
|
||||||
|
|
||||||
req = compat_urllib_request.Request(url)
|
req = compat_urllib_request.Request(url)
|
||||||
req.add_header('Cookie', 'age_verified=1')
|
req.add_header('Cookie', 'age_verified=1')
|
||||||
@@ -38,10 +40,10 @@ class KeezMoviesIE(InfoExtractor):
|
|||||||
embedded_url = mobj.group(1)
|
embedded_url = mobj.group(1)
|
||||||
return self.url_result(embedded_url)
|
return self.url_result(embedded_url)
|
||||||
|
|
||||||
video_title = self._html_search_regex(r'<h1 [^>]*>([^<]+)', webpage, u'title')
|
video_title = self._html_search_regex(r'<h1 [^>]*>([^<]+)', webpage, 'title')
|
||||||
video_url = compat_urllib_parse.unquote(self._html_search_regex(r'video_url=(.+?)&', webpage, u'video_url'))
|
video_url = compat_urllib_parse.unquote(self._html_search_regex(r'video_url=(.+?)&', webpage, 'video_url'))
|
||||||
if webpage.find('encrypted=true')!=-1:
|
if 'encrypted=true' in webpage:
|
||||||
password = self._html_search_regex(r'video_title=(.+?)&', webpage, u'password')
|
password = self._html_search_regex(r'video_title=(.+?)&', webpage, 'password')
|
||||||
video_url = aes_decrypt_text(video_url, password, 32).decode('utf-8')
|
video_url = aes_decrypt_text(video_url, password, 32).decode('utf-8')
|
||||||
path = compat_urllib_parse_urlparse(video_url).path
|
path = compat_urllib_parse_urlparse(video_url).path
|
||||||
extension = os.path.splitext(path)[1][1:]
|
extension = os.path.splitext(path)[1][1:]
|
||||||
|
@@ -1,17 +1,9 @@
|
|||||||
# coding: utf-8
|
# coding: utf-8
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import hashlib
|
|
||||||
import json
|
|
||||||
import re
|
import re
|
||||||
import time
|
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..utils import (
|
|
||||||
compat_parse_qs,
|
|
||||||
compat_str,
|
|
||||||
int_or_none,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class MorningstarIE(InfoExtractor):
|
class MorningstarIE(InfoExtractor):
|
||||||
|
@@ -44,7 +44,7 @@ class MotorsportIE(InfoExtractor):
|
|||||||
e = compat_str(int(time.time()) + 24 * 60 * 60)
|
e = compat_str(int(time.time()) + 24 * 60 * 60)
|
||||||
base_video_url = params['location'] + '?e=' + e
|
base_video_url = params['location'] + '?e=' + e
|
||||||
s = 'h3hg713fh32'
|
s = 'h3hg713fh32'
|
||||||
h = hashlib.md5(s + base_video_url).hexdigest()
|
h = hashlib.md5((s + base_video_url).encode('utf-8')).hexdigest()
|
||||||
video_url = base_video_url + '&h=' + h
|
video_url = base_video_url + '&h=' + h
|
||||||
|
|
||||||
uploader = self._html_search_regex(
|
uploader = self._html_search_regex(
|
||||||
|
27
youtube_dl/extractor/movshare.py
Normal file
27
youtube_dl/extractor/movshare.py
Normal file
@@ -0,0 +1,27 @@
|
|||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
from .novamov import NovaMovIE
|
||||||
|
|
||||||
|
|
||||||
|
class MovShareIE(NovaMovIE):
|
||||||
|
IE_NAME = 'movshare'
|
||||||
|
IE_DESC = 'MovShare'
|
||||||
|
|
||||||
|
_VALID_URL = NovaMovIE._VALID_URL_TEMPLATE % {'host': 'movshare\.(?:net|sx|ag)'}
|
||||||
|
|
||||||
|
_HOST = 'www.movshare.net'
|
||||||
|
|
||||||
|
_FILE_DELETED_REGEX = r'>This file no longer exists on our servers.<'
|
||||||
|
_TITLE_REGEX = r'<strong>Title:</strong> ([^<]+)</p>'
|
||||||
|
_DESCRIPTION_REGEX = r'<strong>Description:</strong> ([^<]+)</p>'
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://www.movshare.net/video/559e28be54d96',
|
||||||
|
'md5': 'abd31a2132947262c50429e1d16c1bfd',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '559e28be54d96',
|
||||||
|
'ext': 'flv',
|
||||||
|
'title': 'dissapeared image',
|
||||||
|
'description': 'optical illusion dissapeared image magic illusion',
|
||||||
|
}
|
||||||
|
}
|
@@ -13,7 +13,8 @@ class NovaMovIE(InfoExtractor):
|
|||||||
IE_NAME = 'novamov'
|
IE_NAME = 'novamov'
|
||||||
IE_DESC = 'NovaMov'
|
IE_DESC = 'NovaMov'
|
||||||
|
|
||||||
_VALID_URL = r'http://(?:(?:www\.)?%(host)s/video/|(?:(?:embed|www)\.)%(host)s/embed\.php\?(?:.*?&)?v=)(?P<videoid>[a-z\d]{13})' % {'host': 'novamov\.com'}
|
_VALID_URL_TEMPLATE = r'http://(?:(?:www\.)?%(host)s/(?:file|video)/|(?:(?:embed|www)\.)%(host)s/embed\.php\?(?:.*?&)?v=)(?P<id>[a-z\d]{13})'
|
||||||
|
_VALID_URL = _VALID_URL_TEMPLATE % {'host': 'novamov\.com'}
|
||||||
|
|
||||||
_HOST = 'www.novamov.com'
|
_HOST = 'www.novamov.com'
|
||||||
|
|
||||||
@@ -36,18 +37,17 @@ class NovaMovIE(InfoExtractor):
|
|||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
mobj = re.match(self._VALID_URL, url)
|
mobj = re.match(self._VALID_URL, url)
|
||||||
video_id = mobj.group('videoid')
|
video_id = mobj.group('id')
|
||||||
|
|
||||||
page = self._download_webpage(
|
page = self._download_webpage(
|
||||||
'http://%s/video/%s' % (self._HOST, video_id), video_id, 'Downloading video page')
|
'http://%s/video/%s' % (self._HOST, video_id), video_id, 'Downloading video page')
|
||||||
|
|
||||||
if re.search(self._FILE_DELETED_REGEX, page) is not None:
|
if re.search(self._FILE_DELETED_REGEX, page) is not None:
|
||||||
raise ExtractorError(u'Video %s does not exist' % video_id, expected=True)
|
raise ExtractorError('Video %s does not exist' % video_id, expected=True)
|
||||||
|
|
||||||
filekey = self._search_regex(self._FILEKEY_REGEX, page, 'filekey')
|
filekey = self._search_regex(self._FILEKEY_REGEX, page, 'filekey')
|
||||||
|
|
||||||
title = self._html_search_regex(self._TITLE_REGEX, page, 'title', fatal=False)
|
title = self._html_search_regex(self._TITLE_REGEX, page, 'title', fatal=False)
|
||||||
|
|
||||||
description = self._html_search_regex(self._DESCRIPTION_REGEX, page, 'description', default='', fatal=False)
|
description = self._html_search_regex(self._DESCRIPTION_REGEX, page, 'description', default='', fatal=False)
|
||||||
|
|
||||||
api_response = self._download_webpage(
|
api_response = self._download_webpage(
|
||||||
|
@@ -7,7 +7,7 @@ class NowVideoIE(NovaMovIE):
|
|||||||
IE_NAME = 'nowvideo'
|
IE_NAME = 'nowvideo'
|
||||||
IE_DESC = 'NowVideo'
|
IE_DESC = 'NowVideo'
|
||||||
|
|
||||||
_VALID_URL = r'http://(?:(?:www\.)?%(host)s/video/|(?:(?:embed|www)\.)%(host)s/embed\.php\?(?:.*?&)?v=)(?P<videoid>[a-z\d]{13})' % {'host': 'nowvideo\.(?:ch|sx|eu)'}
|
_VALID_URL = NovaMovIE._VALID_URL_TEMPLATE % {'host': 'nowvideo\.(?:ch|sx|eu|at|ag|co)'}
|
||||||
|
|
||||||
_HOST = 'www.nowvideo.ch'
|
_HOST = 'www.nowvideo.ch'
|
||||||
|
|
||||||
|
@@ -1,44 +1,81 @@
|
|||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import re
|
import re
|
||||||
|
import json
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..utils import compat_urllib_parse
|
from ..utils import int_or_none
|
||||||
|
|
||||||
|
|
||||||
class PornHdIE(InfoExtractor):
|
class PornHdIE(InfoExtractor):
|
||||||
_VALID_URL = r'(?:http://)?(?:www\.)?pornhd\.com/(?:[a-z]{2,4}/)?videos/(?P<video_id>[0-9]+)/(?P<video_title>.+)'
|
_VALID_URL = r'http://(?:www\.)?pornhd\.com/(?:[a-z]{2,4}/)?videos/(?P<id>\d+)'
|
||||||
_TEST = {
|
_TEST = {
|
||||||
'url': 'http://www.pornhd.com/videos/1962/sierra-day-gets-his-cum-all-over-herself-hd-porn-video',
|
'url': 'http://www.pornhd.com/videos/1962/sierra-day-gets-his-cum-all-over-herself-hd-porn-video',
|
||||||
'file': '1962.flv',
|
'md5': '956b8ca569f7f4d8ec563e2c41598441',
|
||||||
'md5': '35272469887dca97abd30abecc6cdf75',
|
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
"title": "sierra-day-gets-his-cum-all-over-herself-hd-porn-video",
|
'id': '1962',
|
||||||
"age_limit": 18,
|
'ext': 'mp4',
|
||||||
|
'title': 'Sierra loves doing laundry',
|
||||||
|
'description': 'md5:8ff0523848ac2b8f9b065ba781ccf294',
|
||||||
|
'age_limit': 18,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
mobj = re.match(self._VALID_URL, url)
|
mobj = re.match(self._VALID_URL, url)
|
||||||
|
video_id = mobj.group('id')
|
||||||
video_id = mobj.group('video_id')
|
|
||||||
video_title = mobj.group('video_title')
|
|
||||||
|
|
||||||
webpage = self._download_webpage(url, video_id)
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
next_url = self._html_search_regex(
|
title = self._og_search_title(webpage)
|
||||||
r'&hd=(http.+?)&', webpage, 'video URL')
|
TITLE_SUFFIX = ' porn HD Video | PornHD.com '
|
||||||
next_url = compat_urllib_parse.unquote(next_url)
|
if title.endswith(TITLE_SUFFIX):
|
||||||
|
title = title[:-len(TITLE_SUFFIX)]
|
||||||
|
|
||||||
video_url = self._download_webpage(
|
description = self._html_search_regex(
|
||||||
next_url, video_id, note='Retrieving video URL',
|
r'<div class="description">([^<]+)</div>', webpage, 'description', fatal=False)
|
||||||
errnote='Could not retrieve video URL')
|
view_count = int_or_none(self._html_search_regex(
|
||||||
age_limit = 18
|
r'(\d+) views </span>', webpage, 'view count', fatal=False))
|
||||||
|
|
||||||
|
formats = [
|
||||||
|
{
|
||||||
|
'url': format_url,
|
||||||
|
'ext': format.lower(),
|
||||||
|
'format_id': '%s-%s' % (format.lower(), quality.lower()),
|
||||||
|
'quality': 1 if quality.lower() == 'high' else 0,
|
||||||
|
} for format, quality, format_url in re.findall(
|
||||||
|
r'var __video([\da-zA-Z]+?)(Low|High)StreamUrl = \'(http://.+?)\?noProxy=1\'', webpage)
|
||||||
|
]
|
||||||
|
|
||||||
|
mobj = re.search(r'flashVars = (?P<flashvars>{.+?});', webpage)
|
||||||
|
if mobj:
|
||||||
|
flashvars = json.loads(mobj.group('flashvars'))
|
||||||
|
formats.extend([
|
||||||
|
{
|
||||||
|
'url': flashvars['hashlink'].replace('?noProxy=1', ''),
|
||||||
|
'ext': 'flv',
|
||||||
|
'format_id': 'flv-low',
|
||||||
|
'quality': 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
'url': flashvars['hd'].replace('?noProxy=1', ''),
|
||||||
|
'ext': 'flv',
|
||||||
|
'format_id': 'flv-high',
|
||||||
|
'quality': 1,
|
||||||
|
}
|
||||||
|
])
|
||||||
|
thumbnail = flashvars['urlWallpaper']
|
||||||
|
else:
|
||||||
|
thumbnail = self._og_search_thumbnail(webpage)
|
||||||
|
|
||||||
|
self._sort_formats(formats)
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'url': video_url,
|
'title': title,
|
||||||
'ext': 'flv',
|
'description': description,
|
||||||
'title': video_title,
|
'thumbnail': thumbnail,
|
||||||
'age_limit': age_limit,
|
'view_count': view_count,
|
||||||
|
'formats': formats,
|
||||||
|
'age_limit': 18,
|
||||||
}
|
}
|
||||||
|
@@ -18,7 +18,7 @@ class Ro220IE(InfoExtractor):
|
|||||||
'md5': '03af18b73a07b4088753930db7a34add',
|
'md5': '03af18b73a07b4088753930db7a34add',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
"title": "Luati-le Banii sez 4 ep 1",
|
"title": "Luati-le Banii sez 4 ep 1",
|
||||||
"description": "Iata-ne reveniti dupa o binemeritata vacanta. Va astept si pe Facebook cu pareri si comentarii.",
|
"description": "re:^Iata-ne reveniti dupa o binemeritata vacanta\. +Va astept si pe Facebook cu pareri si comentarii.$",
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -9,46 +9,136 @@ from ..utils import (
|
|||||||
parse_duration,
|
parse_duration,
|
||||||
parse_iso8601,
|
parse_iso8601,
|
||||||
unescapeHTML,
|
unescapeHTML,
|
||||||
|
compat_str,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class RTSIE(InfoExtractor):
|
class RTSIE(InfoExtractor):
|
||||||
IE_DESC = 'RTS.ch'
|
IE_DESC = 'RTS.ch'
|
||||||
_VALID_URL = r'^https?://(?:www\.)?rts\.ch/archives/tv/[^/]+/(?P<id>[0-9]+)-.*?\.html'
|
_VALID_URL = r'^https?://(?:www\.)?rts\.ch/(?:[^/]+/){2,}(?P<id>[0-9]+)-.*?\.html'
|
||||||
|
|
||||||
_TEST = {
|
_TESTS = [
|
||||||
'url': 'http://www.rts.ch/archives/tv/divers/3449373-les-enfants-terribles.html',
|
{
|
||||||
'md5': '753b877968ad8afaeddccc374d4256a5',
|
'url': 'http://www.rts.ch/archives/tv/divers/3449373-les-enfants-terribles.html',
|
||||||
'info_dict': {
|
'md5': '753b877968ad8afaeddccc374d4256a5',
|
||||||
'id': '3449373',
|
'info_dict': {
|
||||||
'ext': 'mp4',
|
'id': '3449373',
|
||||||
'duration': 1488,
|
'ext': 'mp4',
|
||||||
'title': 'Les Enfants Terribles',
|
'duration': 1488,
|
||||||
'description': 'France Pommier et sa soeur Luce Feral, les deux filles de ce groupe de 5.',
|
'title': 'Les Enfants Terribles',
|
||||||
'uploader': 'Divers',
|
'description': 'France Pommier et sa soeur Luce Feral, les deux filles de ce groupe de 5.',
|
||||||
'upload_date': '19680921',
|
'uploader': 'Divers',
|
||||||
'timestamp': -40280400,
|
'upload_date': '19680921',
|
||||||
'thumbnail': 're:^https?://.*\.image'
|
'timestamp': -40280400,
|
||||||
|
'thumbnail': 're:^https?://.*\.image'
|
||||||
|
},
|
||||||
},
|
},
|
||||||
}
|
{
|
||||||
|
'url': 'http://www.rts.ch/emissions/passe-moi-les-jumelles/5624067-entre-ciel-et-mer.html',
|
||||||
|
'md5': 'c148457a27bdc9e5b1ffe081a7a8337b',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '5624067',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'duration': 3720,
|
||||||
|
'title': 'Les yeux dans les cieux - Mon homard au Canada',
|
||||||
|
'description': 'md5:d22ee46f5cc5bac0912e5a0c6d44a9f7',
|
||||||
|
'uploader': 'Passe-moi les jumelles',
|
||||||
|
'upload_date': '20140404',
|
||||||
|
'timestamp': 1396635300,
|
||||||
|
'thumbnail': 're:^https?://.*\.image'
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
'url': 'http://www.rts.ch/video/sport/hockey/5745975-1-2-kloten-fribourg-5-2-second-but-pour-gotteron-par-kwiatowski.html',
|
||||||
|
'md5': 'b4326fecd3eb64a458ba73c73e91299d',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '5745975',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'duration': 48,
|
||||||
|
'title': '1/2, Kloten - Fribourg (5-2): second but pour Gottéron par Kwiatowski',
|
||||||
|
'description': 'Hockey - Playoff',
|
||||||
|
'uploader': 'Hockey',
|
||||||
|
'upload_date': '20140403',
|
||||||
|
'timestamp': 1396556882,
|
||||||
|
'thumbnail': 're:^https?://.*\.image'
|
||||||
|
},
|
||||||
|
'skip': 'Blocked outside Switzerland',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
'url': 'http://www.rts.ch/video/info/journal-continu/5745356-londres-cachee-par-un-epais-smog.html',
|
||||||
|
'md5': '9bb06503773c07ce83d3cbd793cebb91',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '5745356',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'duration': 33,
|
||||||
|
'title': 'Londres cachée par un épais smog',
|
||||||
|
'description': 'Un important voile de smog recouvre Londres depuis mercredi, provoqué par la pollution et du sable du Sahara.',
|
||||||
|
'uploader': 'Le Journal en continu',
|
||||||
|
'upload_date': '20140403',
|
||||||
|
'timestamp': 1396537322,
|
||||||
|
'thumbnail': 're:^https?://.*\.image'
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
'url': 'http://www.rts.ch/audio/couleur3/programmes/la-belle-video-de-stephane-laurenceau/5706148-urban-hippie-de-damien-krisl-03-04-2014.html',
|
||||||
|
'md5': 'dd8ef6a22dff163d063e2a52bc8adcae',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '5706148',
|
||||||
|
'ext': 'mp3',
|
||||||
|
'duration': 123,
|
||||||
|
'title': '"Urban Hippie", de Damien Krisl',
|
||||||
|
'description': 'Des Hippies super glam.',
|
||||||
|
'upload_date': '20140403',
|
||||||
|
'timestamp': 1396551600,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
m = re.match(self._VALID_URL, url)
|
m = re.match(self._VALID_URL, url)
|
||||||
video_id = m.group('id')
|
video_id = m.group('id')
|
||||||
|
|
||||||
all_info = self._download_json(
|
def download_json(internal_id):
|
||||||
'http://www.rts.ch/a/%s.html?f=json/article' % video_id, video_id)
|
return self._download_json(
|
||||||
info = all_info['video']['JSONinfo']
|
'http://www.rts.ch/a/%s.html?f=json/article' % internal_id,
|
||||||
|
video_id)
|
||||||
|
|
||||||
|
all_info = download_json(video_id)
|
||||||
|
|
||||||
|
# video_id extracted out of URL is not always a real id
|
||||||
|
if 'video' not in all_info and 'audio' not in all_info:
|
||||||
|
page = self._download_webpage(url, video_id)
|
||||||
|
internal_id = self._html_search_regex(
|
||||||
|
r'<(?:video|audio) data-id="([0-9]+)"', page,
|
||||||
|
'internal video id')
|
||||||
|
all_info = download_json(internal_id)
|
||||||
|
|
||||||
|
info = all_info['video']['JSONinfo'] if 'video' in all_info else all_info['audio']
|
||||||
|
|
||||||
upload_timestamp = parse_iso8601(info.get('broadcast_date'))
|
upload_timestamp = parse_iso8601(info.get('broadcast_date'))
|
||||||
duration = parse_duration(info.get('duration'))
|
duration = info.get('duration') or info.get('cutout') or info.get('cutduration')
|
||||||
|
if isinstance(duration, compat_str):
|
||||||
|
duration = parse_duration(duration)
|
||||||
|
view_count = info.get('plays')
|
||||||
thumbnail = unescapeHTML(info.get('preview_image_url'))
|
thumbnail = unescapeHTML(info.get('preview_image_url'))
|
||||||
|
|
||||||
|
def extract_bitrate(url):
|
||||||
|
return int_or_none(self._search_regex(
|
||||||
|
r'-([0-9]+)k\.', url, 'bitrate', default=None))
|
||||||
|
|
||||||
formats = [{
|
formats = [{
|
||||||
'format_id': fid,
|
'format_id': fid,
|
||||||
'url': furl,
|
'url': furl,
|
||||||
'tbr': int_or_none(self._search_regex(
|
'tbr': extract_bitrate(furl),
|
||||||
r'-([0-9]+)k\.', furl, 'bitrate', default=None)),
|
|
||||||
} for fid, furl in info['streams'].items()]
|
} for fid, furl in info['streams'].items()]
|
||||||
|
|
||||||
|
if 'media' in info:
|
||||||
|
formats.extend([{
|
||||||
|
'format_id': '%s-%sk' % (media['ext'], media['rate']),
|
||||||
|
'url': 'http://download-video.rts.ch/%s' % media['url'],
|
||||||
|
'tbr': media['rate'] or extract_bitrate(media['url']),
|
||||||
|
} for media in info['media'] if media.get('rate')])
|
||||||
|
|
||||||
self._sort_formats(formats)
|
self._sort_formats(formats)
|
||||||
|
|
||||||
return {
|
return {
|
||||||
@@ -57,6 +147,7 @@ class RTSIE(InfoExtractor):
|
|||||||
'title': info['title'],
|
'title': info['title'],
|
||||||
'description': info.get('intro'),
|
'description': info.get('intro'),
|
||||||
'duration': duration,
|
'duration': duration,
|
||||||
|
'view_count': view_count,
|
||||||
'uploader': info.get('programName'),
|
'uploader': info.get('programName'),
|
||||||
'timestamp': upload_timestamp,
|
'timestamp': upload_timestamp,
|
||||||
'thumbnail': thumbnail,
|
'thumbnail': thumbnail,
|
||||||
|
84
youtube_dl/extractor/rtve.py
Normal file
84
youtube_dl/extractor/rtve.py
Normal file
@@ -0,0 +1,84 @@
|
|||||||
|
# encoding: utf-8
|
||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import re
|
||||||
|
import base64
|
||||||
|
|
||||||
|
from .common import InfoExtractor
|
||||||
|
from ..utils import (
|
||||||
|
struct_unpack,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class RTVEALaCartaIE(InfoExtractor):
|
||||||
|
IE_NAME = 'rtve.es:alacarta'
|
||||||
|
IE_DESC = 'RTVE a la carta'
|
||||||
|
_VALID_URL = r'http://www\.rtve\.es/alacarta/videos/[^/]+/[^/]+/(?P<id>\d+)'
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://www.rtve.es/alacarta/videos/balonmano/o-swiss-cup-masculina-final-espana-suecia/2491869/',
|
||||||
|
'md5': '18fcd45965bdd076efdb12cd7f6d7b9e',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '2491869',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Balonmano - Swiss Cup masculina. Final: España-Suecia',
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
def _decrypt_url(self, png):
|
||||||
|
encrypted_data = base64.b64decode(png)
|
||||||
|
text_index = encrypted_data.find(b'tEXt')
|
||||||
|
text_chunk = encrypted_data[text_index-4:]
|
||||||
|
length = struct_unpack('!I', text_chunk[:4])[0]
|
||||||
|
# Use bytearray to get integers when iterating in both python 2.x and 3.x
|
||||||
|
data = bytearray(text_chunk[8:8+length])
|
||||||
|
data = [chr(b) for b in data if b != 0]
|
||||||
|
hash_index = data.index('#')
|
||||||
|
alphabet_data = data[:hash_index]
|
||||||
|
url_data = data[hash_index+1:]
|
||||||
|
|
||||||
|
alphabet = []
|
||||||
|
e = 0
|
||||||
|
d = 0
|
||||||
|
for l in alphabet_data:
|
||||||
|
if d == 0:
|
||||||
|
alphabet.append(l)
|
||||||
|
d = e = (e + 1) % 4
|
||||||
|
else:
|
||||||
|
d -= 1
|
||||||
|
url = ''
|
||||||
|
f = 0
|
||||||
|
e = 3
|
||||||
|
b = 1
|
||||||
|
for letter in url_data:
|
||||||
|
if f == 0:
|
||||||
|
l = int(letter)*10
|
||||||
|
f = 1
|
||||||
|
else:
|
||||||
|
if e == 0:
|
||||||
|
l += int(letter)
|
||||||
|
url += alphabet[l]
|
||||||
|
e = (b + 3) % 4
|
||||||
|
f = 0
|
||||||
|
b += 1
|
||||||
|
else:
|
||||||
|
e -= 1
|
||||||
|
|
||||||
|
return url
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
mobj = re.match(self._VALID_URL, url)
|
||||||
|
video_id = mobj.group('id')
|
||||||
|
info = self._download_json(
|
||||||
|
'http://www.rtve.es/api/videos/%s/config/alacarta_videos.json' % video_id,
|
||||||
|
video_id)['page']['items'][0]
|
||||||
|
png_url = 'http://www.rtve.es/ztnr/movil/thumbnail/default/videos/%s.png' % video_id
|
||||||
|
png = self._download_webpage(png_url, video_id, 'Downloading url information')
|
||||||
|
video_url = self._decrypt_url(png)
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'title': info['title'],
|
||||||
|
'url': video_url,
|
||||||
|
'thumbnail': info['image'],
|
||||||
|
}
|
@@ -9,8 +9,18 @@ from ..utils import (
|
|||||||
|
|
||||||
|
|
||||||
class TeamcocoIE(InfoExtractor):
|
class TeamcocoIE(InfoExtractor):
|
||||||
_VALID_URL = r'http://teamcoco\.com/video/(?P<url_title>.*)'
|
_VALID_URL = r'http://teamcoco\.com/video/(?P<video_id>[0-9]+)?/?(?P<display_id>.*)'
|
||||||
_TEST = {
|
_TESTS = [
|
||||||
|
{
|
||||||
|
'url': 'http://teamcoco.com/video/80187/conan-becomes-a-mary-kay-beauty-consultant',
|
||||||
|
'file': '80187.mp4',
|
||||||
|
'md5': '3f7746aa0dc86de18df7539903d399ea',
|
||||||
|
'info_dict': {
|
||||||
|
'title': 'Conan Becomes A Mary Kay Beauty Consultant',
|
||||||
|
'description': 'Mary Kay is perhaps the most trusted name in female beauty, so of course Conan is a natural choice to sell their products.'
|
||||||
|
}
|
||||||
|
},
|
||||||
|
{
|
||||||
'url': 'http://teamcoco.com/video/louis-ck-interview-george-w-bush',
|
'url': 'http://teamcoco.com/video/louis-ck-interview-george-w-bush',
|
||||||
'file': '19705.mp4',
|
'file': '19705.mp4',
|
||||||
'md5': 'cde9ba0fa3506f5f017ce11ead928f9a',
|
'md5': 'cde9ba0fa3506f5f017ce11ead928f9a',
|
||||||
@@ -19,22 +29,23 @@ class TeamcocoIE(InfoExtractor):
|
|||||||
"title": "Louis C.K. Interview Pt. 1 11/3/11"
|
"title": "Louis C.K. Interview Pt. 1 11/3/11"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
mobj = re.match(self._VALID_URL, url)
|
mobj = re.match(self._VALID_URL, url)
|
||||||
if mobj is None:
|
|
||||||
raise ExtractorError('Invalid URL: %s' % url)
|
|
||||||
url_title = mobj.group('url_title')
|
|
||||||
webpage = self._download_webpage(url, url_title)
|
|
||||||
|
|
||||||
video_id = self._html_search_regex(
|
display_id = mobj.group('display_id')
|
||||||
r'<article class="video" data-id="(\d+?)"',
|
webpage = self._download_webpage(url, display_id)
|
||||||
webpage, 'video id')
|
|
||||||
|
video_id = mobj.group("video_id")
|
||||||
self.report_extraction(video_id)
|
if not video_id:
|
||||||
|
video_id = self._html_search_regex(
|
||||||
|
r'<article class="video" data-id="(\d+?)"',
|
||||||
|
webpage, 'video id')
|
||||||
|
|
||||||
data_url = 'http://teamcoco.com/cvp/2.0/%s.xml' % video_id
|
data_url = 'http://teamcoco.com/cvp/2.0/%s.xml' % video_id
|
||||||
data = self._download_xml(data_url, video_id, 'Downloading data webpage')
|
data = self._download_xml(
|
||||||
|
data_url, display_id, 'Downloading data webpage')
|
||||||
|
|
||||||
qualities = ['500k', '480p', '1000k', '720p', '1080p']
|
qualities = ['500k', '480p', '1000k', '720p', '1080p']
|
||||||
formats = []
|
formats = []
|
||||||
@@ -69,6 +80,7 @@ class TeamcocoIE(InfoExtractor):
|
|||||||
|
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
|
'display_id': display_id,
|
||||||
'formats': formats,
|
'formats': formats,
|
||||||
'title': self._og_search_title(webpage),
|
'title': self._og_search_title(webpage),
|
||||||
'thumbnail': self._og_search_thumbnail(webpage),
|
'thumbnail': self._og_search_thumbnail(webpage),
|
||||||
|
@@ -37,6 +37,7 @@ class TEDIE(SubtitlesInfoExtractor):
|
|||||||
'consciousness, but that half the time our brains are '
|
'consciousness, but that half the time our brains are '
|
||||||
'actively fooling us.'),
|
'actively fooling us.'),
|
||||||
'uploader': 'Dan Dennett',
|
'uploader': 'Dan Dennett',
|
||||||
|
'width': 854,
|
||||||
}
|
}
|
||||||
}, {
|
}, {
|
||||||
'url': 'http://www.ted.com/watch/ted-institute/ted-bcg/vishal-sikka-the-beauty-and-power-of-algorithms',
|
'url': 'http://www.ted.com/watch/ted-institute/ted-bcg/vishal-sikka-the-beauty-and-power-of-algorithms',
|
||||||
@@ -50,10 +51,10 @@ class TEDIE(SubtitlesInfoExtractor):
|
|||||||
}
|
}
|
||||||
}]
|
}]
|
||||||
|
|
||||||
_FORMATS_PREFERENCE = {
|
_NATIVE_FORMATS = {
|
||||||
'low': 1,
|
'low': {'preference': 1, 'width': 320, 'height': 180},
|
||||||
'medium': 2,
|
'medium': {'preference': 2, 'width': 512, 'height': 288},
|
||||||
'high': 3,
|
'high': {'preference': 3, 'width': 854, 'height': 480},
|
||||||
}
|
}
|
||||||
|
|
||||||
def _extract_info(self, webpage):
|
def _extract_info(self, webpage):
|
||||||
@@ -98,12 +99,14 @@ class TEDIE(SubtitlesInfoExtractor):
|
|||||||
talk_info = self._extract_info(webpage)['talks'][0]
|
talk_info = self._extract_info(webpage)['talks'][0]
|
||||||
|
|
||||||
formats = [{
|
formats = [{
|
||||||
'ext': 'mp4',
|
|
||||||
'url': format_url,
|
'url': format_url,
|
||||||
'format_id': format_id,
|
'format_id': format_id,
|
||||||
'format': format_id,
|
'format': format_id,
|
||||||
'preference': self._FORMATS_PREFERENCE.get(format_id, -1),
|
|
||||||
} for (format_id, format_url) in talk_info['nativeDownloads'].items()]
|
} for (format_id, format_url) in talk_info['nativeDownloads'].items()]
|
||||||
|
for f in formats:
|
||||||
|
finfo = self._NATIVE_FORMATS.get(f['format_id'])
|
||||||
|
if finfo:
|
||||||
|
f.update(finfo)
|
||||||
self._sort_formats(formats)
|
self._sort_formats(formats)
|
||||||
|
|
||||||
video_id = compat_str(talk_info['id'])
|
video_id = compat_str(talk_info['id'])
|
||||||
|
@@ -11,7 +11,7 @@ from ..utils import (
|
|||||||
|
|
||||||
|
|
||||||
class UstreamIE(InfoExtractor):
|
class UstreamIE(InfoExtractor):
|
||||||
_VALID_URL = r'https?://www\.ustream\.tv/recorded/(?P<videoID>\d+)'
|
_VALID_URL = r'https?://www\.ustream\.tv/(?P<type>recorded|embed)/(?P<videoID>\d+)'
|
||||||
IE_NAME = 'ustream'
|
IE_NAME = 'ustream'
|
||||||
_TEST = {
|
_TEST = {
|
||||||
'url': 'http://www.ustream.tv/recorded/20274954',
|
'url': 'http://www.ustream.tv/recorded/20274954',
|
||||||
@@ -25,6 +25,13 @@ class UstreamIE(InfoExtractor):
|
|||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
m = re.match(self._VALID_URL, url)
|
m = re.match(self._VALID_URL, url)
|
||||||
|
if m.group('type') == 'embed':
|
||||||
|
video_id = m.group('videoID')
|
||||||
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
desktop_video_id = self._html_search_regex(r'ContentVideoIds=\["([^"]*?)"\]', webpage, 'desktop_video_id')
|
||||||
|
desktop_url = 'http://www.ustream.tv/recorded/' + desktop_video_id
|
||||||
|
return self.url_result(desktop_url, 'Ustream')
|
||||||
|
|
||||||
video_id = m.group('videoID')
|
video_id = m.group('videoID')
|
||||||
|
|
||||||
video_url = 'http://tcdn.ustream.tv/video/%s' % video_id
|
video_url = 'http://tcdn.ustream.tv/video/%s' % video_id
|
||||||
|
26
youtube_dl/extractor/videoweed.py
Normal file
26
youtube_dl/extractor/videoweed.py
Normal file
@@ -0,0 +1,26 @@
|
|||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
from .novamov import NovaMovIE
|
||||||
|
|
||||||
|
|
||||||
|
class VideoWeedIE(NovaMovIE):
|
||||||
|
IE_NAME = 'videoweed'
|
||||||
|
IE_DESC = 'VideoWeed'
|
||||||
|
|
||||||
|
_VALID_URL = NovaMovIE._VALID_URL_TEMPLATE % {'host': 'videoweed\.(?:es|com)'}
|
||||||
|
|
||||||
|
_HOST = 'www.videoweed.es'
|
||||||
|
|
||||||
|
_FILE_DELETED_REGEX = r'>This file no longer exists on our servers.<'
|
||||||
|
_TITLE_REGEX = r'<h1 class="text_shadow">([^<]+)</h1>'
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://www.videoweed.es/file/b42178afbea14',
|
||||||
|
'md5': 'abd31a2132947262c50429e1d16c1bfd',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'b42178afbea14',
|
||||||
|
'ext': 'flv',
|
||||||
|
'title': 'optical illusion dissapeared image magic illusion',
|
||||||
|
'description': ''
|
||||||
|
},
|
||||||
|
}
|
@@ -15,22 +15,24 @@ from ..utils import (
|
|||||||
|
|
||||||
class YahooIE(InfoExtractor):
|
class YahooIE(InfoExtractor):
|
||||||
IE_DESC = 'Yahoo screen'
|
IE_DESC = 'Yahoo screen'
|
||||||
_VALID_URL = r'http://screen\.yahoo\.com/.*?-(?P<id>\d*?)\.html'
|
_VALID_URL = r'https?://screen\.yahoo\.com/.*?-(?P<id>[0-9]+)(?:-[a-z]+)?\.html'
|
||||||
_TESTS = [
|
_TESTS = [
|
||||||
{
|
{
|
||||||
'url': 'http://screen.yahoo.com/julian-smith-travis-legg-watch-214727115.html',
|
'url': 'http://screen.yahoo.com/julian-smith-travis-legg-watch-214727115.html',
|
||||||
'file': '214727115.mp4',
|
|
||||||
'md5': '4962b075c08be8690a922ee026d05e69',
|
'md5': '4962b075c08be8690a922ee026d05e69',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
|
'id': '214727115',
|
||||||
|
'ext': 'mp4',
|
||||||
'title': 'Julian Smith & Travis Legg Watch Julian Smith',
|
'title': 'Julian Smith & Travis Legg Watch Julian Smith',
|
||||||
'description': 'Julian and Travis watch Julian Smith',
|
'description': 'Julian and Travis watch Julian Smith',
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
'url': 'http://screen.yahoo.com/wired/codefellas-s1-ep12-cougar-lies-103000935.html',
|
'url': 'http://screen.yahoo.com/wired/codefellas-s1-ep12-cougar-lies-103000935.html',
|
||||||
'file': '103000935.mp4',
|
|
||||||
'md5': 'd6e6fc6e1313c608f316ddad7b82b306',
|
'md5': 'd6e6fc6e1313c608f316ddad7b82b306',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
|
'id': '103000935',
|
||||||
|
'ext': 'mp4',
|
||||||
'title': 'Codefellas - The Cougar Lies with Spanish Moss',
|
'title': 'Codefellas - The Cougar Lies with Spanish Moss',
|
||||||
'description': 'Agent Topple\'s mustache does its dirty work, and Nicole brokers a deal for peace. But why is the NSA collecting millions of Instagram brunch photos? And if your waffles have nothing to hide, what are they so worried about?',
|
'description': 'Agent Topple\'s mustache does its dirty work, and Nicole brokers a deal for peace. But why is the NSA collecting millions of Instagram brunch photos? And if your waffles have nothing to hide, what are they so worried about?',
|
||||||
},
|
},
|
||||||
@@ -60,10 +62,9 @@ class YahooIE(InfoExtractor):
|
|||||||
'env': 'prod',
|
'env': 'prod',
|
||||||
'format': 'json',
|
'format': 'json',
|
||||||
})
|
})
|
||||||
query_result_json = self._download_webpage(
|
query_result = self._download_json(
|
||||||
'http://video.query.yahoo.com/v1/public/yql?' + data,
|
'http://video.query.yahoo.com/v1/public/yql?' + data,
|
||||||
video_id, 'Downloading video info')
|
video_id, 'Downloading video info')
|
||||||
query_result = json.loads(query_result_json)
|
|
||||||
info = query_result['query']['results']['mediaObj'][0]
|
info = query_result['query']['results']['mediaObj'][0]
|
||||||
meta = info['meta']
|
meta = info['meta']
|
||||||
|
|
||||||
@@ -86,7 +87,6 @@ class YahooIE(InfoExtractor):
|
|||||||
else:
|
else:
|
||||||
format_url = compat_urlparse.urljoin(host, path)
|
format_url = compat_urlparse.urljoin(host, path)
|
||||||
format_info['url'] = format_url
|
format_info['url'] = format_url
|
||||||
|
|
||||||
formats.append(format_info)
|
formats.append(format_info)
|
||||||
|
|
||||||
self._sort_formats(formats)
|
self._sort_formats(formats)
|
||||||
@@ -134,27 +134,25 @@ class YahooSearchIE(SearchInfoExtractor):
|
|||||||
|
|
||||||
def _get_n_results(self, query, n):
|
def _get_n_results(self, query, n):
|
||||||
"""Get a specified number of results for a query"""
|
"""Get a specified number of results for a query"""
|
||||||
|
entries = []
|
||||||
res = {
|
for pagenum in itertools.count(0):
|
||||||
'_type': 'playlist',
|
|
||||||
'id': query,
|
|
||||||
'entries': []
|
|
||||||
}
|
|
||||||
for pagenum in itertools.count(0):
|
|
||||||
result_url = 'http://video.search.yahoo.com/search/?p=%s&fr=screen&o=js&gs=0&b=%d' % (compat_urllib_parse.quote_plus(query), pagenum * 30)
|
result_url = 'http://video.search.yahoo.com/search/?p=%s&fr=screen&o=js&gs=0&b=%d' % (compat_urllib_parse.quote_plus(query), pagenum * 30)
|
||||||
webpage = self._download_webpage(result_url, query,
|
info = self._download_json(result_url, query,
|
||||||
note='Downloading results page '+str(pagenum+1))
|
note='Downloading results page '+str(pagenum+1))
|
||||||
info = json.loads(webpage)
|
|
||||||
m = info['m']
|
m = info['m']
|
||||||
results = info['results']
|
results = info['results']
|
||||||
|
|
||||||
for (i, r) in enumerate(results):
|
for (i, r) in enumerate(results):
|
||||||
if (pagenum * 30) +i >= n:
|
if (pagenum * 30) + i >= n:
|
||||||
break
|
break
|
||||||
mobj = re.search(r'(?P<url>screen\.yahoo\.com/.*?-\d*?\.html)"', r)
|
mobj = re.search(r'(?P<url>screen\.yahoo\.com/.*?-\d*?\.html)"', r)
|
||||||
e = self.url_result('http://' + mobj.group('url'), 'Yahoo')
|
e = self.url_result('http://' + mobj.group('url'), 'Yahoo')
|
||||||
res['entries'].append(e)
|
entries.append(e)
|
||||||
if (pagenum * 30 +i >= n) or (m['last'] >= (m['total'] -1)):
|
if (pagenum * 30 + i >= n) or (m['last'] >= (m['total'] - 1)):
|
||||||
break
|
break
|
||||||
|
|
||||||
return res
|
return {
|
||||||
|
'_type': 'playlist',
|
||||||
|
'id': query,
|
||||||
|
'entries': entries,
|
||||||
|
}
|
||||||
|
@@ -2,6 +2,7 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
import calendar
|
import calendar
|
||||||
|
import codecs
|
||||||
import contextlib
|
import contextlib
|
||||||
import ctypes
|
import ctypes
|
||||||
import datetime
|
import datetime
|
||||||
@@ -909,25 +910,81 @@ def platform_name():
|
|||||||
return res
|
return res
|
||||||
|
|
||||||
|
|
||||||
def write_string(s, out=None):
|
def _windows_write_string(s, out):
|
||||||
|
""" Returns True if the string was written using special methods,
|
||||||
|
False if it has yet to be written out."""
|
||||||
|
# Adapted from http://stackoverflow.com/a/3259271/35070
|
||||||
|
|
||||||
|
import ctypes
|
||||||
|
import ctypes.wintypes
|
||||||
|
|
||||||
|
WIN_OUTPUT_IDS = {
|
||||||
|
1: -11,
|
||||||
|
2: -12,
|
||||||
|
}
|
||||||
|
|
||||||
|
fileno = out.fileno()
|
||||||
|
if fileno not in WIN_OUTPUT_IDS:
|
||||||
|
return False
|
||||||
|
|
||||||
|
GetStdHandle = ctypes.WINFUNCTYPE(
|
||||||
|
ctypes.wintypes.HANDLE, ctypes.wintypes.DWORD)(
|
||||||
|
("GetStdHandle", ctypes.windll.kernel32))
|
||||||
|
h = GetStdHandle(WIN_OUTPUT_IDS[fileno])
|
||||||
|
|
||||||
|
WriteConsoleW = ctypes.WINFUNCTYPE(
|
||||||
|
ctypes.wintypes.BOOL, ctypes.wintypes.HANDLE, ctypes.wintypes.LPWSTR,
|
||||||
|
ctypes.wintypes.DWORD, ctypes.POINTER(ctypes.wintypes.DWORD),
|
||||||
|
ctypes.wintypes.LPVOID)(("WriteConsoleW", ctypes.windll.kernel32))
|
||||||
|
written = ctypes.wintypes.DWORD(0)
|
||||||
|
|
||||||
|
GetFileType = ctypes.WINFUNCTYPE(ctypes.wintypes.DWORD, ctypes.wintypes.DWORD)(("GetFileType", ctypes.windll.kernel32))
|
||||||
|
FILE_TYPE_CHAR = 0x0002
|
||||||
|
FILE_TYPE_REMOTE = 0x8000
|
||||||
|
GetConsoleMode = ctypes.WINFUNCTYPE(
|
||||||
|
ctypes.wintypes.BOOL, ctypes.wintypes.HANDLE,
|
||||||
|
ctypes.POINTER(ctypes.wintypes.DWORD))(
|
||||||
|
("GetConsoleMode", ctypes.windll.kernel32))
|
||||||
|
INVALID_HANDLE_VALUE = ctypes.wintypes.DWORD(-1).value
|
||||||
|
|
||||||
|
def not_a_console(handle):
|
||||||
|
if handle == INVALID_HANDLE_VALUE or handle is None:
|
||||||
|
return True
|
||||||
|
return ((GetFileType(handle) & ~FILE_TYPE_REMOTE) != FILE_TYPE_CHAR
|
||||||
|
or GetConsoleMode(handle, ctypes.byref(ctypes.wintypes.DWORD())) == 0)
|
||||||
|
|
||||||
|
if not_a_console(h):
|
||||||
|
return False
|
||||||
|
|
||||||
|
remaining = len(s)
|
||||||
|
while remaining > 0:
|
||||||
|
ret = WriteConsoleW(
|
||||||
|
h, s, min(len(s), 1024), ctypes.byref(written), None)
|
||||||
|
if ret == 0:
|
||||||
|
raise OSError('Failed to write string')
|
||||||
|
remaining -= written.value
|
||||||
|
return True
|
||||||
|
|
||||||
|
|
||||||
|
def write_string(s, out=None, encoding=None):
|
||||||
if out is None:
|
if out is None:
|
||||||
out = sys.stderr
|
out = sys.stderr
|
||||||
assert type(s) == compat_str
|
assert type(s) == compat_str
|
||||||
|
|
||||||
|
if sys.platform == 'win32' and encoding is None and hasattr(out, 'fileno'):
|
||||||
|
if _windows_write_string(s, out):
|
||||||
|
return
|
||||||
|
|
||||||
if ('b' in getattr(out, 'mode', '') or
|
if ('b' in getattr(out, 'mode', '') or
|
||||||
sys.version_info[0] < 3): # Python 2 lies about mode of sys.stderr
|
sys.version_info[0] < 3): # Python 2 lies about mode of sys.stderr
|
||||||
s = s.encode(preferredencoding(), 'ignore')
|
byt = s.encode(encoding or preferredencoding(), 'ignore')
|
||||||
try:
|
out.write(byt)
|
||||||
|
elif hasattr(out, 'buffer'):
|
||||||
|
enc = encoding or getattr(out, 'encoding', None) or preferredencoding()
|
||||||
|
byt = s.encode(enc, 'ignore')
|
||||||
|
out.buffer.write(byt)
|
||||||
|
else:
|
||||||
out.write(s)
|
out.write(s)
|
||||||
except UnicodeEncodeError:
|
|
||||||
# In Windows shells, this can fail even when the codec is just charmap!?
|
|
||||||
# See https://wiki.python.org/moin/PrintFails#Issue
|
|
||||||
if sys.platform == 'win32' and hasattr(out, 'encoding'):
|
|
||||||
s = s.encode(out.encoding, 'ignore').decode(out.encoding)
|
|
||||||
out.write(s)
|
|
||||||
else:
|
|
||||||
raise
|
|
||||||
|
|
||||||
out.flush()
|
out.flush()
|
||||||
|
|
||||||
|
|
||||||
@@ -1263,9 +1320,11 @@ class PagedList(object):
|
|||||||
|
|
||||||
|
|
||||||
def uppercase_escape(s):
|
def uppercase_escape(s):
|
||||||
|
unicode_escape = codecs.getdecoder('unicode_escape')
|
||||||
return re.sub(
|
return re.sub(
|
||||||
r'\\U[0-9a-fA-F]{8}',
|
r'\\U[0-9a-fA-F]{8}',
|
||||||
lambda m: m.group(0).decode('unicode-escape'), s)
|
lambda m: unicode_escape(m.group(0))[0],
|
||||||
|
s)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
struct.pack(u'!I', 0)
|
struct.pack(u'!I', 0)
|
||||||
|
@@ -1,2 +1,2 @@
|
|||||||
|
|
||||||
__version__ = '2014.04.04.2'
|
__version__ = '2014.04.07.3'
|
||||||
|
Reference in New Issue
Block a user