--- /dev/null
+# coding: utf-8
+from __future__ import unicode_literals
+
+import os.path
+import time
+import datetime
+
+from .common import InfoExtractor
+from ..compat import (compat_urlparse, compat_urllib_parse)
+from ..utils import (ExtractorError, parse_iso8601)
+
+
+class LetvIE(InfoExtractor):
+ _VALID_URL = r'http://www.letv.com/ptv/vplay/(?P<id>\d+).html'
+
+ _TESTS = [{
+ 'url': 'http://www.letv.com/ptv/vplay/22005890.html',
+ 'md5': 'cab23bd68d5a8db9be31c9a222c1e8df',
+ 'info_dict': {
+ 'id': '22005890',
+ 'ext': 'mp4',
+ 'title': '第87届奥斯卡颁奖礼完美落幕 《鸟人》成最大赢家',
+ 'timestamp': 1424747397,
+ 'upload_date': '20150224',
+ }
+ }, {
+ 'url': 'http://www.letv.com/ptv/vplay/1118082.html',
+ 'info_dict': {
+ 'id': '1118082',
+ 'ext': 'mp4',
+ }
+ }]
+
+ @staticmethod
+ def urshift(val, n):
+ return val >> n if val >= 0 else (val + 0x100000000) >> n
+
+ # ror() and calcTimeKey() are reversed from a embedded swf file in KLetvPlayer.swf
+ def ror(self, param1, param2):
+ _loc3_ = 0
+ while _loc3_ < param2:
+ param1 = self.urshift(param1, 1) + ((param1 & 1) << 31)
+ _loc3_ += 1
+ return param1
+
+ def calcTimeKey(self, param1):
+ _loc2_ = 773625421
+ _loc3_ = self.ror(param1, _loc2_ % 13)
+ _loc3_ = _loc3_ ^ _loc2_
+ _loc3_ = self.ror(_loc3_, _loc2_ % 17)
+ return _loc3_
+
+ def _real_extract(self, url):
+ media_id = self._match_id(url)
+ page = self._download_webpage(url, media_id)
+ params = {
+ 'id': media_id,
+ 'platid': 1,
+ 'splatid': 101,
+ 'format': 1,
+ 'tkey': self.calcTimeKey(int(time.time())),
+ 'domain': 'www.letv.com'
+ }
+ play_json = self._download_json(
+ 'http://api.letv.com/mms/out/video/playJson?' + compat_urllib_parse.urlencode(params),
+ media_id, 'playJson data')
+
+ # Check for errors
+ playstatus = play_json['playstatus']
+ if playstatus['status'] == 0:
+ flag = playstatus['flag']
+ if flag == 1:
+ msg = 'Country %s auth error' % playstatus['country']
+ else:
+ msg = 'Generic error. flag = %d' % flag
+ raise ExtractorError(msg, expected=True)
+
+ playurl = play_json['playurl']
+
+ formats = ['350', '1000', '1300', '720p', '1080p']
+ dispatch = playurl['dispatch']
+
+ urls = []
+ for format_id in formats:
+ if format_id in dispatch:
+ media_url = playurl['domain'][0] + dispatch[format_id][0]
+
+ # Mimic what flvxz.com do
+ url_parts = list(compat_urlparse.urlparse(media_url))
+ qs = dict(compat_urlparse.parse_qs(url_parts[4]))
+ qs.update({
+ 'platid': '14',
+ 'splatid': '1401',
+ 'tss': 'no',
+ 'retry': 1
+ })
+ url_parts[4] = compat_urllib_parse.urlencode(qs)
+ media_url = compat_urlparse.urlunparse(url_parts)
+
+ url_info_dict = {
+ 'url': media_url,
+ 'ext': os.path.splitext(dispatch[format_id][1])[1][1:]
+ }
+
+ if format_id[-1:] == 'p':
+ url_info_dict['height'] = format_id[:-1]
+
+ urls.append(url_info_dict)
+
+ publish_time = parse_iso8601(self._html_search_regex(
+ r'发布时间 ([^<>]+) ', page, 'publish time', fatal=False),
+ delimiter=' ', timezone=datetime.timedelta(hours=8))
+
+ return {
+ 'id': media_id,
+ 'formats': urls,
+ 'title': playurl['title'],
+ 'thumbnail': playurl['pic'],
+ 'timestamp': publish_time,
+ }