commit you-get for openSUSE:Factory

Hello community, here is the log from the commit of package you-get for openSUSE:Factory checked in at 2018-06-29 22:35:43 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Comparing /work/SRC/openSUSE:Factory/you-get (Old) and /work/SRC/openSUSE:Factory/.you-get.new (New) ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Package is "you-get" Fri Jun 29 22:35:43 2018 rev:7 rq:619582 version:0.4.1099 Changes: -------- --- /work/SRC/openSUSE:Factory/you-get/you-get.changes 2018-05-16 11:44:10.770651664 +0200 +++ /work/SRC/openSUSE:Factory/.you-get.new/you-get.changes 2018-06-29 22:35:51.302030222 +0200 @@ -1,0 +2,5 @@ +Thu Jun 28 12:37:52 UTC 2018 - aloisio@gmx.com + +- Update to version 0.4.1099 (no changelog supplied) + +------------------------------------------------------------------- Old: ---- you-get-0.4.1077.tar.gz New: ---- you-get-0.4.1099.tar.gz ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Other differences: ------------------ ++++++ you-get.spec ++++++ --- /var/tmp/diff_new_pack.hwgHgx/_old 2018-06-29 22:35:51.826029767 +0200 +++ /var/tmp/diff_new_pack.hwgHgx/_new 2018-06-29 22:35:51.830029764 +0200 @@ -17,7 +17,7 @@ Name: you-get -Version: 0.4.1077 +Version: 0.4.1099 Release: 0 Summary: Dumb downloader that scrapes the web License: MIT ++++++ you-get-0.4.1077.tar.gz -> you-get-0.4.1099.tar.gz ++++++ diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/.travis.yml new/you-get-0.4.1099/.travis.yml --- old/you-get-0.4.1077/.travis.yml 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/.travis.yml 2018-06-27 23:35:15.000000000 +0200 @@ -8,6 +8,9 @@ - "3.6" - "nightly" - "pypy3" +before_install: pip install flake8 +before_script: + - if [[ $TRAVIS_PYTHON_VERSION != '3.2'* ]]; then flake8 . --count --select=E901,E999,F821,F822,F823 --show-source --statistics; fi script: make test sudo: false notifications: diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/README.md new/you-get-0.4.1099/README.md --- old/you-get-0.4.1077/README.md 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/README.md 2018-06-27 23:35:15.000000000 +0200 @@ -416,6 +416,7 @@ | 西瓜视频 | <https://www.ixigua.com/> |✓| | | | 快手 | <https://www.kuaishou.com/> |✓|✓| | | 抖音 | <https://www.douyin.com/> |✓| | | +| 中国体育(TV) | <http://v.zhibo.tv/> </br><http://video.zhibo.tv/> |✓| | | For all other sites not on the list, the universal extractor will take care of finding and downloading interesting resources from the page. diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/common.py new/you-get-0.4.1099/src/you_get/common.py --- old/you-get-0.4.1077/src/you_get/common.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/common.py 2018-06-27 23:35:15.000000000 +0200 @@ -24,6 +24,7 @@ SITES = { '163' : 'netease', '56' : 'w56', + '365yg' : 'toutiao', 'acfun' : 'acfun', 'archive' : 'archive', 'baidu' : 'baidu', @@ -64,6 +65,7 @@ 'iqiyi' : 'iqiyi', 'ixigua' : 'ixigua', 'isuntv' : 'suntv', + 'iwara' : 'iwara', 'joy' : 'joy', 'kankanews' : 'bilibili', 'khanacademy' : 'khan', @@ -82,6 +84,7 @@ 'mixcloud' : 'mixcloud', 'mtv81' : 'mtv81', 'musicplayon' : 'musicplayon', + 'miaopai' : 'yixia', 'naver' : 'naver', '7gogo' : 'nanagogo', 'nicovideo' : 'nicovideo', @@ -118,14 +121,12 @@ 'xiaojiadianvideo' : 'fc2video', 'ximalaya' : 'ximalaya', 'yinyuetai' : 'yinyuetai', - 'miaopai' : 'yixia', 'yizhibo' : 'yizhibo', 'youku' : 'youku', - 'iwara' : 'iwara', 'youtu' : 'youtube', 'youtube' : 'youtube', 'zhanqi' : 'zhanqi', - '365yg' : 'toutiao', + 'zhibo' : 'zhibo', } dry_run = False diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/__init__.py new/you-get-0.4.1099/src/you_get/extractors/__init__.py --- old/you-get-0.4.1077/src/you_get/extractors/__init__.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/extractors/__init__.py 2018-06-27 23:35:15.000000000 +0200 @@ -88,3 +88,4 @@ from .khan import * from .zhanqi import * from .kuaishou import * +from .zhibo import * \ No newline at end of file diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/acfun.py new/you-get-0.4.1099/src/you_get/extractors/acfun.py --- old/you-get-0.4.1077/src/you_get/extractors/acfun.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/extractors/acfun.py 2018-06-27 23:35:15.000000000 +0200 @@ -65,7 +65,7 @@ elif sourceType == 'tudou': tudou_download_by_iid(sourceId, title, output_dir=output_dir, merge=merge, info_only=info_only) elif sourceType == 'qq': - qq_download_by_vid(sourceId, title, output_dir=output_dir, merge=merge, info_only=info_only) + qq_download_by_vid(sourceId, title, True, output_dir=output_dir, merge=merge, info_only=info_only) elif sourceType == 'letv': letvcloud_download_by_vu(sourceId, '2d8c027396', title, output_dir=output_dir, merge=merge, info_only=info_only) elif sourceType == 'zhuzhan': diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/baidu.py new/you-get-0.4.1099/src/you_get/extractors/baidu.py --- old/you-get-0.4.1077/src/you_get/extractors/baidu.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/extractors/baidu.py 2018-06-27 23:35:15.000000000 +0200 @@ -129,6 +129,15 @@ html = get_html(url) title = r1(r'title:"([^"]+)"', html) + vhsrc = re.findall(r'vhsrc="([^"]+)"', html) + if vhsrc is not None: + ext = 'mp4' + size = url_size(vhsrc[0]) + print_info(site_info, title, ext, size) + if not info_only: + download_urls(vhsrc, title, ext, size, + output_dir=output_dir, merge=False) + items = re.findall( r'//imgsrc.baidu.com/forum/w[^"]+/([^/"]+)', html) urls = ['http://imgsrc.baidu.com/forum/pic/item/' + i diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/bilibili.py new/you-get-0.4.1099/src/you_get/extractors/bilibili.py --- old/you-get-0.4.1077/src/you_get/extractors/bilibili.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/extractors/bilibili.py 2018-06-27 23:35:15.000000000 +0200 @@ -169,15 +169,17 @@ tc_flashvars = tc_flashvars.group(1) if tc_flashvars is not None: self.out = True - qq_download_by_vid(tc_flashvars, self.title, output_dir=kwargs['output_dir'], merge=kwargs['merge'], info_only=kwargs['info_only']) + qq_download_by_vid(tc_flashvars, self.title, True, output_dir=kwargs['output_dir'], merge=kwargs['merge'], info_only=kwargs['info_only']) return has_plist = re.search(r'"page":2', self.page) - if has_plist: + if has_plist and not kwargs.get('playlist'): log.w('This page contains a playlist. (use --playlist to download all videos.)') try: - cid = re.search(r'cid=(\d+)', self.page).group(1) + page_list = json.loads(re.search(r'"pages":(\[.*?\])', self.page).group(1)) + index_id = int(re.search(r'index_(\d+)', self.url).group(1)) + cid = page_list[index_id-1]['cid'] # change cid match rule except: cid = re.search(r'"cid":(\d+)', self.page).group(1) if cid is not None: @@ -339,6 +341,7 @@ def bilibili_download_playlist_by_url(url, **kwargs): url = url_locations([url])[0] + kwargs['playlist'] = True # a bangumi here? possible? if 'live.bilibili' in url: site.download_by_url(url) @@ -357,7 +360,7 @@ page_cnt = len(page_list) for no in range(1, page_cnt+1): page_url = 'http://www.bilibili.com/video/av{}/index_{}.html'.format(aid, no) - subtitle = page_list[no-1]['pagename'] + subtitle = '#%s. %s'% (page_list[no-1]['page'], page_list[no-1]['pagename']) Bilibili().download_by_url(page_url, subtitle=subtitle, **kwargs) site = Bilibili() diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/douyutv.py new/you-get-0.4.1099/src/you_get/extractors/douyutv.py --- old/you-get-0.4.1077/src/you_get/extractors/douyutv.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/extractors/douyutv.py 2018-06-27 23:35:15.000000000 +0200 @@ -45,9 +45,9 @@ douyutv_video_download(url, output_dir=output_dir, merge=merge, info_only=info_only, **kwargs) return - url = re.sub(r'[w.]*douyu.com', 'm.douyu.com', url) + url = re.sub(r'.*douyu.com','https://m.douyu.com/room', url) html = get_content(url, headers) - room_id_patt = r'room_id\s*:\s*(\d+),' + room_id_patt = r'"rid"\s*:\s*(\d+),' room_id = match1(html, room_id_patt) if room_id == "0": room_id = url[url.rfind('/') + 1:] diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/instagram.py new/you-get-0.4.1099/src/you_get/extractors/instagram.py --- old/you-get-0.4.1077/src/you_get/extractors/instagram.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/extractors/instagram.py 2018-06-27 23:35:15.000000000 +0200 @@ -27,6 +27,8 @@ for edge in edges: title = edge['node']['shortcode'] image_url = edge['node']['display_url'] + if 'video_url' in edge['node']: + image_url = edge['node']['video_url'] ext = image_url.split('.')[-1] size = int(get_head(image_url)['Content-Length']) print_info(site_info, title, ext, size) @@ -39,6 +41,8 @@ else: title = info['entry_data']['PostPage'][0]['graphql']['shortcode_media']['shortcode'] image_url = info['entry_data']['PostPage'][0]['graphql']['shortcode_media']['display_url'] + if 'video_url' in info['entry_data']['PostPage'][0]['graphql']['shortcode_media']: + image_url =info['entry_data']['PostPage'][0]['graphql']['shortcode_media']['video_url'] ext = image_url.split('.')[-1] size = int(get_head(image_url)['Content-Length']) print_info(site_info, title, ext, size) diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/ixigua.py new/you-get-0.4.1099/src/you_get/extractors/ixigua.py --- old/you-get-0.4.1077/src/you_get/extractors/ixigua.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/extractors/ixigua.py 2018-06-27 23:35:15.000000000 +0200 @@ -1,5 +1,6 @@ #!/usr/bin/env python -__all__ = ['ixigua_download', 'ixigua_download_playlist'] +__all__ = ['ixigua_download'] + from .toutiao import download as toutiao_download from .toutiao import download_playlist as toutiao_download_playlist @@ -10,4 +11,4 @@ site_info = "ixigua.com" download = ixigua_download -download_playlist = toutiao_download_playlist \ No newline at end of file +download_playlist = toutiao_download_playlist diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/qq.py new/you-get-0.4.1099/src/you_get/extractors/qq.py --- old/you-get-0.4.1077/src/you_get/extractors/qq.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/extractors/qq.py 2018-06-27 23:35:15.000000000 +0200 @@ -8,8 +8,15 @@ from .qie_video import download_by_url as qie_video_download from urllib.parse import urlparse,parse_qs -def qq_download_by_vid(vid, title, output_dir='.', merge=True, info_only=False): - info_api = 'http://vv.video.qq.com/getinfo?otype=json&appver=3.2.19.333&platform=11&defnpayver=1&vid={}'.format(vid) +def qq_download_by_vid(vid, title, default_from, output_dir='.', merge=True, info_only=False): + + if default_from: + platform = 11 + else: + # fix return {,"msg":"cannot play outside"} + platform = 4100201 + + info_api = 'http://vv.video.qq.com/getinfo?otype=json&appver=3.2.19.333&platform=11&defnpayver=1&defn=shd&vid={}'.format(vid) info = get_content(info_api) video_json = json.loads(match1(info, r'QZOutputJson=(.*)')[:-1]) @@ -17,7 +24,8 @@ title = video_json['vl']['vi'][0]['ti'] host = video_json['vl']['vi'][0]['ul']['ui'][0]['url'] streams = video_json['fl']['fi'] - seg_cnt = video_json['vl']['vi'][0]['cl']['fc'] + seg_cnt = fc_cnt = video_json['vl']['vi'][0]['cl']['fc'] + filename = video_json['vl']['vi'][0]['fn'] if seg_cnt == 0: seg_cnt = 1 @@ -25,21 +33,13 @@ fn_pre, magic_str, video_type = filename.split('.') best_quality = streams[-1]['name'] - #part_format_id = streams[-1]['id'] part_urls= [] total_size = 0 for part in range(1, seg_cnt+1): - #if seg_cnt == 1 and video_json['vl']['vi'][0]['vh'] <= 480: - # filename = fn_pre + '.mp4' - #else: - # filename = fn_pre + '.p' + str(part_format_id % 10000) + '.' + str(part) + '.mp4' - #filename = fn_pre + '.p' + str(part_format_id % 10000) + '.' + str(part) + '.mp4' - - # fix some error cases("check vid&filename failed" and "format invalid") - # https://v.qq.com/x/page/q06058th9ll.html - # https://v.qq.com/x/page/t060789a21e.html - if seg_cnt == 1: + if fc_cnt == 0: + # fix json parsing error + # example:https://v.qq.com/x/page/w0674l9yrrh.html part_format_id = video_json['vl']['vi'][0]['cl']['keyid'].split('.')[-1] else: part_format_id = video_json['vl']['vi'][0]['cl']['ci'][part - 1]['keyid'].split('.')[1] @@ -112,6 +112,8 @@ def qq_download(url, output_dir='.', merge=True, info_only=False, **kwargs): """""" + default_from = True + if re.match(r'https?://egame.qq.com/live\?anchorid=(\d+)', url): from . import qq_egame qq_egame.qq_egame_download(url, output_dir=output_dir, merge=merge, info_only=info_only, **kwargs) @@ -134,7 +136,7 @@ content = get_content(url) vids = matchall(content, [r'\?vid=(\w+)']) for vid in vids: - qq_download_by_vid(vid, vid, output_dir, merge, info_only) + qq_download_by_vid(vid, vid, default_from, output_dir, merge, info_only) return if 'kuaibao.qq.com' in url or re.match(r'http://daxue.qq.com/content/content/id/\d+', url): @@ -165,7 +167,12 @@ title = match1(content, r'"title":"([^"]+)"') if not title else title title = vid if not title else title #general fallback - qq_download_by_vid(vid, title, output_dir, merge, info_only) + if 'v.sports.qq.com' in url: + # fix url forbidden + # example:http://v.sports.qq.com/#/cover/t0fqsm1y83r8v5j/a0026nvw5jr + default_from = False + + qq_download_by_vid(vid, title, default_from, output_dir, merge, info_only) site_info = "QQ.com" download = qq_download diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/twitter.py new/you-get-0.4.1099/src/you_get/extractors/twitter.py --- old/you-get-0.4.1077/src/you_get/extractors/twitter.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/extractors/twitter.py 2018-06-27 23:35:15.000000000 +0200 @@ -19,7 +19,7 @@ url = 'https://' + match1(url, r'//mobile\.(.+)') if re.match(r'https?://twitter\.com/i/moments/', url): # moments - html = get_html(url) + html = get_html(url, faker=True) paths = re.findall(r'data-permalink-path="([^"]+)"', html) for path in paths: twitter_download('https://twitter.com' + path, @@ -29,7 +29,7 @@ **kwargs) return - html = get_html(url) + html = get_html(url, faker=True) screen_name = r1(r'data-screen-name="([^"]*)"', html) or \ r1(r'<meta name="twitter:title" content="([^"]*)"', html) item_id = r1(r'data-item-id="([^"]*)"', html) or \ diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/universal.py new/you-get-0.4.1099/src/you_get/extractors/universal.py --- old/you-get-0.4.1077/src/you_get/extractors/universal.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/extractors/universal.py 2018-06-27 23:35:15.000000000 +0200 @@ -67,12 +67,12 @@ urls = [] for i in media_exts: - urls += re.findall(r'(https?://[^;"\'\\]+' + i + r'[^;"\'\\]*)', page) + urls += re.findall(r'(https?://[^ ;"\'\\]+' + i + r'[^ ;"\'\\]*)', page) p_urls = re.findall(r'(https?%3A%2F%2F[^;&]+' + i + r'[^;&]*)', page) urls += [parse.unquote(url) for url in p_urls] - q_urls = re.findall(r'(https?:\\\\/\\\\/[^;"\']+' + i + r'[^;"\']*)', page) + q_urls = re.findall(r'(https?:\\\\/\\\\/[^ ;"\']+' + i + r'[^ ;"\']*)', page) urls += [url.replace('\\\\/', '/') for url in q_urls] # a link href to an image is often an interesting one diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/extractors/zhibo.py new/you-get-0.4.1099/src/you_get/extractors/zhibo.py --- old/you-get-0.4.1077/src/you_get/extractors/zhibo.py 1970-01-01 01:00:00.000000000 +0100 +++ new/you-get-0.4.1099/src/you_get/extractors/zhibo.py 2018-06-27 23:35:15.000000000 +0200 @@ -0,0 +1,52 @@ +#!/usr/bin/env python + +__all__ = ['zhibo_download'] + +from ..common import * + +def zhibo_vedio_download(url, output_dir = '.', merge = True, info_only = False, **kwargs): + # http://video.zhibo.tv/video/details/d103057f-663e-11e8-9d83-525400ccac43.htm... + + html = get_html(url) + title = r1(r'<title>([\s\S]*)</title>', html) + total_size = 0 + part_urls= [] + + video_html = r1(r'<script type="text/javascript">([\s\S]*)</script></head>', html) + + # video_guessulike = r1(r"window.xgData =([s\S'\s\.]*)\'\;[\s\S]*window.vouchData", video_html) + video_url = r1(r"window.vurl = \'([s\S'\s\.]*)\'\;[\s\S]*window.imgurl", video_html) + part_urls.append(video_url) + ext = video_url.split('.')[-1] + + print_info(site_info, title, ext, total_size) + if not info_only: + download_urls(part_urls, title, ext, total_size, output_dir=output_dir, merge=merge) + + +def zhibo_download(url, output_dir = '.', merge = True, info_only = False, **kwargs): + if 'video.zhibo.tv' in url: + zhibo_vedio_download(url, output_dir=output_dir, merge=merge, info_only=info_only, **kwargs) + return + + # if 'v.zhibo.tv' in url: + # http://v.zhibo.tv/31609372 + html = get_html(url) + title = r1(r'<title>([\s\S]*)</title>', html) + is_live = r1(r"window.videoIsLive=\'([s\S'\s\.]*)\'\;[\s\S]*window.resDomain", html) + if is_live is not "1": + raise ValueError("The live stream is not online! (Errno:%s)" % is_live) + + ourStreamName = r1(r"window.ourStreamName=\'([s\S'\s\.]*)\'\;[\s\S]*window.rtmpDefaultSource", html) + rtmpPollUrl = r1(r"window.rtmpPollUrl=\'([s\S'\s\.]*)\'\;[\s\S]*window.hlsDefaultSource", html) + + #real_url = 'rtmp://220.194.213.56/live.zhibo.tv/8live/' + ourStreamName + real_url = rtmpPollUrl + ourStreamName + + print_info(site_info, title, 'flv', float('inf')) + if not info_only: + download_url_ffmpeg(real_url, title, 'flv', params={}, output_dir=output_dir, merge=merge) + +site_info = "zhibo.tv" +download = zhibo_download +download_playlist = playlist_not_supported('zhibo') diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/processor/ffmpeg.py new/you-get-0.4.1099/src/you_get/processor/ffmpeg.py --- old/you-get-0.4.1077/src/you_get/processor/ffmpeg.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/processor/ffmpeg.py 2018-06-27 23:35:15.000000000 +0200 @@ -3,6 +3,7 @@ import logging import os.path import subprocess +import sys from ..util.strings import parameterize from ..common import print_more_compatible as print @@ -25,8 +26,8 @@ try: version = [int(i) for i in vers[2].split('.')] except: - print('It seems that your ffmpeg is a nightly build.') - print('Please switch to the latest stable if merging failed.') + print('It seems that your ffmpeg is a nightly build.', file=sys.stderr) + print('Please switch to the latest stable if merging failed.', file=sys.stderr) version = [1, 0] return cmd, 'ffprobe', version except: diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/you-get-0.4.1077/src/you_get/version.py new/you-get-0.4.1099/src/you_get/version.py --- old/you-get-0.4.1077/src/you_get/version.py 2018-05-15 21:22:51.000000000 +0200 +++ new/you-get-0.4.1099/src/you_get/version.py 2018-06-27 23:35:15.000000000 +0200 @@ -1,4 +1,4 @@ #!/usr/bin/env python script_name = 'you-get' -__version__ = '0.4.1077' +__version__ = '0.4.1099'
participants (1)
-
root