mirror of
https://source.perfectable.org/nivertius/napi.git
synced 2025-07-02 03:14:46 +02:00
Convert to python3
This commit is contained in:
parent
dd857db598
commit
aeddbd3e22
1 changed files with 12 additions and 10 deletions
22
napi
22
napi
|
@ -1,4 +1,4 @@
|
||||||
#!/usr/bin/python
|
#!/usr/bin/python3
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# reversed napi 0.16.3.1
|
# reversed napi 0.16.3.1
|
||||||
# by gim,krzynio,dosiu,hash 2oo8.
|
# by gim,krzynio,dosiu,hash 2oo8.
|
||||||
|
@ -7,6 +7,8 @@
|
||||||
import hashlib
|
import hashlib
|
||||||
import sys
|
import sys
|
||||||
import urllib
|
import urllib
|
||||||
|
from urllib.parse import urlencode
|
||||||
|
from urllib.request import urlopen
|
||||||
import os
|
import os
|
||||||
import argparse
|
import argparse
|
||||||
import base64
|
import base64
|
||||||
|
@ -31,10 +33,10 @@ for movie_filename in args.movies:
|
||||||
|
|
||||||
digest = hashlib.md5()
|
digest = hashlib.md5()
|
||||||
try:
|
try:
|
||||||
with open(movie_filename, "r") as movie_file:
|
with open(movie_filename, "rb") as movie_file:
|
||||||
digest.update(movie_file.read(10 * 1024 * 1024))
|
digest.update(movie_file.read(10 * 1024 * 1024))
|
||||||
except IOError:
|
except IOError:
|
||||||
print "%s: No such file" % movie_filename
|
print("%s: No such file" % movie_filename)
|
||||||
continue
|
continue
|
||||||
|
|
||||||
movie_digest = digest.hexdigest()
|
movie_digest = digest.hexdigest()
|
||||||
|
@ -46,14 +48,14 @@ for movie_filename in args.movies:
|
||||||
"downloaded_subtitles_lang" : args.language
|
"downloaded_subtitles_lang" : args.language
|
||||||
}
|
}
|
||||||
|
|
||||||
request_data = urllib.urlencode(request_data)
|
request_data = urlencode(request_data).encode()
|
||||||
response = urllib.urlopen(base_url, request_data)
|
response = urlopen(base_url, request_data)
|
||||||
response_data = response.read()
|
response_data = response.read()
|
||||||
response_xml = XML(response_data)
|
response_xml = XML(response_data)
|
||||||
|
|
||||||
response_status = response_xml.find('status')
|
response_status = response_xml.find('status')
|
||||||
if response_status is None or response_status.text != 'success':
|
if response_status is None or response_status.text != 'success':
|
||||||
print "%s: No subtitles found for language %s" % (movie_filename, args.language)
|
print("%s: No subtitles found for language %s" % (movie_filename, args.language))
|
||||||
continue
|
continue
|
||||||
|
|
||||||
subtitle_xml = response_xml.find("subtitles")
|
subtitle_xml = response_xml.find("subtitles")
|
||||||
|
@ -64,7 +66,7 @@ for movie_filename in args.movies:
|
||||||
|
|
||||||
content = subtitle_xml.find("content").text
|
content = subtitle_xml.find("content").text
|
||||||
decoded_content = base64.b64decode(content)
|
decoded_content = base64.b64decode(content)
|
||||||
with open(input_filename, "w") as input_file:
|
with open(input_filename, "wb") as input_file:
|
||||||
input_file.write(decoded_content)
|
input_file.write(decoded_content)
|
||||||
subtitle_filename = movie_filename[:-4] + ".srt"
|
subtitle_filename = movie_filename[:-4] + ".srt"
|
||||||
|
|
||||||
|
@ -83,10 +85,10 @@ for movie_filename in args.movies:
|
||||||
result = process.wait()
|
result = process.wait()
|
||||||
|
|
||||||
if result != 0:
|
if result != 0:
|
||||||
print "%s: 7zip returned non-zero code: %s" % (movie_filename, subtitle_filename)
|
print("%s: 7zip returned non-zero code: %s" % (movie_filename, subtitle_filename))
|
||||||
os.remove(subtitle_filename)
|
os.remove(subtitle_filename)
|
||||||
continue
|
continue
|
||||||
|
|
||||||
print "%s: downloaded subtitles: language '%s', uploader '%s', author '%s', date '%s'" % (movie_filename,
|
print("%s: downloaded subtitles: language '%s', uploader '%s', author '%s', date '%s'" % (movie_filename,
|
||||||
args.language, subtitle_uploader, subtitle_author, subtitle_date)
|
args.language, subtitle_uploader, subtitle_author, subtitle_date))
|
||||||
os.remove(input_filename)
|
os.remove(input_filename)
|
||||||
|
|
Loading…
Add table
Reference in a new issue