diff --git a/app.py b/app.py index 7571dfc..757817e 100644 --- a/app.py +++ b/app.py @@ -1,7 +1,7 @@ import argparse import sys from flask import Flask -from flask.ext.restful import reqparse, Api, Resource +from flask_restful import reqparse, Api, Resource from decoder import Decoder from languages import new_lang_from_long_english_name from text import PreProcessor, PostProcessor @@ -12,7 +12,7 @@ api = Api(app) http_parser = reqparse.RequestParser() -http_parser.add_argument('inputText', type=unicode, location='json') +http_parser.add_argument('inputText', type=str, location='json') http_parser.add_argument('inputLanguage', type=str, location='json') decoders = {} diff --git a/decoder.py b/decoder.py index 3a9cfe5..5136917 100644 --- a/decoder.py +++ b/decoder.py @@ -35,17 +35,17 @@ def translate(self, input_text, sock=None): sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock.connect(('localhost', self.port)) - tx_msg = input_text + u'\n' + tx_msg = input_text + '\n' sock.send(tx_msg.encode('utf8')) num_lines = len(input_text.split('\n')) - rx_msg = u'' + rx_msg = '' for i in range(num_lines): rx_msg += sock.recv(1024).decode('utf8') return rx_msg def start_decoder_server(self): - runner_path = os.path.join(self.bundle_dir, 'run-joshua.sh') + runner_path = os.path.join(self.bundle_dir, 'joshua') options = ['-server-port', str(self.port)] subprocess.Popen([runner_path] + options, env=os.environ) diff --git a/text.py b/text.py index 039910c..1631374 100644 --- a/text.py +++ b/text.py @@ -1,3 +1,4 @@ +import base64 import os from subprocess import Popen, PIPE import sys @@ -11,7 +12,6 @@ def _tokenize(lang_short_code, text): runner_path = os.path.join( os.environ['JOSHUA'], 'scripts', - 'preparation', 'tokenize.pl' ) options = ['-l', lang_short_code] @@ -23,15 +23,14 @@ def _tokenize(lang_short_code, text): env=os.environ ) out, err = p.communicate(text.encode('utf8')) - sys.stderr.write(err.encode('utf8') + '\n') - return unicode(out.strip(), encoding='utf8').split('\n') + sys.stderr.write(err.decode('utf8') + '\n') + return str(out.strip(), encoding='utf8').split('\n') def _detokenize(lang_short_code, text): runner_path = os.path.join( os.environ['JOSHUA'], 'scripts', - 'preparation', 'detokenize.pl' ) options = ['-l', lang_short_code] @@ -43,8 +42,8 @@ def _detokenize(lang_short_code, text): env=os.environ ) out, err = p.communicate(text.encode('utf8')) - sys.stderr.write(err.encode('utf8') + '\n') - return unicode(out.strip(), encoding='utf8') + sys.stderr.write(err.decode('utf8') + '\n') + return str(out.strip(), encoding='utf8') def tokenize(lang_short_code, sentences): @@ -114,14 +113,14 @@ def merge_lines(translation): next_line = lines.pop(0) if prev_line == '': if next_line == '': - result += u'\n\n' + result += '\n\n' else: result += next_line else: if next_line == '': - result += u'\n\n' + result += '\n\n' else: - result = u'{0} {1}'.format(result, next_line) + result = '{0} {1}'.format(result, next_line) prev_line = next_line