forked from eladhoffer/seq2seq.pytorch
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathapp.py
More file actions
82 lines (70 loc) · 3.36 KB
/
app.py
File metadata and controls
82 lines (70 loc) · 3.36 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
from flask import Flask, render_template, request
import torch
from torch.utils.model_zoo import load_url
from seq2seq.models import Img2Seq
from seq2seq.tools.inference import CaptionGenerator
import numpy as np
from seq2seq.datasets.pix2codedataset import Pix2CodeDataset
import logging
import re
import os
import cv2
app = Flask(__name__)
checkpoint = load_url(
'model_best.pth.tar', model_dir="results/pix2code_devsupport_resnet50_finetune/", map_location={'gpu:0': 'cpu'})
model = Img2Seq(**checkpoint['config'].model_config)
model.load_state_dict(checkpoint['state_dict'])
img_transform, target_tok = checkpoint['tokenizers'].values()
beam_size = 3
caption_model = CaptionGenerator(model,
img_transform=img_transform,
target_tok=target_tok,
beam_size=beam_size,
get_attention=True,
max_sequence_length=250,
length_normalization_factor=5.0,
cuda=True,
length_normalization_const=5)
dset = Pix2CodeDataset("/home/vigi99/devsupportai_ui_gen/eval_set")
def sample(dset, caption_model, index=None):
sample_index = np.random.choice(len(dset)) if index is None else index
sample_img_filename = dset.file_names[sample_index] + ".png"
sample_img, sample_target = dset.get_sample(sample_index)
predicted_target, attentions = caption_model.describe(sample_img)
return (sample_img, sample_img_filename), sample_target, predicted_target.decode('utf-8').split(' '), attentions
def return_list_with_tabs(li):
elems = map(lambda x: x.replace("\\n", "\n").replace("\\t", "\t"), li)
return ' '.join(elems)
def get_np_array_from_file_object(file):
'''converts a buffer from a tar file in np.array'''
return np.asarray(bytearray(file.read()), dtype=np.uint8)
@app.before_first_request
def setup_logging():
if not app.debug:
# In production mode, add log handler to sys.stderr.
formatter = logging.Formatter(
"[%(asctime)s] {%(pathname)s:%(lineno)d} %(levelname)s - %(message)s")
handler = logging.StreamHandler()
handler.setFormatter(formatter)
app.logger.addHandler(handler)
app.logger.setLevel(logging.INFO)
@app.route("/")
def random_example():
(img, img_fname), target, predicted, attentions = sample(dset, caption_model)
img_fname = re.sub(r'^.*eval_set','/static', img_fname)
actual_text = return_list_with_tabs(target)
predicted_text = return_list_with_tabs(predicted)
app.logger.info('ImageFile: %s, Actual Text: %s, Predicted Text: %s ', img_fname, target, predicted)
return render_template('index.html', img_filepath=img_fname, actual_text=actual_text, predicted_text=predicted_text)
@app.route("/upload", methods=['GET', 'POST'])
def upload_example():
predicted_target = ''
if request.method == 'POST':
file = request.files['uploaded_image']
img = cv2.imdecode(get_np_array_from_file_object(file), 1)
img = cv2.resize(img, (256,256))
predicted_target_list, attentions = caption_model.describe(img)
predicted_target = return_list_with_tabs(predicted_target_list.decode('utf-8').split(' '))
return render_template('upload.html', text=predicted_target)
if __name__ == "__main__":
app.run()