Upload inference.py
Browse filesUpdated the file to return the input tokens as well
- inference.py +1 -1
inference.py
CHANGED
@@ -33,5 +33,5 @@ def predict(x, model, max_output_len = 30):
|
|
33 |
# if xcol == config.end_token:
|
34 |
# break
|
35 |
|
36 |
-
return x[:,
|
37 |
|
|
|
33 |
# if xcol == config.end_token:
|
34 |
# break
|
35 |
|
36 |
+
return x[:, :] # B, max_output_len
|
37 |
|