to benefit an improved understanding of how to miti-
gate computational complexity in an MC engine. Fi-
nally, we plan to experiment with MSMARCO V2.1
that was just released and comprises over one million
user queries. This version allows to explore additional
evaluation metrics to overcome semantic-equivalence
limitations of ROUGE and BLEU.
ACKNOWLEDGMENTS
We would like to thank the anonymous reviewers for
their insightful suggestions and feedback.
REFERENCES
Andress, J. and Zanoci, C. (2017). Coattention-
based neural network for question answer-
ing. Technical report, Stanford University.
http://stanford.edu/class/cs224n/reports/2762015.pdf.
Chen, Z., Yang, R., Cao, B., Zhao, Z., Cai, D., and He,
X. (2017). Smarnet: Teaching machines to read
and comprehend like human. CoRR, abs/1710.02772.
http://arxiv.org/abs/1710.02772.
Cho, K., van Merrienboer, B., Bahdanau, D., and Ben-
gio, Y. (2014). On the properties of neural machine
translation: Encoder-decoder approaches. CoRR,
abs/1409.1259. http://arxiv.org/abs/1409.1259.
Chung, J., G
¨
ulc¸ehre, c., Cho, K., and Bengio, Y. (2014).
Empirical evaluation of gated recurrent neural net-
works on sequence modeling. CoRR, abs/1412.3555.
http://arxiv.org/abs/1412.3555.
Cui, Y., Chen, Z., Wei, S., Wang, S., Liu, T., and Hu,
G. (2016). Attention-over-attention neural networks
for reading comprehension. CoRR, abs/1607.04423.
http://arxiv.org/abs/1607.04423.
Dhingra, B., Liu, H., Cohen, W. W., and Salakhut-
dinov, R. (2017). Gated-attention readers for
text comprehension. CoRR, abs/1606.01549.
http://arxiv.org/abs/1606.01549.
Glorot, X. and Bengio, Y. (2010). Understanding the diffi-
culty of training deep feedforward neural networks. In
Artificial Intelligence and Statistics, pages 249–256,
Sardinia, Italy.
Higgins, B. and Nho, E. (2017). LSTM
encoder-decoder architecture with atten-
tion mechanism for machine comprehen-
sion. Technical report, Stanford University.
http://stanford.edu/class/cs224n/reports/2784756.pdf.
Hoang, C. D. V., Cohn, T., and Haffari, G. (2016). In-
corporating side information into recurrent neural net-
work language models. In Human Language Tech-
nologies: North American Chapter of the Association
for Computational Linguistics (NAACL), pages 1250–
1255, San Diego, California.
Hochreiter, S. and Schmidhuber, J. (1997). Long short-term
memory. Neural Computation, 9(8):1735–1780.
Joshi, M., Choi, E., Weld, D., and Zettlemoyer, L.
(2017). TriviaQA: A large scale distantly super-
vised challenge dataset for reading comprehension.
In Annual Meeting of the Association for Computa-
tional Linguistics (ACL), pages 1601–1611, Vancou-
ver, Canada.
Kingma, D. P. and Ba, J. (2014). Adam: A method
for stochastic optimization. CoRR, abs/1412.6980.
http://arxiv.org/abs/1412.6980.
Lee, K., Kwiatkowski, T., Parikh, A. P., and Das, D. (2016).
Learning recurrent span representations for extrac-
tive question answering. CoRR, abs/1611.01436.
http://arxiv.org/abs/1611.01436.
Lin, C. (2004). ROUGE: a package for automatic evaluation
of summaries. In Workshop on Text Summarization
Branches Out of the Association for Computational
Linguistics (ACL).
Luong, M., Pham, H., and Manning, C. D. (2015).
Effective approaches to attention-based neural
machine translation. CoRR, abs/1508.04025.
http://arxiv.org/abs/1508.04025.
Mikolov, T., Chen, K., Corrado, G. S., and Dean, J.
(2013a). Efficient estimation of word represen-
tations in vector space. CoRR, abs/1301.3781.
http://arxiv.org/abs/1301.3781.
Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., and
Dean, J. (2013b). Distributed representations of words
and phrases and their compositionality. In Advances in
Neural Information Processing Systems (NIPS), pages
3111–3119. Curran Associates, Inc., Red Hook, NY.
Nguyen, T., Rosenberg, M., Song, X., Gao, J., Ti-
wary, S., Majumder, R., and Deng, L. (2016). MS
MARCO: A human generated MAchine Reading
COmprehension dataset. CoRR, abs/1611.09268.
http://arxiv.org/abs/1611.09268.
Papineni, K., Roukos, S., Ward, T., and jing Zhu, W. (2002).
BLEU: a method for automatic evaluation of machine
translation. In Annual Meeting of the Association for
Computational Linguistics (ACL).
Pennington, J., Socher, R., and Manning, C. D. (2014).
GloVe: Global vectors for word representation. In
Empirical Methods in Natural Language Processing
(EMNLP), pages 1532–1543, Doha, Qatar.
R Core Team (2013). R: A Language and Environment
for Statistical Computing. R Foundation for Sta-
tistical Computing, Vienna, Austria. http://www.R-
project.org/.
Raiman, J. and Miller, J. (2017). Globally normal-
ized reader. In Empirical Methods in Natural
Language Processing (EMNLP), pages 1059–1069,
Copenhagen, Denmark.
Rajpurkar, P., Zhang, J., Lopyrev, K., and Liang, P. (2016).
SQuAD: 100,000+ questions for machine compre-
hension of text. In Empirical Methods in Natural
Language Processing (EMNLP), pages 2383–2392,
Austin, Texas.
Seo, M. J., Kembhavi, A., Farhadi, A., and Ha-
jishirzi, H. (2017). Bidirectional attention flow for
machine comprehension. CoRR, abs/1611.01603.
http://arxiv.org/abs/1611.01603.
KDIR 2018 - 10th International Conference on Knowledge Discovery and Information Retrieval
70