standard significance level α = 0.05 in every case.
In the case of the Sigmoid activation function, the
experimental results do not provide any evidences
that the best-ranked appro ach is different from the
second one since the adjusted p-value is greater
than the standard significance level; however, there
is enough evidence to prove that it is different from
the remaining ones since the adjusted p-value is
smaller th an the significance level. So, we selected
the GRU-BiGRU alternative in both ca ses.
Next, we compare d our best alternatives to the
baselines. The results of the comparison a re shown
in Table 2.c, where the subindex d enotes the activa-
tion function used. According to the statistical test,
there is not enough evidence in the experimental
data to make a difference between our proposals,
but there is enough evidence to prove that they
both are significantly better than the baseline s.
5 CONCLUSIONS
We have motivated the need for mining conditi-
ons and we have presented a novel approach to the
problem. It relies on a encoder-decoder model as
a means to overcome the drawbacks that we have
found in o ther proposals, namely: it does not rely
on user-defined patterns, it does not require any
specific-purpose dictionaries, taxo nomies, or heu-
ristics, and it ca n mine conditions in both factual
and opinion sen tences. Furthermor e it only needs
two components that are read ily available, namely:
a stemmer and a word embedder. We have also per-
formed a comprehensive experimental analysis on a
large multi-topic dataset with 3 779 000 sentences in
English and Spanish that we make publicly availa-
ble. Our results confirm that our proposal is similar
to the state-of-the -art proposals in terms of preci-
sion, but it improves recall enough to beat them
in terms of the F
1
score. We have backed up the
previous conclusions using sound statistical tests.
ACKNOWLEDGEMENTS
The work described in this paper was supported
by Opileak.es and the Spa nish R&D programme
(grants TIN2013-40848-R and TIN2013-40848-R).
The computing facilities were provided by the An-
dalusian Scientific Computing Cen tre (CICA).
REFERENCES
Bengio, Y., Simard, P. Y., and Frasconi, P. ( 1994). Le-
arning long-term dependencies with gradient descent
is difficult. IEEE Trans. Neural Networks, 5(2):157–
166.
Caruana, R., Lawrence, S., and Giles, C. L. (2000). Over-
fitting in neural nets: Backpropagation, conjugate
gradient, and early stopping. In NIPS, pages 402–
408.
Chen, H., Huang, S., Chiang, D., and Chen, J. (2017).
Improved neural machine translation with a syntax-
aware encoder and decoder. In ACL, pages 1936–
1945.
Chikersal, P., Poria, S., Cambria, E., Gelbukh, A . F., and
Siong, C. E . (2015). Modelling public sentiment in
Twitter. In CICLing (2), pages 49–65.
Cho, K., van Merrienboer, B., Bahdanau, D. , and Ben-
gio, Y. (2014). On the properties of neural ma-
chine translation: Encoder-decoder approaches. In
EMNLP, pages 103–111.
Chopra, S ., Auli, M., and Rush, A. M. (2016). Ab-
stractive sentence summarization with attentive re-
current neural networks. In NAACL, pages 93–98.
Etzioni, O., Fader, A., Christensen, J. , Soderland, S., and
Mausam (2011). Open information extraction: the
second generation. In IJCAI, pages 3–10.
Han, H. , Zhang, S., and Qiao, J. (2017). An adaptive gro-
wing and pruning algorithm for designing recurrent
neural network. Neurocomputing, 242:51–62.
Hochreiter, S. and Schmidhuber, J. (1997). Long short-
term memory. Neural Comput., 9(8):1735–1780.
Kingma, D. P. and Ba, J. (2014). Adam: A method for
stochastic optimization. CoRR, abs/1412.6980.
Ma, X. and Hovy, E. H. (2016). End-to-end sequence la-
beling via bi-directional LSTM-CNNs-CRF. In ACL,
pages 1–11.
Mausam (2016). Open information extraction systems
and downstream applications. In IJCAI, pages 4074–
4077.
Mausam, Schmitz, M., Soderland, S., Bart, R., and Etzi-
oni, O. (2012). Open language learning for informa-
tion extraction. In EMNLP-CoNLL, pages 523–534.
Mitchell, T. M., Cohen, W. W., Hruschka, E. R. , Ta-
lukdar, P. P., Betteridge, J., Carlson, A., Mishra,
B. D., Gardner, M., Kisiel, B., Krishnamurthy, J.,
Lao, N., Mazaitis, K., Mohamed, T., Nakashole, N.,
Platanios, E. A., Ritter, A., Samadi, M., Settles, B.,
Wang, R. C., Wijaya, D. T., Gupta, A., Chen, X.,
Saparov, A., Greaves, M., and Welling, J. (2015).
Never-ending learning. In AAAI, pages 2302–2310.
Nakayama, Y. and Fujii, A. (2015). Extracti ng condition-
opinion relations toward fine-grained opinion mi-
ning. In EMNLP, pages 622–631.
Nallapati, R., Zhou, B., dos Santos, C. N., Gülçehre, Ç.,
and Xiang, B. (2016). Abstractive text summariza-
tion using sequence-to-sequence rnns and beyond. In
CoNLL, pages 280–290.