the applicability of our approach for other languages. This should allow us to evaluate
the relative efficiency of the system in language dependency. There are cases where
nouns cannot be extracted by this approach. However, we believe that our approach
can be used to extract present component-integral part-whole relation from a well
balanced-corpus to help building resources like WordNet or an ontology in a given
language. For applications requiring sentence-based semantics, the approach needs to
be modified in order to cover sentences containing nouns with a part-whole relation,
but whose head and tail are separated by one or several nouns.
References
1. Beamer, B., Rozovskaya, A. and Girju, A., (2008). Automatic Semantic Relation Extraction
with Multiple Boundary Generation. Association for the Advancement of Artificial
Intelligence.
2. Cederberg, S. and D. Widdows (2003). Using LSA and Noun Coordination Information to
Improve the Precision and Recall of Automatic Hyponymy Extraction. In Conference on
Natural Language Learning (CoNLL-2003), Edmonton, Canada, pp. 111-118.
3. Firth, J. R., (1957). A synopsis of linguistic theory 1930-1955. In Studies in Linguistic
Analysis, pp. 1-32. Oxford: Philological Society.
4. Girju R., Moldovan D., Tatu, M. and Antohe, D., (2005). Automatic Discovery of Part–
Whole Relations. ACM 32(1)
5. Girju, R., Hearst, M., Nakov, P., Nastase, V., Szpakowicz; S., Turney, P. and Yuret D.,
(2007). Classification of Semantic Relations between Nominals: Dataset for Task 4 in
SemEval, 4th International Workshop on Semantic Evaluations, Prague, Czech Republic.
6. Grad, B., (1995). Extraction of semantic relations from bioscience text. University of
Trieste, Italy.
7. Hage, W., Kolb, H. and Schreiber, G., (2006). A Method for Learning Part-Whole
Relations. TNO Science & Industry Delft, Vrije Universiteit Amsterdam.
8. Harris, Z., (1954). Distributional structure. Word 10 (23), 46–162.
9. Harshman, R., (1970). Foundations of the parafac procedure: Models and conditions for an
“explanatory” multi-modal factor analysis. UCLA Working Papers in Phonetics, 16.
10. Landauer, T., Foltz P., & Laham, D., (1998). Introduction to Latent Semantic Analysis.
Discourse processes, 25, 259-284.
11. Lund, K. and Burgess, C., (1996). Producing high-dimensional semantic spaces from
lexical co-occurrence. Behavior Research Methods, Instruments, and Computers, 28(2),
203–208.
12. Mark, D., (2011). N-grams and word frequency data from the Corpus of Historical
American English (COHA). http://www.ngrams.info.
13. Matthew, B. and Charniak, E., (1999). Finding parts in very large corpora. In Proceedings
of the 37th Annual Meeting of the Association for Computational Linguistics (ACL1999),
pages 57–64, University of Maryland.
14. Salton, G., Wong, A. and Yang, C.-S., (1975). A vector space model for automatic
indexing. Communications of the ACM, 18(11), 613–620.
15. Winston, M., Chaffin, R. and Hermann, D., (1987). Taxonomy of part-whole relations.
Cognitive Science, 11(4), 417–444.
16. Wittgenstein, L., (1922). Tractatus Logico-Philosophicus. London: Routledge & Kegan
Paul.
17. Zock, M.: Needles in a haystack and how to find them? The case of lexical access. In E.
Miyares Bermudez and L. Ruiz Miyares (Eds.) "Linguistics in the Twenty First Century".
Cambridge Scholars Press, 2006, United Kingdom, pp. 155-162.
139