meaning and use of the schema/ontology under con-
struction. Therefore, it should not come as a surprise
that a large number of tools for ontology learning and
schema/ontology matching include some lexical re-
sources (mainly WordNet
1
) as a component, and use
it in some intermediate step to annotate schema el-
ements and ontology classes/properties with lexical
knowledge. To sum up, lexical annotation seems to
be a critical task to develop smart methods for ontol-
ogy learning and matching.
In this context, we developed MELIS (Meaning
Elicitation and Lexical Integration System), a method
and a software tool for the annotation of data sources.
The distinguishing feature and the novelty of MELIS
is its incremental annotation method: the more
sources (including a number of different schemas) are
processed, the more background/domain knowledge
is cumulated in the system, the better the performance
of the systems on new sources. MELIS supports three
important tasks: (1) the source annotation process,
i.e. the operation of associating an element of a lexi-
cal reference database (WordNet in our implementa-
tion, but the method is independent from this choice)
to all source elements, (2) the customization of the
lexical reference with the introduction of new lexi-
cal knowledge (glossa, lemma and lexical relation-
ships), and (3) the extraction of lexical/semantic re-
lationships across elements of different data sources.
Works related to the issues discussed in this paper
are in the area of languages and tools for annotations
((Bechhofer et al., 2002), (Staab et al., 2001) and
(Handschuh et al., 2003) where an approach similar
to our is adopted), techniques for extending WordNet
((Gangemi et al., 2003), (Montoyo et al., 2001) and
(Pazienza and Stellato, 2006) where a system coupled
with Prot`eg`e
2
for enriching and annotating sources
is proposed), and systems for ontology management
(see the the Ontoweb
3
and the Knowledgeweb Net-
work of Excellence
4
technical reports for complete
surveys).
2 MELIS: THE LEXICAL
KNOWLEDGE COMPONENT
In most real world applications, ontology elements
are labeled by natural language expressions. In our
opinion, the crucial reason for this aspect of ontol-
1
See
http://wordnet.princeton.edu
for more in-
formation on WordNet.
2
http://protege.stanford.edu/
3
http://www.ontoweb.org, in particular deliverable 1.4
4
http://knowledgeweb.semanticweb.org/
ogy engineering is the following: while conceptual
annotations provide a specification of how some ter-
minology is used to describe some domain (the stan-
dard role of OWL ontologies), natural language la-
bels (lexical annotations) provide a natural and rich
connection between formal objects (e.g. OWL classes
and properties) and their intended meaning. The in-
tuition is that grasping the intended interpretation of
an ontology requires not only an understanding of the
formal properties of the conceptual schema, but also
knowledge about the meaning of labels used for the
ontology elements. In other words, an OWL ontology
can be viewed as a collection of formal constraints be-
tween terms, whose intended meaning also depends
on lexical knowledge.
In most cases, lexical knowledge is used for an-
notating schema/ontology labels with lexical infor-
mation, typically WordNet senses. However, lexical
annotation is a difficult task, and making it accurate
may require a heavy user involvement. Typical prob-
lems are: coverage (a complete lexical database in-
cluding all possible terms does not exist); polysemy
(in natural language, many terms have multiple mean-
ings); compound terms (schemas and ontologies are
often labeled with compound nominal expressions,
like “full professor”, “table leg”, “football team”, and
the choice of the right lexical meaning often depends
on determining the relationship between terms); in-
tegration (a standard model/language for describing
lexical databases does not exist).
That is why several tools which were developed
for annotating sources only provide a GUI for sup-
porting the user in the manual execution of the task.
However, this manual work can be highly time con-
suming, and very tedious for humans.
MELIS tries to make annotation as automatic
as possible by providing a candidate lexical anno-
tation of the sources as the combination of lexical
knowledge (from WordNet) and domain knowledge
(if available). In addition, MELIS uses the WNEdi-
tor (Benassi et al., 2004) to support customized ex-
tensions of WordNet with missing words and senses.
In the following we describe the MELIS method,
its heuristic rules and the main features of WNEditor.
2.1 The MELIS Method
The way MELIS works is depicted in Figure 1. We
start from a collection of data sources which cover re-
lated domains, e.g. hotels and restaurants. In general
we do not assume that a domain ontology is initially
available, though this may be the case. The process is
a cycle which goes as follows:
1. a schema, which can be already partially anno-
MELIS - An Incremental Method for the Lexical Annotation of Domain Ontologies
241