Figure 10: Precision@K (K = 1, 3) and MRR scores on
the (top-3) answers achieved by AdProc and other query
processing approaches for the 1,750 test queries
and May 27, 2023. The metric P@1, P@3, and MRR
were computed based on the evaluation provided by
the 135 Facebook users on the 1,750 test cases, which
serve as the ground truth for this empirical study.
As shown in Figure 10, QuePR outperforms the
other four querying systems based on P@1, P@3,
and MRR, which verifies the effectiveness of AdProc.
The results are statistically significant based on the
Wilcoxon Signed-Ranks Test (p < 0.01).
Among all the five approaches, the P@1, P@3,
and MRR values for FAQFinder are the lowest, ex-
cept the Random approach, since FAQFinder uses a
simple method that does not compare numerical at-
tributes. On individual category, we observed that the
lowest scores on the three measures for AdProc occur
in the jobs category. For this category, appraisers did
not consider the answers based on their similarity to
the original query. For example, a Java programmer
job is closely related to a C++ programmer job, but
the appraisers considered the answers based on which
result is more relevant to their own expertise and ex-
perience, which is different from one user to another.
6 CONCLUSION
We have introduced AdProc, a closed domain natural
languagequery processing system on multiple ads do-
mains, which (i) automates the process of classifying,
extracting, and populating data from online ads to its
underlying database, (ii) relies on simple probabilistic
models to determine the domain an ad query belongs,
and (iii) generates answers that match the informa-
tion needs expressed in an ad query. Empirical stud-
ies conducted on a set of 80,000 online ads show that
AdProc is highly effective in classifying ads in mul-
tiple domains and labeling and extracting their data,
with accuracy in the ninety percentile. Furthermore,
the approaches adopted by AdProc outperform other
machine learning approaches (up to 9%) in accom-
plishing the same task. In addition, a conducted study
has verified the effectiveness of AdProc in answering
natural language queries in multiple ads domains.
For future work, we intend to further enhance
AdProc so that it can (i) automatically define the
schema of the underlying database for storing ads
from multiple domains, and (ii) handle online ads that
include multiple products within the same ad, such as
video games ads.
REFERENCES
Allison, B. (2008). An Improved Hierarchical Bayesian
Model of Language for Document Classification. In
Proc. of COLING, pages 25–32.
Burke, R., Hammond, K., Kulyukin, V., Lytinen, S., To-
muro, N., and Schoenberg, S. (1997). Question An-
swering from Frequently Asked Question Files: Ex-
periences with the FAQ Finder System. AI Magazine,
18(2):57–57.
Callan, J. and Connell, M. (2001). Query-Based Sampling
of Text Databases. ACM TOIS, 19(2):97–130.
Chieu, H. and Ng, H. (2003). Named Entity Recognition
with a Maximum Entropy Approach. In Proc. of Conf.
on Natural Language Learning, pages 160–163.
Chouaib, H., Cloppet, F., and Tabbone, S. (2009). Generic
Feature Selection and Document Processing. In Proc.
of ICDAR, pages 356–360.
Cohen, W. (1995). Fast and Effective Rule Induction. In
Proc. of ICML, pages 115–123.
Dayanik, A., Lewis, D., Madigan, D., Menkov, V., and
Genkin, A. (2006). Constructing Informative Prior
Distri- butions from Domain Knowledge in Text
Classifica- tion. In Proc. of SIGIR, pages 493–500.
Demner-Fushman, D. and Lin, J. (2007). Answering Clin-
ical Questions with Knowledge-Based and Statistical
Techniques. Comp. Linguistics, 33(1):63–103.
Hall, M. and Frank, E. (2008). Combining Naive Bayes and
Decision Tables. In Proc. of Florida Artificial Intelli-
gence Research Society Conf.
Khare, R. and An, Y. (2009). An Empirical Study on Using
Hidden Markov Model for Search Interface Segment-
ation. In Proc. of ACM CIKM, pages 17–26.
Li, B. and Han, L. (2013). Distance Weighted Cosine Sim-
ilarity Measure for Text Classification. In Proc. of
IDEAL, pages 611–618. Springer.
Liu, Y., Lin, Y., and Chen, Z. (2003). Using Hidden Markov
Model for Information Extraction Based on Multiple
Templates. In Proc. of NLP-KE, pages 394–399.
Liu, Y. and Zheng, Y. (2005). One-Against-All Multi-Class
SVM Classification Using Reliability Measures. In
Proc. of IJCNN, pages 849–854.
McCallum, A. and Nigam, K. (1998). A Comparison of
Event Models for Naive Bayes Text Classification. In
Proc. of AAAI Workshop on LTC, pages 41–48.