are also required in the related tasks. We multiplied
the estimations of CTR and impression values and ob-
tained the click prediction for the next day.
The results show that the highest R
2
obtained by
multiplying CTR and impression was 0.81. The other
success criterion, which can be regarded as the total
success, is based on comparing the sum of actual and
predicted values over all hotels. We have achieved
95% SumSuccess criterion, which shows the effec-
tiveness of the features extracted from the original
dataset.
We applied Support Vector Regression (SVR) and
random forest algorithms which are known to be suc-
cessful regression algorithms. The results showed
that decision tree-based boosting algorithms outper-
formed SVR and random forest on this dataset. The
highest R-Squared value obtained in the prediction
of individual-hotel based CTR and impression values
are 0.65 and 0.84, respectively, both achieved by XG-
Boost. Another contribution is to observe that a sub-
set of features selected by mRMR technique achieves
comparable performance to using all of the features
in the machine learning model. The obtained results
showed that the most important features are the bid of
the last day and rating of the hotel for both CTR and
impression prediction. We should also note that the
variables representing the length of the closest holi-
day, the region of the hotel, and the position of the
advertisement of the OTA for the related hotel are
among the top-ranked variables in both CTR and im-
pression prediction problems. These results show that
they carry important and complementary information
about the target variables. As a future direction, we
aim to construct sequential models using different ar-
chitectures of recurrent neural networks for click pre-
diction.
REFERENCES
Adam-Bourdarios, C., Cowan, G., Germain-Renaud, C.,
Guyon, I., K
´
egl, B., and Rousseau, D. (2015).
The higgs machine learning challenge. In Journal
of Physics: Conference Series, volume 664, page
072015. IOP Publishing.
Agresti, A. (1996). An introduction to categorical data anal-
ysis.
Babajide Mustapha, I. and Saeed, F. (2016). Bioactive
molecule prediction using extreme gradient boosting.
Molecules, 21(8):983.
Balfer, J. and Bajorath, J. (2015). Systematic artifacts in
support vector regression-based compound potency
prediction revealed by statistical and activity land-
scape analysis. PloS one, 10(3):e0119301.
Chapelle, O., Manavoglu, E., and Rosales, R. (2015). Sim-
ple and scalable response prediction for display adver-
tising. ACM Transactions on Intelligent Systems and
Technology (TIST), 5(4):61.
Chen, T., He, T., Benesty, M., et al. (2015). Xgboost: ex-
treme gradient boosting. R package version 0.4-2,
pages 1–4.
Cheng, H., Zwol, R. v., Azimi, J., Manavoglu, E., Zhang,
R., Zhou, Y., and Navalpakkam, V. (2012). Multime-
dia features for click prediction of new ads in display
advertising. In Proceedings of the 18th ACM SIGKDD
international conference on Knowledge discovery and
data mining, pages 777–785. ACM.
Cootes, T. F., Ionita, M. C., Lindner, C., and Sauer, P.
(2012). Robust and accurate shape model fitting using
random forest regression voting. In European Confer-
ence on Computer Vision, pages 278–291. Springer.
Dave, K. S. and Varma, V. (2010). Learning the click-
through rate for rare/new ads from similar ads. In Pro-
ceedings of the 33rd international ACM SIGIR con-
ference on Research and development in information
retrieval, pages 897–898. ACM.
Effendi, M. J. and Ali, S. A. (2017). Click through rate
prediction for contextual advertisment using linear re-
gression. arXiv preprint arXiv:1701.08744.
Esmael, B., Arnaout, A., Fruhwirth, R., and Thonhauser, G.
(2015). A statistical feature-based approach for opera-
tions recognition in drilling time series. International
Journal of Computer Information Systems and Indus-
trial Management Applications, 5:454–461.
Friedman, J. H. (1997). On bias, variance, 0/1—loss, and
the curse-of-dimensionality. Data mining and knowl-
edge discovery, 1(1):55–77.
Friedman, J. H. (2001). Greedy function approximation: a
gradient boosting machine. Annals of statistics, pages
1189–1232.
Kotsiantis, S., Kanellopoulos, D., and Pintelas, P. (2006).
Data preprocessing for supervised leaning. Interna-
tional Journal of Computer Science, 1(2):111–117.
Li, J., Zhang, P., Cao, Y., Liu, P., and Guo, L. (2012). Effi-
cient behavior targeting using svm ensemble indexing.
In 2012 IEEE 12th International Conference on Data
Mining, pages 409–418. IEEE.
Loshin, D. (2001). Enterprise knowledge management: The
data quality approach.
Lup Low, W., Lee, M., and Ling, T. (2001). A knowledge-
based approach for duplicate elimination in data
cleaning. Information Systems, 26:585–606.
Malani, J., Sinha, N., Prasad, N., and Lokesh, V. Forecast-
ing bike sharing demand.
M
¨
uller, A. C., Guido, S., et al. (2016). Introduction to ma-
chine learning with Python: a guide for data scien-
tists, volume 35. ” O’Reilly Media, Inc.”.
Nabi-Abdolyousefi, R. (2015). Conversion rate prediction
in search engine marketing. PhD thesis.
Richardson, M., Dominowska, E., and Ragno, R. (2007).
Predicting clicks: Estimating the click-through rate
for new ads. In Proceedings of the 16th International
Conference on World Wide Web, WWW ’07, pages
521–530, New York, NY, USA. ACM.
ICPRAM 2019 - 8th International Conference on Pattern Recognition Applications and Methods
628