moid b function with the additional trainable param-
eter b was introduced to tackle the limitations of the
standard sigmoid function used in the negative sam-
pling process. The improved update equations of the
SkipGram
b
model were used in the processes of DW
and n2v for generating graph embeddings, while the
k-means algorithm and the logistic regression model
were utilised to detect the communities of the graph
and to predict links between the nodes, respectively,
using the (calculated) embedded nodes. The experi-
mental results in real datasets showed that DW
b
and
n2v
b
converged faster than DW and n2v, respectively,
and attained higher ARI, Mod and AUC score in
the CiteSeer graph, as well as higher AUC score in
the Cora and PubMed graphs. Finally, the proposed
SkipGram
b
provided a more robust performance in
convergence speed than the standard SkipGram algo-
rithm, considering different values of learning rates.
ACKNOWLEDGEMENTS
This research is part of projects that have received
funding from the European Union’s H2020 research
and innovation programme under AIDA (GA No.
883596) and CREST (GA No. 833464).
REFERENCES
Banerjee., K., C.., V., Gupta., R., Vyas., K., H.., A., and
Mishra., B. (2021). Exploring alternatives to soft-
max function. In Proceedings of the 2nd International
Conference on Deep Learning Theory and Applica-
tions - DeLTA,, pages 81–86. INSTICC, SciTePress.
Cai, H., Zheng, V. W., and Chang, K. C.-C. (2018). A com-
prehensive survey of graph embedding: Problems,
techniques, and applications. IEEE Transactions on
Knowledge and Data Engineering, 30(9):1616–1637.
Cavallari, S., Zheng, V. W., Cai, H., Chang, K. C.-C., and
Cambria, E. (2017). Learning community embed-
ding with community detection and node embedding
on graphs. In Proceedings of the 2017 ACM on Con-
ference on Information and Knowledge Management,
pages 377–386.
Chowdhary, K. (2020). Natural language processing. In
Fundamentals of Artificial Intelligence, pages 603–
649. Springer.
Fawcett, T. (2006). An introduction to roc analysis. Pattern
recognition letters, 27(8):861–874.
Goyal, P. and Ferrara, E. (2018). Graph embedding tech-
niques, applications, and performance: A survey.
Knowledge-Based Systems, 151:78–94.
Grover, A. and Leskovec, J. (2016). node2vec: Scal-
able feature learning for networks. In Proceedings
of the 22nd ACM SIGKDD international conference
on Knowledge discovery and data mining, pages 855–
864.
Hartigan, J. A. and Wong, M. A. (1979). Algorithm as
136: A k-means clustering algorithm. Journal of the
royal statistical society. series c (applied statistics),
28(1):100–108.
He, T., Zhang, Z., Zhang, H., Zhang, Z., Xie, J., and Li,
M. (2019). Bag of tricks for image classification
with convolutional neural networks. In Proceedings
of the IEEE Conference on Computer Vision and Pat-
tern Recognition, pages 558–567.
Hochreiter, S. (1998). The vanishing gradient problem dur-
ing learning recurrent neural nets and problem solu-
tions. International Journal of Uncertainty, Fuzziness
and Knowledge-Based Systems, 6(02):107–116.
Hosmer Jr, D. W., Lemeshow, S., and Sturdivant, R. X.
(2013). Applied logistic regression, volume 398. John
Wiley & Sons.
Luo, X., Chang, X., and Ban, X. (2016). Regression and
classification using extreme learning machine based
on l1-norm and l2-norm. Neurocomputing, 174:179–
186.
Mikolov, T., Chen, K., Corrado, G., and Dean, J. (2013a).
Efficient estimation of word representations in vector
space. arXiv preprint arXiv:1301.3781.
Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., and
Dean, J. (2013b). Distributed representations of words
and phrases and their compositionality. Advances
in neural information processing systems, 26:3111–
3119.
Nguyen, G. H., Lee, J. B., Rossi, R. A., Ahmed, N. K., Koh,
E., and Kim, S. (2018). Continuous-time dynamic net-
work embeddings. In Companion Proceedings of the
The Web Conference 2018, pages 969–976.
Perozzi, B., Al-Rfou, R., and Skiena, S. (2014). Deepwalk:
Online learning of social representations. In Proceed-
ings of the 20th ACM SIGKDD international confer-
ence on Knowledge discovery and data mining, pages
701–710.
Rozemberczki, B., Davies, R., Sarkar, R., and Sutton, C.
(2019). Gemsec: Graph embedding with self clus-
tering. In Proceedings of the 2019 IEEE/ACM inter-
national conference on advances in social networks
analysis and mining, pages 65–72.
Sen, P., Namata, G., Bilgic, M., Getoor, L., Galligher, B.,
and Eliassi-Rad, T. (2008). Collective classification in
network data. AI magazine, 29(3):93–106.
Vinh, N. X., Epps, J., and Bailey, J. (2010). Information the-
oretic measures for clusterings comparison: Variants,
properties, normalization and correction for chance.
The Journal of Machine Learning Research, 11:2837–
2854.
Zhang, D., Yin, J., Zhu, X., and Zhang, C. (2018). Network
representation learning: A survey. IEEE transactions
on Big Data.
A Faster Converging Negative Sampling for the Graph Embedding Process in Community Detection and Link Prediction Tasks
93