For measuring the metrics we randomly select 100
data points and create the testing dataset using their
class labels and generated explanations. The stability
metric is measured by applying K-means clustering
with two clusters to group explanations in the testing
dataset. For simplicity, we use the three explanations
generated by the decision tree (DT), converting each
explanation string into a unique numerical value to
form an integer array. The assigned cluster labels are
then compared with the predicted class labels to eval-
uate whether instances of the same class have simi-
lar explanations. To measure the separability metric,
two subsets S1 and S2 of the testing dataset are se-
lected corresponding to different class labels. Then,
for each instance in S1, its explanation is compared
with all other explanations of instances in S2. If the
explanation have no duplicates, it satisfies the sepa-
rability metric. Finally, the identity of the explana-
tions offered by the various deterministic techniques
may be easily measured theoretically. The explana-
tions generated by the decision tree is rule based thus
conforming to complete identity conservation. Addi-
tionally, due to the nature of KNN alrogithm identical
instances will have identical explanations.
8.1.2 Results
The experimental findings can be seen in Table 10.
The figures in this table show the percentage of in-
stances that meet the specified metrics. From the table
we can infer that identity metric is 100%, as identical
instances will have a similar explanations. The stabil-
ity is very high, thus conforming that instances with
same class labels have comparable interpretations. Fi-
nally, the separability is also very high, thus acknowl-
edging that dissimilar instances have dissimilar expla-
nations.
9 CONCLUSION
This paper demonstrates the effectiveness of an
ensemble-based classifier using a tweet’s diffusion
pattern for accurate misinformation detection. We im-
prove the classification by using features inspired by
epidemiology and recent COVID-19 research, while
providing understandable predictions. The intrinsic
explanations help users to understand the predicted
class label without compromising accuracy.
Future work will focus on the following areas:
• Incorporating statistical and qualitative measures
to evaluate the results and generated explanations.
• Expanding the model’s applicability to other so-
cial networks such as Instagram and Facebook.
• Investigate and document how hyperparameters,
such as the value of k in k-NN, sampling rate, af-
fect model performance.
• Conduct deeper analysis on the consistency and
comparability of explanations generated by differ-
ent models (e.g., k-NN vs. DT).
Table 10: Metrics for the evaluation of explanations.
Metric Score
Stability 89%
Separability 97%
Identity 100%
REFERENCES
Brainard, J., Jones, N. R., Harrison, F. C., Hammer, C. C.,
and Lake, I. R. (2023). Super-spreaders of novel coro-
naviruses that cause sars, mers and covid-19: a sys-
tematic review. Annals of Epidemiology, 82:66–76.e6.
Carlos Castillo, M. M. and Poblete, B. (2011). Information
credibility on twitter. page 675–684.
Du, M., Liu, N., and Hu, X. (2018). Techniques for inter-
pretable machine learning.
ElShawi, R., Sherif, Y., Al-Mallah, M., and Sakr, S. (2021).
Interpretability in healthcare: A comparative study
of local machine learning interpretability techniques.
Computational Intelligence, 37(4):1633–1650.
Friggeri, A., Adamic, L., Eckles, D., and Cheng, J. (2014).
Rumor cascades. Proceedings of the 8th International
Conference on Weblogs and Social Media, ICWSM
2014, pages 101–110.
Juul, J. L. and Ugander, J. (2021). Comparing informa-
tion diffusion mechanisms by matching on cascade
size. Proceedings of the National Academy of Sci-
ences, 118(46).
Kwon, S., Cha, M., Jung, K., Chen, W., and Wang, Y.
(2013). Prominent features of rumor propagation in
online social media. In 2013 IEEE 13th International
Conference on Data Mining, pages 1103–1108.
Liu, Y. and Wu, Y.-F. (2018). Early detection of fake news
on social media through propagation path classifica-
tion with recurrent and convolutional networks. Pro-
ceedings of the AAAI Conference on Artificial Intelli-
gence, 32(1).
Lu, Y.-J. and Li, C.-T. (2020). GCAN: Graph-aware co-
attention networks for explainable fake news detec-
tion on social media. In Proceedings of the 58th An-
nual Meeting of the Association for Computational
Linguistics.
Ma, J., Gao, W., Wei, Z., Lu, Y., and Wong, K.-F. (2015).
Detect rumors using time series of social context in-
formation on microblogging websites.
Ma, J., Gao, W., and Wong, K.-F. (2017). Detect rumors in
microblog posts using propagation structure via kernel
learning. In Proceedings of the 55th Annual Meeting
of the Association for Computational Linguistics (Vol-
ume 1: Long Papers).
KDIR 2024 - 16th International Conference on Knowledge Discovery and Information Retrieval
356