of classes to a specific class. In case of bimodal and
flip noise patterns, EM achieves second best results in
terms of accuracy and error rate.
5 CONCLUSION
Motivated by the need for accurate data labeling of
crowd workers and using the provided labels for train-
ing DNNs, we design an iterative method for label
aggregation and training DNN together. The prior art
performs label aggregation and training classifier in
two separate processes. We propose LABNET that
considers aggregation and training in contact with
each other. In our model, the classifier extracts the
prior knowledge for passing to the aggregation algo-
rithm. Also, the estimated correct labels by aggrega-
tion algorithm are used to train the classifier. In ad-
dition, we design an algorithm to decide when DNN
needs to be trained through the aggregation algorithm
iteration. Compared to the baselines, LABNET out-
performs in most scenarios, especially for in challeng-
ing scenarios with large number of classes.
ACKNOWLEDGEMENTS
This work has been partly funded by the Swiss
National Science Foundation NRP75 project
407540 167266.
REFERENCES
Cousineau, D. and Helie, S. (2013). Improving maximum
likelihood estimation using prior probabilities: A tuto-
rial on maximum a posteriori estimation and an exam-
ination of the weibull distribution. Tutorials in Quan-
titative Methods for Psychology, 9(2):61–71.
Dawid, A. P. and Skene, A. M. (1979). Maximum likeli-
hood estimation of observer error-rates using the em
algorithm. Applied statistics, pages 20–28.
Gaunt, A., Borsa, D., and Bachrach, Y. (2016). Train-
ing deep neural nets to aggregate crowdsourced re-
sponses. In Proceedings of the Thirty-Second Con-
ference on Uncertainty in Artificial Intelligence. AUAI
Press, page 242251.
Ghiassi, A., Birke, R., Han, R., and Chen, L. Y. (2021).
Labelnet: Recovering noisy labels. In International
Joint Conference on Neural Networks (IJCNN), pages
1–8. IEEE.
Ghiassi, A., Younesian, T., Zhao, Z., Birke, R., Schiavoni,
V., and Chen, L. Y. (2019). Robust (deep) learning
framework against dirty labels and beyond. In Inter-
national Conference on Trust, Privacy and Security in
Intelligent Systems and Applications (TPS-ISA), pages
236–244. IEEE.
Han, B., Yao, Q., Yu, X., Niu, G., Xu, M., Hu, W., Tsang,
I., and Sugiyama, M. (2018). Co-teaching: Robust
training of deep neural networks with extremely noisy
labels. In NeurIPS, pages 8527–8537.
Hendrycks, D., Mazeika, M., Wilson, D., and Gimpel, K.
(2018). Using trusted data to train deep networks on
labels corrupted by severe noise. In NeurIPS, pages
10456–10465.
Hong, C., Ghiassi, A., Zhou, Y., Birke, R., and Chen,
L. Y. (2021). Online label aggregation: A variational
bayesian approach. In Web Conference 2021, WWW
’21, page 1904–1915. ACM.
Imran, M., Mitra, P., and Castillo, C. (2016). Twitter as
a lifeline: Human-annotated twitter corpora for NLP
of crisis-related messages. In Calzolari, N., Choukri,
K., Declerck, T., Goggi, S., Grobelnik, M., Maegaard,
B., Mariani, J., Mazo, H., Moreno, A., Odijk, J., and
Piperidis, S., editors, Language Resources and Evalu-
ation LREC. European Language Resources Associa-
tion (ELRA).
Jiang, L., Zhou, Z., Leung, T., Li, L., and Fei-Fei, L. (2018).
Mentornet: Learning data-driven curriculum for very
deep neural networks on corrupted labels. In ICML,
pages 2309–2318.
Khetan, A., Lipton, Z. C., and Anandkumar, A. (2018).
Learning from noisy singly-labeled data. In ICLR.
Kim, H.-C. and Ghahramani, Z. (2012). Bayesian classifier
combination. In Artificial Intelligence and Statistics,
pages 619–627.
Krizhevsky, A., Nair, V., and Hinton, G. (2009). Cifar-
10/100 (Canadian Institute for Advanced Research).
Li, J., Socher, R., and Hoi, S. C. (2020). Dividemix: Learn-
ing with noisy labels as semi-supervised learning. In
ICLR.
Liu, Q., Peng, J., and Ihler, A. T. (2012). Variational infer-
ence for crowdsourcing. In Advances in neural infor-
mation processing systems, pages 692–700.
Patrini, G., Rozza, A., Krishna Menon, A., Nock, R., and
Qu, L. (2017). Making deep neural networks robust
to label noise: A loss correction approach. In IEEE
CVPR, pages 1944–1952.
Shu, J., Xie, Q., Yi, L., Zhao, Q., Zhou, S., Xu, Z., and
Meng, D. (2019). Meta-weight-net: Learning an ex-
plicit mapping for sample weighting. In NIPS, pages
1919–1930.
Simpson, E. D., Venanzi, M., Reece, S., Kohli, P., Guiver,
J., Roberts, S. J., and Jennings, N. R. (2015). Lan-
guage understanding in the wild: Combining crowd-
sourcing and machine learning. In Proceedings of
the 24th international conference on world wide web,
pages 992–1002. International World Wide Web Con-
ferences Steering Committee.
Venanzi, M., Guiver, J., Kazai, G., Kohli, P., and Shokouhi,
M. (2014). Community-based bayesian aggregation
models for crowdsourcing. In Proceedings of the 23rd
international conference on World wide web, pages
155–164. ACM.
LABNET: A Collaborative Method for DNN Training and Label Aggregation
65