Table 5: Standard deviation of the F
1
scores for all the 18
labels, considering the two training sets (Random and Fair).
k Random Fair Difference
1000 0.194 0.157 0.037
2000 0.191 0.163 0.029
3000 0.193 0.163 0.030
4000 0.196 0.169 0.027
5000 0.191 0.171 0.020
10000 0.187 0.174 0.013
4 CONCLUSIONS
In this work, we present and discuss a new algorithm
to generate fair subsets from unbalanced datasets.
The results of ILP algorithm in the multi-label image
classification task showed consistent improvements
compared to the random sub-selection of the original
training set, considering both the global scope (macro
F1 score), and the F1 score of the less frequent labels.
As future research directions, we envision the in-
vestigation of the computational complexity of the
Fairer Coverage Problem and the application of our
method to different datasets. The HPA dataset is a
special case where we have a single characteristic for
each sample, but our method could easily be adapted
to select a fairer dataset from a more complex dataset,
i.e., containing more than one attribute. We also be-
lieve the method will be useful when applied to large
datasets that cannot be used in full for the training
phase due to computational limitations.
ACKNOWLEDGEMENTS
The authors would like to thank the S
˜
ao Paulo
Research Foundation [grants #2015/11937-9,
#2017/12646-3, #2020/16439-5]; Coordination for
the Improvement of Higher Education Person-
nel; and the National Council for Scientific and
Technological Development [grants #304380/2018-0,
#306454/2018-1, #309330/2018-1, #161015/2021-2].
REFERENCES
Angwin, J., Larson, J., Mattu, S., and Kirchner, L. (2016).
Machine Bias. ProPublica.
Asudeh, A., Berger-Wolf, T., DasGupta, B., and Sidiropou-
los, A. (2022). Maximizing coverage while en-
suring fairness: a tale of conflicting objective.
arXiv:2007.08069v3, pages 1–44.
Christin, A., Rosenblat, A., and Boyd, D. (2015). Courts
and predictive algorithms. In Data & CivilRight,
Washington, DC.
Chung, J. (2022). Racism In, Racism Out - A Primer on
Algorithmic Racism. Public Citizen.
Galhotra, S., Shanmugam, K., Sattigeri, P., and Varshney,
K. R. (2020). Causal Feature Selection for Algorith-
mic Fairness. arXiv:2006.06053v2, pages 1–12.
Gichoya, J. W., Banerjee, I., Bhimireddy, A. R., Burns,
J. L., Celi, L. A., Chen, L.-C., Correa, R., Dullerud,
N., Ghassemi, M., Huang, S.-C., et al. (2022). Ai
recognition of patient race in medical imaging: a mod-
elling study. The Lancet Digital Health.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep
residual learning for image recognition. In Conference
on Computer Vision and Pattern Recognition (CVPR),
pages 770–778. IEEE.
Hundt, A., Agnew, W., Zeng, V., Kacianka, S., and Gombo-
lay, M. (2022). Robots Enact Malignant Stereotypes.
In ACM Conference on Fairness, Accountability, and
Transparency, pages 743–756.
Ingold, D. and Soper, S. (2016). Amazon Doesn’t Consider
the Race of Its Customers. Should It? Bloomberg.
Available at http://bloom.bg/3p0DHKz.
Kingma, D. P. and Ba, J. (2017). Adam: A Method for
Stochastic Optimization. arXiv:1412.6980, pages 1–
15.
Kleinberg, J., Ludwig, J., Mullainathan, S., and Ram-
bachan, A. (2018). Algorithmic Fairness. AEA Papers
and Proceedings, 108:22–27.
Koenecke, A., Nam, A., Lake, E., Nudell, J., Quartey, M.,
Mengesha, Z., Toups, C., Rickford, J. R., Jurafsky,
D., and Goel, S. (2020). Racial disparities in auto-
mated speech recognition. Proceedings of the Na-
tional Academy of Sciences, 117(14):7684–7689.
Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012).
ImageNet classification with deep convolutional neu-
ral networks. In Advances in Neural Information
Processing Systems (NIPS), pages 1097–1105. Curran
Associates, Inc.
Lin, Y., Guan, Y., Asudeh, A., and Jagadish, H. V. J. (2020).
Identifying Insufficient Data Coverage in Databases
with Multiple Relations. Proceedings of the VLDB
Endowment, 13(12):2229–2242.
O’Neil, C. (2017). Weapons of math destruction: How big
data increases inequality and threatens democracy.
Crown.
Roh, Y., Lee, K., Whang, S., and Suh, C. (2021). Sam-
ple selection for fair and robust training. Advances in
Neural Information Processing Systems, 34:815–827.
Rose, A. (2010). Are Face-Detection Cameras Racist?
Time. Available at https://bit.ly/3A7lIsC.
Silva, T. (2020). Algorithmic Racism Timeline. Available
at http://bit.ly/3O6RJYC.
Wang, F., Xie, Z., Pei, Z., and Liu, D. (2022). Emer-
gency Relief Chain for Natural Disaster Response
Based on Government-Enterprise Coordination. Inter-
national Journal of Environmental Research and Pub-
lic Health, 19(18).
Waters, A. and Miikkulainen, R. (2014). GRADE: Machine
learning support for graduate admissions. AI Maga-
zine, 35(1):64–64.
VISAPP 2023 - 18th International Conference on Computer Vision Theory and Applications
744