Authors:
Ryo Meguro
1
;
Hiroya Kato
2
;
Shintaro Narisada
2
;
Seira Hidano
2
;
Kazuhide Fukushima
2
;
Takuo Suganuma
1
and
Masahiro Hiji
1
Affiliations:
1
Tohoku University, Miyagi, Japan
;
2
KDDI Research, Inc., Saitama, Japan
Keyword(s):
Graph Neural Networks, AI Security, Backdoor Attacks.
Abstract:
Graph neural networks (GNNs) can obtain useful information from graph structured data. Although its great capability is promising, GNNs are vulnerable to backdoor attacks, which plant a marker called trigger in victims’ models to cause them to misclassify poisoned data with triggers into a target class. In particular, a clean label backdoor attack (CLBA) on the GNNs remains largely unexplored. Revealing characteristics of the CLBA is vital from the perspective of defense. In this paper, we propose the first gradient based CLBA on GNNs for graph classification tasks. Our attack consists of two important phases, the graph embedding based pairing and the gradient based trigger injection. Our pairing makes pairs from graphs of the target class and the others to successfully plant the backdoor in the target class area in the graph embedding space. Our trigger injection embeds triggers in graphs with gradient-based scores, yielding effective poisoned graphs. We conduct experiments on multi
ple datasets and GNN models. Our results demonstrate that our attack outperforms the existing CLBA using fixed triggers. Our attack surpasses attack success rates of the existing CLBA by up to 50%. Furthermore, we show that our attack is difficult to detect with an existing defense.
(More)