Table 2: Results for all experiments. Experiment 1: Full dataset, in-range scenes. Experiment 2: Full dataset, out-of-range
scenes. Experiment 3: Minimal dataset, in-range scenes. Experiment 4: Minimal dataset, out of range scenes. The subscript
rb indicates robot/ball and t indicates target.
Experiment 1
Type Total Hit Target Hit Ball/Missed Target Missed Ball
# # # Mean
rb
Med
rb
Mean
t
Med
t
# Mean
rb
Med
rb
MLP 9 0 5 0.1438 0.1404 0.4504 0.4607 4 0.1375 0.1287
MLP/GAN 11 3 4 0.2023 0.2022 0.3246 0.3772 4 0.2069 0.2382
Experiment 2
MLP 8 0 0 - - - - 8 0.2985 0.3312
MLP/GAN 8 0 2 0.2816 0.2816 0.5370 0.5370 6 0.3022 0.3395
Experiment 3
MLP 36 12 11 0.2106 0.1990 0.3368 0.2989 13 0.2467 0.2785
MLP/GAN 36 21 7 0.2254 0.2223 0.3854 0.3321 8 0.2524 0.2432
Experiment 4
MLP 16 6 4 0.1791 0.1898 0.3399 0.2624 6 0.43 0.3977
MLP/GAN 16 6 1 0.2128* 0.2128* 0.3039* 0.3039* 9 0.3439 0.2427
*The mean and median results match as there is only a single trial where the ball was hit but the target was missed.
to train the generator on an intermediate layer of
the discriminator, which is the actual features of the
dataset, as opposed to a singular output that inade-
quately combines those features.
Another line of inquiry is to use a more recent
GAN architecture. We picked the WGAN (Arjovsky
et al., 2017) to minimize mode-collapse. However,
mode-collapse was still a problem in Experiment 4.
A natural response is to use a GAN such as that de-
scribed in (Gulrajani et al., 2017), which incorporates
further technical improvements.
Regardless of these future developments, we be-
lieve that we have shown GAN-augmentation is a vi-
able path for developing robust controllers for phys-
ical robots in scenarios where a sufficient number of
the physical trials is difficult to obtain.
ACKNOWLEDGEMENTS
These experiments were carried out while Todd Flyr
was a visiting student at King’s College London. We
are grateful to King’s for hosting him, and for allow-
ing us to use the robotics facilities in the Department
of Engineering.
REFERENCES
Arjovsky, M., Chintala, S., and Bottou, L. (2017). Wasser-
stein generative adversarial networks. In Precup, D.
and Teh, Y. W., editors, Proceedings of the 34th Inter-
national Conference on Machine Learning, volume 70
of Proceedings of Machine Learning Research, pages
214–223, International Convention Centre, Sydney,
Australia. PMLR.
Bongard, J. C. (2015). Using robots to investigate the evo-
lution of adaptive behavior. Current Opinion in Be-
havioral Sciences, 6:168–173.
Bowles, C., Chen, L., Guerrero, R., Bentley, P., Gunn, R.,
Hammers, A., Dickie, D. A., Hern
´
andez, M. V., Ward-
law, J., and Rueckert, D. (2018). GAN augmentation:
Augmenting training data using generative adversarial
networks. arXiv preprint arXiv:1810.10863.
Chatfield, K., Simonyan, K., Vedaldi, A., and Zisserman,
A. (2014). Return of the devil in the details: Delving
deep into convolutional nets. In Proceedings of the
British Machine Vision Conference. BMVA Press.
Flyr, T. W. and Parsons, S. (2019). Towards adversar-
ial training for mobile robots. In Proceedings of the
20th Towards Autonomous Robotic Systems Confer-
ence, London.
Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., and
Courville, A. C. (2017). Improved training of Wasser-
stein GANs. In Guyon, I., Luxburg, U. V., Bengio, S.,
Wallach, H., Fergus, R., Vishwanathan, S., and Gar-
nett, R., editors, Advances in Neural Information Pro-
cessing Systems 30, pages 5767–5777. Curran Asso-
ciates, Inc.
Gupta, A., Johnson, J., Fei-Fei, L., Savarese, S., and Alahi,
A. (2018). Social GAN: socially acceptable trajec-
tories with generative adversarial networks. CoRR,
abs/1803.10892.
Lee, J. and Ryoo, M. S. (2017). Learning robot activities
from first-person human videos using convolutional
future regression. CoRR, abs/1703.01040.
Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V.,
Radford, A., Chen, X., and Chen, X. (2016). Im-
proved techniques for training GANs. In Lee, D. D.,
Sugiyama, M., Luxburg, U. V., Guyon, I., and Garnett,
R., editors, Advances in Neural Information Process-
ing Systems 29, pages 2234–2242. Curran Associates,
Inc.
Improving Learning in a Mobile Robot using Adversarial Training
89