deploying the trained models to real and simulation
environments.
Our results cast a new light on the performance
and accuracy of the trained models. The models'
overall accuracy and performance increased
proportionally with the continuous addition of
simulation data.
7 CONCLUSIONS AND FUTURE
WORK
In this paper, we have investigated the ability of
CNNs models to generalize and learn from a mixture
of real and artificial data. Two CNN models have
been evaluated where the first was trained from
scratch and the other was fine-tuned to fit the dataset
classes of the lane tracking task. The CNN models are
trained six times wherein each a different
combination of the collected real and simulation
datasets is used. The results show that the models
trained with a dataset collected from a particular
environment can work only in this environment and
fails when it is transferred to an unseen target
environment. Another promising finding was that the
models' performance increased significantly and were
able to generalize to both real and simulation
environments with the inclusion of simulation data.
On this basis, we conclude that a mixture of
simulation and real data can help the CNN models to
generalize in cases where datasets are scarce and
when models trained in a particular domain are
transferred to an unseen target domain. This paper
provides a good starting point for further research. In
our future research, we intend to examine more
complex model architectures and environments.
REFERENCES
Bayraktar, E., Yigit, C., & Boyraz, P. (2019). A hybrid
image dataset toward bridging the gap between real and
simulation environments for robotics. (pp. 23–40).
Machine Vision and Applications.
Bousmalis, K., & Levine, S. (2017, October 30). Closing
the Simulation-to-Reality Gap for Deep Robotic
Learning. Retrieved March 3, 2020, from
https://ai.googleblog.com/2017/10/closing-simulation-
to-reality-gap-for.html.
Cutler, M., Walsh, T. J., & How, J. P. (2014).
Reinforcement learning with multi-fidelity simulators.
(pp. 3888-3895). IEEE.
Duan, L., Xu, D., & Tsang, I. (2012). Learning with
augmented features for heterogeneous domain
adaptation. (pp. 667–674). arXiv.
Gamal, O., Imran, M., Roth, H., & Wahrburg, J. (2020).
Assistive Parking Systems Knowledge Transfer to End-
to-End Deep Learning for Autonomous Parking. (pp.
216-221). IEEE.
Hoffman, J., Guadarrama, S., Tzeng, E. S., Hu, R.,
Donahue, J., Girshick, R., Darrell, T., Saenko, K.
(2014). LSDA: Large scale detection through
adaptation. (pp. 3536-3544). arXiv.
James, S., Davison, A. J., & Johns, E. (2017). Transferring
end-to-end visuomotor control from simulation to real
world for a multi-stage task. arXiv.
Kulis, B., Saenko, K., & Darrell, T. (2011). What you saw
is not what you get: Domain adaptation using
asymmetric kernel transforms. (pp. 1785-1792). IEEE.
Li, Y., Wang, N., Shi, J., Liu, J., & Hou, X. (2016).
Revisiting batch normalization for practical domain
adaptation. arXiv.
Lin, M., Chen, Q., & Yan, S. (2013). Network in network.
arXiv.
Loquercio, A., Kaufmann, E., Ranftl, R., Dosovitskiy, A.,
Koltun, V., & Scaramuzza, D. (2019). Deep drone
racing: From simulation to reality with domain
randomization. 36. IEEE Transactions on Robotics.
Peng, X., Sun, B., Ali, K., & Saenko, K. (2014). Exploring
invariances in deep convolutional neural networks
using synthetic images. arXiv.
Sadeghi, F., & Levine, S. (2016). Cad2rl: Real single-image
flight without a single real image. arXiv.
Stark, M., Goesele, M., & Schiele, B. (2010). Back to the
Future: Learning Shape Models from 3D CAD Data. 2,
p. 5. BMVC.
Su, H., Qi, C. R., Li, Y., & Guibas, L. J. (2015). Render for
cnn: Viewpoint estimation in images using cnns trained
with rendered 3d model views. (pp. 2686-2694). ICCV.
Sünderhauf, N., Brock, O., Scheirer, W., Hadsell, R., Fox,
D., Leitner, J., Upcroft, B., Abbeel, P., Burgard, W.,
Milford, M., Corke, P. (2018). The limits and potentials
of deep learning for robotics. 37, pp. 405-420. IJRR.
Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S.,
Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich,
A. (2015). Going deeper with convolutions. (pp. 1-9).
IEEE.
Tkacz, M. (2005). Artificial neural networks in incomplete
data sets processing. In M. Kłopotek, S. Wierzchoń, &
T. K., Intelligent Information Processing and Web
Mining (Vol. 31, pp. 577-583). Berlin, Heidelberg,
Berlin: Springer.
Tobin, J., Fong, R., Ray, A., Schneider, J., Zaremba, W., &
Abbeel, P. (2017). Domain randomization for
transferring deep neural networks from simulation to
the real world. (pp. 23-30). IEEE.
Tzeng, E., Hoffman, J., Zhang, N., Saenko, K., & Darrell,
T. (2014). Deep Domain Confusion: Maximizing for
Domain Invariance. arXiv.
Yan, M., Frosio, I., Tyree, S., & Kautz, J. (2017). Sim-to-
real transfer of accurate grasping with eye-in-hand
observations and continuous control. arXiv.
Yosinski, J., Clune, J., Bengio, Y., & Lipson, H. (2014).
How transferable are features in deep neural networks?
NIPS ’14. NIPS Foundation.