If we had the optimal (or existing) solution to the
problem in training data we could just plug the neural
network into the search as a heuristic and evaluate dif-
ference between the expected solution and the newly
found one. That would be the best metric to tell how
well does the neural network work as a heuristic func-
tion.
6 CONCLUSION
We proposed three novel planning problem represen-
tations suitable for training neural networks that can
be created from PDDL problem definition. We an-
alyzed IPC planning domains to filter out domains
which are explicitly defined on a grid which makes
them great candidates for our grid-based representa-
tions. We also picked domains which are not defined
on a grid and showed that it is still possible to use
a grid representation even when no grid is defined in
their PDDL.
We showed examples of all proposed representa-
tions and described their advantages and drawbacks.
The biggest advantage of problems represented by
grids is the grid structure which can be easily used
for neural network training. If we used such repre-
sentation on outputs from the available PDDL data
generators it would mean more reliable training data
for architectures that require the grid-based input.
We also discussed the domain-independence of
the representations as well as possible automatising
in terms of creating PDDL parsers that would be able
to create grid representations solely from the PDDL
files.
Finally, we shortly discussed training neural net-
works for planning. We pointed out problems like
obtaining high quality data, properties that the data
should have, hardness of the problems and choosing
correct loss function for training.
Main drawback of these representations is the lack
of action information. In order to correctly compute
heuristic or solve a planning problem, it is necessary
to understand the transition system of the problem and
how it is created. In this case, we only aim to repre-
sent the problem in a convenient manner and assume
that further information is provided as another input
to the architecture.
In the future, we would like to focus on N-
dimensional representation that would be able to add
definition of actions into the grid representation as
well. Next focus is on the PDDL to grid parsers which
could be beneficial set of tools for trying out different
machine learning techniques such as the mentioned
bootstrapping.
ACKNOWLEDGEMENTS
The work of Michaela Urbanovsk
´
a was
supported by the OP VVV funded project
CZ.02.1.01/0.0/0.0/16019/0000765 “Research
Center for Informatics” and the work of Anton
´
ın
Komenda was supported by the Czech Science
Foundation (grant no. 21-33041J).
REFERENCES
Aeronautiques, C., Howe, A., Knoblock, C., McDermott,
I. D., Ram, A., Veloso, M., Weld, D., SRI, D. W., Bar-
rett, A., Christianson, D., et al. (1998). Pddl— the
planning domain definition language. Technical re-
port, Technical Report.
Arfaee, S. J., Zilles, S., and Holte, R. C. (2010). Boot-
strap learning of heuristic functions. In Felner, A. and
Sturtevant, N. R., editors, Proceedings of the Third
Annual Symposium on Combinatorial Search, SOCS
2010, Stone Mountain, Atlanta, Georgia, USA, July
8-10, 2010. AAAI Press.
Asai, M. and Fukunaga, A. (2017). Classical planning in
deep latent space: From unlabeled images to PDDL
(and back). In Besold, T. R., d’Avila Garcez, A. S.,
and Noble, I., editors, Proceedings of the Twelfth In-
ternational Workshop on Neural-Symbolic Learning
and Reasoning, NeSy 2017, London, UK, July 17-18,
2017, volume 2003 of CEUR Workshop Proceedings.
CEUR-WS.org.
Asai, M. and Fukunaga, A. (2018). Classical planning in
deep latent space: Bridging the subsymbolic-symbolic
boundary. In Thirty-Second AAAI Conference on Ar-
tificial Intelligence.
Bento, D. S., Pereira, A. G., and Lelis, L. H. S. (2019).
Procedural generation of initial states of sokoban. In
Kraus, S., editor, Proceedings of the Twenty-Eighth
International Joint Conference on Artificial Intelli-
gence, IJCAI 2019, Macao, China, August 10-16,
2019, pages 4651–4657. ijcai.org.
Franc
`
es, G., Corr
ˆ
ea, A. B., Geissmann, C., and Pommeren-
ing, F. (2019). Generalized potential heuristics for
classical planning. In Proceedings of the Twenty-
Eighth International Joint Conference on Artificial In-
telligence, IJCAI-19, pages 5554–5561. International
Joint Conferences on Artificial Intelligence Organiza-
tion.
Geffner, H. (2018). Model-free, model-based, and general
intelligence. CoRR, abs/1806.02308.
Groshev, E., Tamar, A., Goldstein, M., Srivastava, S., and
Abbeel, P. (2018). Learning generalized reactive poli-
cies using deep neural networks. In 2018 AAAI Spring
Symposium Series.
LeCun, Y., Boser, B., Denker, J. S., Henderson, D., Howard,
R. E., Hubbard, W., and Jackel, L. D. (1989). Back-
propagation applied to handwritten zip code recogni-
tion. Neural computation, 1(4):541–551.
Grid Representation in Neural Networks for Automated Planning
879