
users, making it easier for a wider audience to lever-
age RL’s potential. Our no-code solution allows users
to quickly build and test RL models without extensive
programming skills, hence enabling faster prototyp-
ing and experimentation. RLML is developed to be
easily extensible to support a wide range of RL algo-
rithms. To the best of our knowledge, this work is a
first step in this direction for reinforcement learning.
With the use of the language workbench MPS, we
built a domain-specific modelling environment sup-
porting model editing, syntax checking, constraints
checking and validation, as well as code generation.
RLML achieves the abstraction needed in RL applica-
tions, by providing a configuration-like model to pro-
vide input values of the RL problem environment and
a choice of the RL algorithm. From that point, our
framework can generate executable code, run it and
display the results. The environment also provides a
comparator to compare results obtained with differ-
ent RL algorithms. It supports both Java and Python
implementations.
We demonstrated the use of our proposed frame-
work with the path finding and blackjack RL applica-
tions. It can also be used for business applications as
well as to get feedback from RL users at different lev-
els of expertise. Moreover, RLML can be helpful in
academia for making reinforcement learning accessi-
ble for non-technical students.
This work is a starting point towards developing
a framework for supporting various types of RL tech-
nologies, both model-free and model-based, with the
ultimate goal of democratizing access to advanced AI
capabilities. We are currently working on incorporat-
ing probability distributions and custom actions into
RLML, which will allow it to model real world use
cases more effectively.
ACKNOWLEDGEMENTS
This work has been partially supported by Natu-
ral Sciences and Engineering Research Council of
Canada (NSERC) and Toronto Metropolitan Univer-
sity. The authors would like to extend their thanks to
Prof. Nariman Farsad for his feedback on this work.
REFERENCES
Brunton, S. and Kutz, J. (2019). Data-Driven Science and
Engineering. Cambridge University Press.
Bucchiarone, A., Cabot, J., Paige, R., and Pierantonio, A.
(2020). Grand challenges in model-driven engineer-
ing: an analysis of the state of the research. Software
and Systems Modeling, 19:1–9.
Ding, Z., Huang, Y., Yuan, H., and Dong, H. (2020). Intro-
duction to reinforcement learning. In Deep Reinforce-
ment Learning, pages 47–123. Springer Singapore.
Franc¸ois-Lavet, V., Henderson, P., Islam, R., Bellemare,
M. G., Pineau, J., et al. (2018). An introduction
to deep reinforcement learning. Foundations and
Trends® in Machine Learning, 11(3-4):219–354.
Graesser, L. and Keng, W. L. (2019). Foundations of
Deep Reinforcement Learning: Theory and Practice
in Python. Addison-Wesley Professional, Boston.
Jiang, H. (2021). Machine learning fundamentals: a con-
cise introduction. Cambridge University Press.
Liaskos, S., Khan, S. M., Golipour, R., and Mylopoulos, J.
(2022). Towards goal-based generation of reinforce-
ment learning domain simulations. In iStar, pages 22–
28.
Meacham, S., Pech, V., and Nauck, D. (2020). Classifica-
tion algorithms framework (CAF) to enable intelligent
systems using jetbrains MPS domain-specific lan-
guages environment. IEEE access, 8:14832–14840.
Naveed, H., Arora, C., Khalajzadeh, H., Grundy, J., and
Haggag, O. (2024). Model driven engineering for
machine learning components: A systematic litera-
ture review. Information and Software Technology,
169:107423.
Radford, A., Wu, J., Amodei, D., Amodei, D., Clark, J.,
Brundage, M., and Sutskever, I. (2019). Better lan-
guage models and their implications. OpenAI blog,
1(2).
Ravichandiran, S. (2018). Hands-On Reinforcement Learn-
ing with Python: Master Reinforcement and Deep Re-
inforcement Learning Using OpenAI Gym and Tensor-
Flow. Packt Publishing, Limited, Birmingham.
Sapio, F. and Ratini, R. (2022). Developing and testing
a new reinforcement learning toolkit with unreal en-
gine. In Artificial Intelligence in HCI, pages 317–334,
Cham. Springer.
Schmidt, D. C. (2006). Model-driven engineering.
Computer-IEEE Computer Society, 39(2):25.
Sinani, N. (2022). RLML: A domain-specific modelling
language for reinforcement learning. MRP Report.
Sutton, R. S. and Barto, A. (2018). Reinforcement learning:
An introduction. A Bradford Book.
Verma, P., Dhanre, U., Khekare, S., Sheikh, S., and
Khekare, G. (2020). The optimal path finding algo-
rithm based on reinforcement learning. International
Journal of Software Science and Computational Intel-
ligence, 12:1–18.
Voelter, M., Benz, S., Dietrich, C., Engelmann, B., He-
lander, M., Kats, L. C. L., Visser, E., and Wachsmuth,
G. (2013). DSL Engineering - Designing, Implement-
ing and Using Domain-Specific Languages. CreateS-
pace Independent Publishing Platform, US.
Zhao, T., Huang, X., and Cao, Y. (2017). DeepDSL: A
compilation-based domain-specific language for deep
learning. In Proceedings of the 5th International Con-
ference on Learning Representations,.
Towards a Domain-Specific Modelling Environment for Reinforcement Learning
51