quencies and elevation rates: combinations of larger
radii and slow elevations differ significantly from
tasks involving fast elevation changes and small radii.
Fine-tuning on the same task family (multi-
sinusoidal) with controlled joint randomization led to
a significant increase in overall prediction accuracy,
depending on the number of examples provided.
Fine-tuning on Spiral tasks and zero-shot testing
on the Circle task resulted in poor performance. This
underscores the difficulty in predicting OSC tasks
generally, particularly during initial transient phases.
Direct fine-tuning on the Circle task yielded better
results compared to fine-tuning on Spiral tasks and
testing on Spiral. This observation suggests that the
model’s effectiveness is constrained to in-distribution
(ID) and slightly out-of-distribution (OOD) tasks.
6 CONCLUSIONS
This work tackles the problem of learning a meta-
model of robot dynamics using an encoder-decoder
Transformer architecture. The challenge lies in the
simulation domain, where the meta-model accurately
predicts complex systems over long sequences based
on a 20% context and 80% prediction of the over-
all trajectory. The results indicate that Transformer-
based models can learn dynamics in a zero-shot or
few-shot fashion within control action distributions,
suggesting their potential for use in robotics. The
results also highlight fine-tuning is advantageous in
these scenarios and more practical compared to train-
ing such models from scratch.
Current limitations are mainly related to general-
ization. Following a black-box approach to gener-
alize a single robotic arm independently of the type
of control action appears structurally unfeasible. Re-
sults show distinctions between in-distribution (ID)
and out-of-distribution (OOD) control actions, high-
lighting the critical role of the model’s inputs.
For future work, the results presented pave the
way for pre-compensating control actions in unknown
systems, particularly where estimating parameters
such as payload, joint stiffness, and damping is chal-
lenging. This approach can be extended beyond
control distributions to encompass Transfer Learning
across diverse robot morphologies.
ACKNOWLEDGMENTS
This paper has received funding from the Hasler
Foundation under the GENERAI (GENerative
Robotics AI) Project.
REFERENCES
Chen, C., Wang, T., Liu, C., Liu, Y., and Cheng, L.
(2023). Lightweight convolutional transformers en-
hanced meta-learning for compound fault diagnosis of
industrial robot. IEEE Transactions on Instrumenta-
tion and Measurement, 72:1–12.
Forgione, M., Pura, F., and Piga, D. (2023). From sys-
tem models to class models: An in-context learning
paradigm. IEEE Control Systems Letters, 7:3513–
3518.
Gupta, A., Fan, L., Ganguli, S., and Fei-Fei, L. (2022).
Metamorph: Learning universal controllers with
transformers. arXiv preprint arXiv:2203.11931.
Hospedales, T., Antoniou, A., Micaelli, P., and Storkey, A.
(2020). Meta-learning in neural networks: A survey.
Kirsch, L., Harrison, J., Sohl-Dickstein, J., and Metz,
L. (2024). General-purpose in-context learning
by meta-learning transformers. arXiv preprint
arXiv:2212.04458.
Li, Z., Kovachki, N., Azizzadenesheli, K., Liu, B., Bhat-
tacharya, K., Stuart, A., and Anandkumar, A. (2020).
Fourier neural operator for parametric partial differen-
tial equations. arXiv preprint arXiv:2010.08895.
Lin, T., Wang, Y., Liu, X., and Qiu, X. (2022). A survey of
transformers. AI Open.
Makoviychuk, V., Wawrzyniak, L., Guo, Y., Lu, M., Storey,
K., Macklin, M., Hoeller, D., Rudin, N., Allshire, A.,
Handa, A., and State, G. (2021). Isaac gym: High
performance gpu-based physics simulation for robot
learning. CoRR, abs/2108.10470.
Piga, D., Pura, F., and Forgione, M. (2024). On the adap-
tation of in-context learners for system identification.
In Proceedings of the 20th IFAC Symposium on Sys-
tem Identification, Boston, MA.
Radford, A., Wu, J., Child, R., Luan, D., Amodei, D.,
Sutskever, I., et al. (2019). Language models are un-
supervised multitask learners. OpenAI blog, 1(8):9.
Schmidhuber, J., Hochreiter, S., et al. (1997). Long short-
term memory. Neural Comput, 9(8):1735–1780.
Shahid, A. A., Piga, D., Braghin, F., and Roveda, L. (2022).
Continuous control actions learning and adaptation for
robotic manipulation through reinforcement learning.
Autonomous Robots, 46(3):483–498.
Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux,
M.-A., Lacroix, T., Rozi
`
ere, B., Goyal, N., Hambro,
E., Azhar, F., et al. (2023). Llama: Open and ef-
ficient foundation language models. arXiv preprint
arXiv:2302.13971.
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones,
L., Gomez, A. N., Kaiser, Ł., and Polosukhin, I.
(2017). Attention is all you need. Advances in neural
information processing systems, 30.
Yang, L., Liu, S., Meng, T., and Osher, S. J. (2023). In-
context operator learning with data prompts for differ-
ential equation problems. Proceedings of the National
Academy of Sciences, 120(39):e2310142120.
ICINCO 2024 - 21st International Conference on Informatics in Control, Automation and Robotics
156