Machine learning (ML) has lately achieved impressive breakthroughs in several fields, enabling a plethora of exciting applications. However, mainstream ML techniques often have an undesirable property: they are not directly understandable by humans, thus humans cannot trust them in high-stakes or life-critical scenarios. A subfield of AI called interpretable AI (IAI) addresses this problem: generating models that are easy to understand for humans and, consequently, trustworthy. While several approaches apply IAI techniques to reinforcement learning problems, addressing the case in which an agent has to act in a continuous action space is still an open question. In this work, we propose a cooperative co-evolutionary approach based on grammatical evolution, Q-learning, and the univariate marginal distribution algorithm, specifically designed to train IAI agents (in the form of binary decision trees) capable of acting in environments with continuous action spaces. The experimental results show that our method is able to solve two well-known OpenAI Gym test cases reaching state-of-the-art performance. Moreover, a quantitative post hoc analysis reveals that the obtained solutions are more interpretable than those reported in the literature.
A co-evolutionary approach to interpretable reinforcement learning in environments with continuous action spaces / Custode, Leonardo Lucio; Iacca, Giovanni. - (2021), pp. 1-8. (Intervento presentato al convegno 2021 IEEE Symposium Series on Computational Intelligence, SSCI 2021 tenutosi a Orlando, FL, USA nel 5th December 2021-7th December 2021) [10.1109/SSCI50451.2021.9660048].
A co-evolutionary approach to interpretable reinforcement learning in environments with continuous action spaces
Custode, Leonardo Lucio;Iacca, Giovanni
2021-01-01
Abstract
Machine learning (ML) has lately achieved impressive breakthroughs in several fields, enabling a plethora of exciting applications. However, mainstream ML techniques often have an undesirable property: they are not directly understandable by humans, thus humans cannot trust them in high-stakes or life-critical scenarios. A subfield of AI called interpretable AI (IAI) addresses this problem: generating models that are easy to understand for humans and, consequently, trustworthy. While several approaches apply IAI techniques to reinforcement learning problems, addressing the case in which an agent has to act in a continuous action space is still an open question. In this work, we propose a cooperative co-evolutionary approach based on grammatical evolution, Q-learning, and the univariate marginal distribution algorithm, specifically designed to train IAI agents (in the form of binary decision trees) capable of acting in environments with continuous action spaces. The experimental results show that our method is able to solve two well-known OpenAI Gym test cases reaching state-of-the-art performance. Moreover, a quantitative post hoc analysis reveals that the obtained solutions are more interpretable than those reported in the literature.File | Dimensione | Formato | |
---|---|---|---|
Continuous_action_interpretable_reinforcement_learning_with_decision_trees.pdf
Solo gestori archivio
Tipologia:
Post-print referato (Refereed author’s manuscript)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
397.91 kB
Formato
Adobe PDF
|
397.91 kB | Adobe PDF | Visualizza/Apri |
A_co-evolutionary_approach_to_interpretable_reinforcement_learning_in_environments_with_continuous_action_spaces.pdf
Solo gestori archivio
Tipologia:
Versione editoriale (Publisher’s layout)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
229.41 kB
Formato
Adobe PDF
|
229.41 kB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione