TL;DR submit an interpretable policy to address a control problem and win
Submissions closed, thanks to everyone who participated! We’ll notify winners at Gecco.
Check out the competition details below and be sure to join our Discord server https://discord.gg/dA8jpFVa9t for questions or news
Control systems play a vital role in managing and regulating various processes and devices across diverse application domains. Their pervasive nature makes them a cornerstone of modern technology. Safety-critical applications, in particular, demand control systems that are not only efficient but also interpretable, ensuring trustworthiness, reliability, and accountability. However, a prevalent issue in the field is the over-reliance on opaque systems, such as deep neural networks, known for their efficiency and optimization potential. This preference is rooted in the prevailing belief that interpretability is of secondary importance, with performance taking precedence. Furthermore, the scarcity of objective metrics to assess the degree of interpretability in a system exacerbates this problem. In fact, though the Evolutionary Computation (EC) community is starting to promote explainable and/or interpretable AI, significant challenges still persist in achieving comprehensive solutions. The goal of this competition is thus to ignite the research domain of interpretable control, with two specific goals in mind. First, we want to create a basis of comparison for different techniques emphasizing the trade-offs between performance and interpretability. Second, through the involvement of a panel of human evaluators, we strive to uncover the key characteristics that enhance the interpretability of control policies, making them more accessible to the general user.
We have chosen two tasks: Walker2D for continuous control, and 2048 for discrete control. We provide details and examples for both tasks in the competition repository. For both tasks we set a limit of 200000 episodes for optimizing the policies. An episode is a simulation of 1000 steps for Walker2D and a game for 2048.
Participants can take part in either or both of the tracks. Participants will have the freedom to apply their preferred methods to generate and interpret policies that effectively address the proposed task. However, we promote the inclusion of EC techniques into either the policy generation or explanation process, as a valuable component of addressing the proposed task.
Each submission will have to include the following:
All submission files will have to be uploaded to a github repository. The submission will then have to be made through the following form: https://forms.gle/bUKLKmKRGQ1niNZm7
If you want to ask any question or provide some feedback, join us at our Discord server: https://discord.gg/dA8jpFVa9t.
20th June 2024 AoE.
13th June 2024 AoE.
The final score for each entry will be determined by combining the two following terms.
These two ranks will be combined using the geometric mean to compute the overall global rank for each competition entry.
Our objective is to raise awareness regarding the importance of interpretability within the realm of control systems. To achieve this, we aim at collecting a wide variety of methodologies and publishing the results in a comprehensive report. We also plan to extend an invitation to select participants to become co-authors of this publication.
The winner(s) will be awarded a certificate and a cash prize (totaling 1000€) sponsored by Aindo.
Giorgia Nadizar, University of Trieste, giorgia.nadizar@phd.units.it
Giorgia Nadizar is a third year PhD student at the University of Trieste, Italy.
Her research interests lie at the intersection of embodied AI and explainable/interpretable AI.
Luigi Rovito, University of Trieste, luigi.rovito@phd.units.it
Luigi Rovito is a third year PhD student at the University of Trieste, Italy.
His research interests are genetic programming for cryptography and interpretable ML.
Dennis G. Wilson, ISAE-SUPAERO, University of Toulouse, dennis.wilson@isae.fr
Dennis G. Wilson is an Associate Professor at ISAE-Supaero in Toulouse, France.
They research evolutionary algorithms, deep learning, and applications of AI to climate problems.
Eric Medvet, University of Trieste, emedvet@units.it
Eric Medvet is an Associate Professor at the University of Trieste, Italy.
His research interests include embodied AI, artificial life, and evolutionary optimization.