- 9.3 Attention-Driven Multi-Agent Reinforcement Learning: Enhancing Decisions with Expertise-Informed Tasks
- Authors: Andre R Kuroswiski, Annie S Wu, Angelo Passaro
- Reason: Innovates in MARL with the integration of attention-based policy mechanisms and domain knowledge which represents a significant step in reducing complexity and enhancing learning efficiency in cooperative multi-agent settings.
- 9.1 Learning Heuristics for Transit Network Design and Improvement with Deep Reinforcement Learning
- Authors: Andrew Holliday, Ahmed El-Geneidy, Gregory Dudek
- Reason: Represents a practical application of DRL in optimizing real-world transit networks, offering significant cost savings and performance improvements over existing networks, making this work directly influential on urban planning and resource management.
- 9.1 Diverse Randomized Value Functions: A Provably Pessimistic Approach for Offline Reinforcement Learning
- Authors: Xudong Yu, Chenjia Bai, Hongyi Guo, Changhong Wang, Zhen Wang
- Reason: Pioneering a novel strategy for Offline RL with theoretical and empirical validation.
- 8.9 Policy-Guided Diffusion
- Authors: Matthew Thomas Jackson, Michael Tryfan Matthews, Cong Lu, Benjamin Ellis, Shimon Whiteson, Jakob Foerster
- Reason: Innovative use of diffusion models in an offline RL setting with significant improvements demonstrated.
- 8.8 Computing Transition Pathways for the Study of Rare Events Using Deep Reinforcement Learning
- Authors: Bo Lin, Yangzheng Zhong, Weiqing Ren
- Reason: Addresses the challenging task of computing transition pathways in high-dimensional systems, a topic meaningful to various scientific fields, and successfully applies a DRL-based actor-critic method which could influence computational studies in physics, chemistry, and biology.
- 8.7 Generative Pre-Trained Transformer for Symbolic Regression Base In-Context Reinforcement Learning
- Authors: Yanjie Li, Weijun Li, Lina Yu, Min Wu, Jingyi Liu, Wenqiang Li, Meilan Hao, Shu Wei, Yusong Deng
- Reason: Introduces a novel integration of Transformer architecture with reinforcement learning for symbolic regression.
- 8.5 Parameter-Adaptive Approximate MPC: Tuning Neural-Network Controllers without Re-Training
- Authors: Henrik Hose, Alexander Gräfe, Sebastian Trimpe
- Reason: Contributes to the field of adaptive control with a novel architecture that allows online tuning without retraining, facilitating deployment in real-world systems with resource constraints, and showcases generalization which is commendable in the context of control systems.
- 8.5 Graph Reinforcement Learning for Combinatorial Optimization: A Survey and Unifying Perspective
- Authors: Victor-Alexandru Darvariu, Stephen Hailes, Mirco Musolesi
- Reason: Provides a comprehensive survey and offers a unifying perspective in the intersection of graph theory and RL, suggesting future research directions.
- 8.3 Adaptable Recovery Behaviors in Robotics: A Behavior Trees and Motion Generators(BTMG) Approach for Failure Management
- Authors: Faseeh Ahmad, Matthias Mayr, Sulthan Suresh-Fazeela, Volker Kreuger
- Reason: Addresses a practical robotics problem using RL for adaptable recovery behaviors with validated efficiency improvements.
- 8.2 Negative Preference Optimization: From Catastrophic Collapse to Effective Unlearning
- Authors: Ruiqi Zhang, Licong Lin, Yu Bai, Song Mei
- Reason: Tackles the essential challenge of unlearning in LLMs while maintaining model utilities, offering a balance between two vital aspects of machine learning, which could have a significant impact on privacy, security, and the functionality of large-scale AI systems.