- 8.9 Decomposing Control Lyapunov Functions for Efficient Reinforcement Learning
- Authors: Antonio Lopez, David Fridovich-Keil
- Reason: This paper addresses a significant real-world problem of RL’s high data requirements in robotics, proposes a novel method for shaping reward functions using DCLFs, and seems to be the first to tackle the computation of DCLFs for high-dimensional systems.
- 8.9 Policy Bifurcation in Safe Reinforcement Learning
- Authors: Wenjun Zou, Yao Lv, Jie Li, Yujie Yang, Shengbo Eben Li, Jingliang Duan, Xianyuan Zhan, Jingjing Liu, Yaqin Zhang, Keqiang Li
- Reason: Introduces a novel phenomenon in safe RL, policy bifurcation, with thorough theoretical analysis and a new algorithm, which indicates substantial potential influence on the domain of safe RL.
- 8.7 Reinforcement Learning from Delayed Observations via World Models
- Authors: Armin Karamzade, Kyungmin Kim, Montek Kalsi, Roy Fox
- Reason: Offers a novel approach to deal with delayed observations in RL, leveraging world models, which could be very influential given how common delayed feedback is in real-world applications of RL.
- 8.7 On Safety in Safe Bayesian Optimization
- Authors: Christian Fiedler, Johanna Menn, Lukas Kreisköther, Sebastian Trimpe
- Reason: Addresses critical safety issues in Safe Bayesian Optimization, introduces Real-beta-SafeOpt, and proposes solutions that can strengthen the real-world application of BO under safety constraints.
- 8.5 Improving LoRA in Privacy-preserving Federated Learning
- Authors: Youbang Sun, Zitao Li, Yaliang Li, Bolin Ding
- Reason: Published at a recognized conference (ICLR 2024), addresses the important area of privacy-preserving FL with empirical demonstrations, and improves upon a popular method (LoRA) for federated learning which is of high current interest.
- 8.5 Yell At Your Robot: Improving On-the-Fly from Language Corrections
- Authors: Lucy Xiaoyang Shi, Zheyuan Hu, Tony Z. Zhao, Archit Sharma, Karl Pertsch, Jianlan Luo, Sergey Levine, Chelsea Finn
- Reason: Explores interactive feedback for hierarchical policies in robotic tasks, suggesting the potential for significant impact on the field of robotics and reinforcement learning.
- 8.3 Stochastic Halpern iteration in normed spaces and applications to reinforcement learning
- Authors: Mario Bravo, Juan Pablo Contreras
- Reason: Proposes a method that surpasses the current state-of-the-art in terms of oracle complexity, which is a core challenge in RL, and applies this to MDPs showing potential for practical impact on RL algorithms.
- 8.3 D-Cubed: Latent Diffusion Trajectory Optimisation for Dexterous Deformable Manipulation
- Authors: Jun Yamada, Shaohong Zhong, Jack Collins, Ingmar Posner
- Reason: Proposes a new trajectory optimization method for dexterous manipulation, leveraging latent diffusion models, which may influence methods for robotics and RL in complex manipulation tasks.
- 8.1 Offline Imitation of Badminton Player Behavior via Experiential Contexts and Brownian Motion
- Authors: Kuang-Da Wang, Wei-Yao Wang, Ping-Chun Hsieh, Wen-Chih Peng
- Reason: Focuses on the niche but innovative task of modelling badminton player behavior which could open up new avenues in sports analytics and behavior modelling using hierarchical offline imitation learning.
- 8.1 Sample Complexity of Offline Distributionally Robust Linear Markov Decision Processes
- Authors: He Wang, Laixi Shi, Yuejie Chi
- Reason: Discusses sample-efficient distributionally robust policy learning in offline RL, which is crucial for practical applications, indicating considerable relevance in the area of offline RL.