1. 9.5 Unexpected Improvements to Expected Improvement for Bayesian Optimization
  2. 9.3 A Kernel Perspective on Behavioural Metrics for Markov Decision Processes
  3. 9.3 Offline RL with Observation Histories: Analyzing and Improving Sample Complexity
  4. 9.0 SERA: Sample Efficient Reward Augmentation in offline-to-online Reinforcement Learning
  5. 8.9 Autonomous Robotic Reinforcement Learning with Asynchronous Human Feedback
  6. 8.7 Improved Communication Efficiency in Federated Natural Policy Gradient via ADMM-based Gradient Updates
  7. 8.7 Sample-Efficient and Safe Deep Reinforcement Learning via Reset Deep Ensemble Agents
  8. 8.5 Handover Protocol Learning for LEO Satellite Networks: Access Delay and Collision Minimization
  9. 8.3 Efficient Exploration in Continuous-time Model-based Reinforcement Learning
  10. 8.1 Posterior Sampling for Competitive RL: Function Approximation and Partial Observation