1. 9.5 Large Language Models as Generalizable Policies for Embodied Tasks
  2. 9.1 Lifting the Veil: Unlocking the Power of Depth in Q-learning
  3. 9.0 Train Once, Get a Family: State-Adaptive Balances for Offline-to-Online Reinforcement Learning
  4. 8.9 Understanding when Dynamics-Invariant Data Augmentations Benefit Model-Free Reinforcement Learning Updates
  5. 8.9 Ask more, know better: Reinforce-Learned Prompt Questions for Decision Making with Large Language Models
  6. 8.7 Reward Scale Robustness for Proximal Policy Optimization via DreamerV3 Tricks
  7. 8.7 Improving Intrinsic Exploration by Creating Stationary Objectives
  8. 8.6 Model-free Posterior Sampling via Learning Rate Randomization
  9. 8.3 Transfer of Reinforcement Learning-Based Controllers from Model- to Hardware-in-the-Loop
  10. 8.1 Function Space Bayesian Pseudocoreset for Bayesian Neural Networks