- 9.4 Probabilistic Inference in Reinforcement Learning Done Right
- Authors: Jean Tarbouriech, Tor Lattimore, Brendan O’Donoghue
- Reason: This paper confronts a fundamental issue in the probability modeling of reinforcement learning, offering a novel solution with tractable algorithms. The involvement of an author like Brendan O’Donoghue, with his significant contributions to the field, further underscores its potential impact. Presented at NeurIPS, a leading conference, it indicates a high likelihood of influencing future research and practices.
- 9.2 From Images to Connections: Can DQN with GNNs learn the Strategic Game of Hex?
- Authors: Yannik Keller, Jannis Blüml, Gopika Sudhakaran, Kristian Kersting
- Reason: Investigating the application of graph neural networks (GNNs) in the context of reinforcement learning in board games, which typically require understanding complex relational structures, this paper can significantly influence strategies for similar problems. Kristian Kersting is known for his work in machine learning, and the paper’s focus on a strategic shift in RL through the use of GNNs makes it remarkably influential.
- 8.9 Nav-Q: Quantum Deep Reinforcement Learning for Collision-Free Navigation of Self-Driving Cars
- Authors: Akash Sinha, Antonio Macaluso, Matthias Klusch
- Reason: This paper pioneers the integration of quantum computation into deep reinforcement learning for self-driving cars, which could significantly accelerate training performance and improve policy quality, potentially impacting both the fields of quantum computing and autonomous driving.
- 8.9 Combinatorial Optimization with Policy Adaptation using Latent Space Search
- Authors: Felix Chalumeau, Shikha Surana, Clement Bonnet, Nathan Grinsztajn, Arnu Pretorius, Alexandre Laterre, Thomas D. Barrett
- Reason: This research delves into a novel RL approach for combinatorial optimization and demonstrates superior generalization and performance across a range of problems. The influence is bolstered by its impressive benchmarks against state-of-the-art approaches and the overall quality of the contributing authors.
- 8.7 InteRACT: Transformer Models for Human Intent Prediction Conditioned on Robot Actions
- Authors: Kushal Kedia, Atiksh Bhardwaj, Prithwish Dan, Sanjiban Choudhury
- Reason: The proposal of a novel architecture that conditions human intent prediction models on robot actions is innovative, and the use of transfer learning from human-human to human-robot interactions could be highly influential in collaborative robot systems.
- 8.7 Risk-sensitive Markov Decision Process and Learning under General Utility Functions
- Authors: Zhengqi Wu, Renyuan Xu
- Reason: Addressing risk-sensitive decision-making in RL with general utility functions, this paper covers an area of growing importance in practical applications. Though it may be a little more niche than the other papers, its mathematical rigor and applicability to real-world problems like portfolio management warrant a high importance score.
- 8.5 Learning to Fly in Seconds
- Authors: Jonas Eschmann, Dario Albani, Giuseppe Loianno
- Reason: Presents breakthrough training speeds for deep RL in quadrotor control and shows promise for practical applications, with demonstrated success in both simulation and real-world deployment, which is crucial for the adoption of RL in physical systems.
- 8.5 Stable Unlearnable Example: Enhancing the Robustness of Unlearnable Examples via Stable Error-Minimizing Noise
- Authors: Yixin Liu, Kaidi Xu, Xun Chen, Lichao Sun
- Reason: While not strictly about reinforcement learning, this paper impacts the broader field of machine learning with its focus on data privacy and robustness, which are critical in today’s data-rich environments. It offers a state-of-the-art solution to enhance data resistance to unauthorized model training. The paper boasts extensive experiments and tangible performance improvements, asserting its potential influence.
- 8.3 DroneOptiNet: A Framework for Optimal Drone-based Load Redistribution Mechanism for 5G and Beyond Solar Small Cell Networks
- Authors: Daksh Dave, Vinay Chamola, Sandeep Joshi, Sherali Zeadally
- Reason: Addresses a topical issue in 5G networks and presents a novel drone-based mechanism for load balancing, combining optimization and machine learning, with significant implications for next-generation wireless systems.
- 8.2 Clustered Policy Decision Ranking
- Authors: Mark Levin, Hana Chockler
- Reason: Offers new insights into understanding complex reinforcement learning policies by introducing a method for clustering and ranking decisions according to their importance, which is valuable for model interpretability and diagnostics.