- 8.9 Decentralised Q-Learning for Multi-Agent Markov Decision Processes with a Satisfiability Criterion
- Authors: Keshav P. Keval, Vivek S. Borkar
- Reason: Proposed a novel reinforcement learning algorithm with potential applications and contribution to multi-agent systems.
- 8.7 Provable Representation with Efficient Planning for Partially Observable Reinforcement Learning
- Authors: Hongming Zhang, Tongzheng Ren, Chenjun Xiao, Dale Schuurmans, Bo Dai
- Reason: The methodology addresses a fundamental challenge in reinforcement learning (partially observable states), and is built upon a solid theoretical foundation with empirical demonstrations. Contributions from authors affiliated with recognizable institutions, and the inclusion of a well-known researcher in the field, Dale Schuurmans, suggest strong potential influence.
- 8.7 Differentiable Sampling of Categorical Distributions Using the CatLog-Derivative Trick
- Authors: Lennert De Smet, Emanuele Sansone, Pedro Zuidberg Dos Martires
- Reason: Presents a significant improvement in the field of gradient estimation for discrete distributions, which could advance the efficiency of REINFORCE and related algorithms in reinforcement learning.
- 8.5 Multi-Objective Reinforcement Learning based on Decomposition: A taxonomy and framework
- Authors: Florian Felten, El-Ghazali Talbi, Grégoire Danoy
- Reason: The paper proposes a taxonomy and framework in a niche but growing area of reinforcement learning, which could significantly influence further research. The potential publication in JAIR and the authoritative backgrounds of the authors contribute to the impact score.
- 8.5 Machine-Guided Discovery of a Real-World Rogue Wave Model
- Authors: Dion Häfner, Johannes Gemmrich, Markus Jochum
- Reason: The paper showcases the intersection of machine learning and physical sciences, potentially guiding future research in reinforcement learning applications for scientific discovery.
- 8.2 Koopman Learning with Episodic Memory
- Authors: William T. Redman, Dean Huang, Maria Fonoberova, Igor Mezić
- Reason: Integrates a key machine learning technique with a dynamical systems approach, potentially enhancing the learning capability of Koopman operator methods.
- 8.1 An efficient likelihood-free Bayesian inference method based on sequential neural posterior estimation
- Authors: Yifei Xiong, Xiliang Yang, Sanguo Zhang, Zhijian He
- Reason: The paper enhances an existing method in simulation-based models, key in certain reinforcement learning scenarios. The comprehensive methodological improvements backed by theoretical analysis and empirical evidence suggest a moderate level of prospective influence.
- 8.0 Improving Source-Free Target Adaptation with Vision Transformers Leveraging Domain Representation Images
- Authors: Gauransh Sawhney, Daksh Dave, Adeel Ahmed, Jiechao Gao, Khalid Saleem
- Reason: Introduces an innovative method to enhance Vision Transformer performance in unsupervised domain adaptation, which can indirectly affect reinforcement learning practices in domain generalization.
- 7.9 Summary of the DISPLACE Challenge 2023 – DIarization of SPeaker and LAnguage in Conversational Environments
- Authors: Shikha Baghel, Shreyas Ramoji, Somil Jain, Pratik Roy Chowdhuri, Prachi Singh, Deepu Vijayasenan, Sriram Ganapathy
- Reason: Challenges and competitions can propel advancement by benchmarking state-of-the-art methods, but this paper is more of a summary report than a direct contribution to reinforcement learning methodology, leading to a slightly lower potential influence score.
- 7.7 Explainable Anomaly Detection using Masked Latent Generative Modeling
- Authors: Daesoo Lee, Sara Malacarne, Erlend Aune
- Reason: While the focus is on anomaly detection, the explainability aspect of the proposed method could have trickle-down effects on reinforcement learning, especially in improving the understanding of agent behaviors. Yet, its indirect relation to core RL diminishes its impact rating.