Papers I Read¶
约 129 个字 预计阅读时间不到 1 分钟
Table of Contents¶
- 「KDD'2025」Achieving Nearly-Optimal Regret and Sample Complexity in Dueling Bandits with Applications in Online Recommendations
- 「JMLR'2006」Action Elimination and Stopping Conditions for the Multi-Armed Bandit and Reinforcement Learning Problems
- 「ICML'2017」Dueling Bandits with Weak Regret
- 「ICML'2011」Beat the Mean Bandit
- 「ICML'2017」Maximum Selection and Ranking under Noisy Comparisons
- Welcome to the Era of Experience
原文目录
- 「KDD'2025」Achieving Nearly-Optimal Regret and Sample Complexity in Dueling Bandits with Applications in Online Recommendations
- 「JMLR'2006」Action Elimination and Stopping Conditions for the Multi-Armed Bandit and Reinforcement Learning Problems
- 「ICML'2017」Dueling Bandits with Weak Regret
- 「ICML'2011」Beat the Mean Bandit
- 「ICML'2017」Maximum Selection and Ranking under Noisy Comparisons
- Welcome to the Era of Experience