Author of the publication

No-Regret Exploration in Goal-Oriented Reinforcement Learning.

, , , , and . ICML, volume 119 of Proceedings of Machine Learning Research, page 9428-9437. PMLR, (2020)

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed. You can also use the button next to the name to display some publications already assigned to the person.

 

Other publications of authors with the same name

A Unified Framework for Conservative Exploration., , , , , , , and . CoRR, (2021)Adversarial Attacks on Linear Contextual Bandits., , , , , , and . NeurIPS, (2020)No-Regret Exploration in Goal-Oriented Reinforcement Learning., , , , and . CoRR, (2019)Bandits with Side Observations: Bounded vs. Logarithmic Regret., , and . UAI, page 467-476. AUAI Press, (2018)Privacy Amplification via Shuffling for Linear Contextual Bandits., , , and . ALT, volume 167 of Proceedings of Machine Learning Research, page 381-407. PMLR, (2022)A Reduction-Based Framework for Conservative Bandits and Reinforcement Learning., , , , , , , and . ICLR, OpenReview.net, (2022)Local Differential Privacy for Regret Minimization in Reinforcement Learning., , , and . NeurIPS, page 10561-10573. (2021)Improved Algorithms for Conservative Exploration in Bandits., , , and . AAAI, page 3962-3969. AAAI Press, (2020)Top K Ranking for Multi-Armed Bandit with Noisy Evaluations., , , and . AISTATS, volume 151 of Proceedings of Machine Learning Research, page 6242-6269. PMLR, (2022)No-Regret Exploration in Goal-Oriented Reinforcement Learning., , , , and . ICML, volume 119 of Proceedings of Machine Learning Research, page 9428-9437. PMLR, (2020)