You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thanks for the excellent work! For this research direction, I think you maybe interested in work on using RL in practical applications where user ratings or preferences are used to improve models. New challenges arise when we deploy RL methods to these real-world scenarios (e.g., noisy reward, sample efficiency). Addressing these challenges is an important research direction.
Hi Authors,
Thanks for the excellent work! For this research direction, I think you maybe interested in work on using RL in practical applications where user ratings or preferences are used to improve models. New challenges arise when we deploy RL methods to these real-world scenarios (e.g., noisy reward, sample efficiency). Addressing these challenges is an important research direction.
Reliability and learnability of human bandit feedback for sequence-to-sequence reinforcement learning (https://arxiv.org/abs/1707.07402)
Reinforcement learning for bandit neural machine translation with simulated human feedback (https://arxiv.org/abs/1805.10627)
Learning from human feedback: https://deepmind.com/blog/learning-through-human-feedback/
The text was updated successfully, but these errors were encountered: