https://lilianweng.github.io/posts/2024-11-28-reward-hacking Reward Hacking in Reinforcement LearningLilian WengNovember 28, 2024 “Reward hacking occurs when a reinforcement learning (RL) agent exploits flaws or ambiguities in the reward function to achieve high rewards,
What I Read: passively learned, causality
What can be passively learned about causality?Simons InstituteAndrew Lampinen (Google DeepMind)Jun 25, 2024 “What could language models learn about causality and experimentation from their passive training?”
What I Read: Contextual Bandit, LinUCB:
https://truetheta.io/concepts/reinforcement-learning/lin-ucb A Reliable Contextual Bandit Algorithm: LinUCBDJ RichAugust 6, 2024 “A user visits a news website. Which articles should they be shown?”
What I Read: Summarization, LLMs
https://cameronrwolfe.substack.com/p/summarization-and-the-evolution-of Summarization and the Evolution of LLMsCameron R. Wolfe, Ph.D.Jun 03, 2024 “How research on abstractive summarization changed language models forever…”