https://www.cs.princeton.edu/~smalladi/blog/2024/07/09/dpo-infinity The Hidden Infinity in Preference LearningSadhika MalladiJuly 09 2024 “I demonstrate from first principles how offline preference learning algorithms (e.g., SimPO) can benefit from length normalization, especially when training
What I Read: AI Engineers, Search
https://softwaredoug.com/blog/2024/06/25/what-ai-engineers-need-to-know-search What AI Engineers Should Know about SearchDoug TurnbullJune 25th, 2024 “Things AI Engineers Should Know about Search”
What I Read: GPT, Ranking
https://messyprogress.substack.com/p/gpt-is-rather-good-at-feed-ranking GPT is Rather Good at Feed RankingRob EnnalsMar 7 “If ranking is as easy as saying what should rank highly, then lots of interesting things happen.”
What I Read: learning-to-rank
https://www.amazon.science/blog/using-learning-to-rank-to-precisely-locate-where-to-deliver-packages Using learning-to-rank to precisely locate where to deliver packagesModels adapted from information retrieval deal well with noisy GPS input and can leverage map information.By George FormanSeptember 15, 2021 “For