https://www.harmdevries.com/post/model-size-vs-compute-overhead/ Go smol or go homeWhy we should train smaller LLMs on more tokensHarm de VriesApr 13, 2023 “However, for most use cases you should not train a compute-optimal LLM
What I Read: LLM applications, production
https://huyenchip.com//2023/04/11/llm-engineering.html Building LLM applications for productionChip HuyenApr 11, 2023 “It’s easy to make something cool with LLMs, but very hard to make something production-ready with them.”
What I Read: human touch, LLMs
https://mewelch.substack.com/p/putting-the-human-touch-on-llms Putting the human touch on LLMsMolly WelchMar 30 “Techniques like RLHF help align large language models with people’s values and preferences. Is that a good thing?”
What I Read: GPT, Ranking
https://messyprogress.substack.com/p/gpt-is-rather-good-at-feed-ranking GPT is Rather Good at Feed RankingRob EnnalsMar 7 “If ranking is as easy as saying what should rank highly, then lots of interesting things happen.”