Dimitri Coelho Mollo<p>Another of my forays into AI ethics is just out! This time the focus is on the ethics (or lack thereof) of Reinforcement Learning Feedback (RLF) techniques aimed at increasing the 'alignment' of LLMs.</p><p>The paper is fruit of the joint work of a great team of collaborators, among whom <span class="h-card" translate="no"><a href="https://social.accum.se/@pettter" class="u-url mention" rel="nofollow noopener" target="_blank">@<span>pettter</span></a></span> and <span class="h-card" translate="no"><a href="https://akademienl.social/@roeldobbe" class="u-url mention" rel="nofollow noopener" target="_blank">@<span>roeldobbe</span></a></span>.</p><p><a href="https://link.springer.com/article/10.1007/s10676-025-09837-2" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">link.springer.com/article/10.1</span><span class="invisible">007/s10676-025-09837-2</span></a></p><p>1/</p><p><a href="https://social.sunet.se/tags/aiethics" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>aiethics</span></a> <a href="https://social.sunet.se/tags/LLMs" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>LLMs</span></a> <a href="https://social.sunet.se/tags/rlhf" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>rlhf</span></a> <a href="https://social.sunet.se/tags/llmsafety" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>llmsafety</span></a></p>