The benefits, risks and bounds of personalizing the alignment of large language models to individuals

Large language models (LLMs) undergo ‘alignment’ so that they better reflect human values or preferences, and are safer or more useful. However, alignment is intrinsically difficult because the hundreds of millions of people who now interact with LLMs have different preferences for language and conv...

Πλήρης περιγραφή

Λεπτομέρειες βιβλιογραφικής εγγραφής
Κύριοι συγγραφείς: Kirk, HR, Vidgen, B, Röttger, P, Hale, SA
Μορφή: Journal article
Γλώσσα:English
Έκδοση: Springer Nature 2024

Παρόμοια τεκμήρια