User preference for models fine-tuned on the moral graph
Determine whether users prefer interacting with a language model fine-tuned on the moral graph alignment target, as compared to their current interactions with existing systems.
References
Finally, we don’t yet know if users will prefer interacting with a model fine-tuned on the moral graph. We are in the process of fine-tuning a model on a new, larger moral graph, and will be able to answer this question soon.
— What are human values, and how do we align AI to them?
(2404.10636 - Klingefjord et al., 27 Mar 2024) in Subsection “Limitations” (Fine-Tuning), Section Discussion