#Model Safety & Alignment
Latest news articles tagged with "Model Safety & Alignment". Follow the timeline of events, related topics, and entities.
Articles (1)
-
πΊπΈ ProMoral-Bench: Evaluating Prompting Strategies for Moral Reasoning and Safety in LLMs
[USA]
arXiv:2602.13274v1 Announce Type: new Abstract: Prompt design significantly impacts the moral competence and safety alignment of large language models (LLMs), yet empirical comparisons remain fragmen...
Related: #Artificial Intelligence Ethics, #Natural Language Processing Evaluation, #Robustness Testing in LLMs, #Prompt Engineering