SP
BravenNow
Design Behaviour Codes (DBCs): A Taxonomy-Driven Layered Governance Benchmark for Large Language Models
| USA | technology | βœ“ Verified - arxiv.org

Design Behaviour Codes (DBCs): A Taxonomy-Driven Layered Governance Benchmark for Large Language Models

#Design Behaviour Codes #DBCs #large language models #taxonomy-driven #layered governance #benchmark #AI regulation

πŸ“Œ Key Takeaways

  • Design Behaviour Codes (DBCs) are a new governance benchmark for large language models.
  • DBCs use a taxonomy-driven approach to categorize and manage model behaviors.
  • The framework implements layered governance for structured oversight of AI systems.
  • It aims to standardize evaluation and regulation of LLM conduct and outputs.

πŸ“– Full Retelling

arXiv:2603.04837v1 Announce Type: new Abstract: We introduce the Dynamic Behavioral Constraint (DBC) benchmark, the first empirical framework for evaluating the efficacy of a structured, 150-control behavioral governance layer, the MDBC (Madan DBC) system, applied at inference time to large language models (LLMs). Unlike training time alignment methods (RLHF, DPO) or post-hoc content moderation APIs, DBCs constitute a system prompt level governance layer that is model-agnostic, jurisdiction-map

🏷️ Themes

AI Governance, Model Benchmarking

πŸ“š Related People & Topics

Regulation of artificial intelligence

Guidelines and laws to regulate AI

Regulation of artificial intelligence is the development of public sector policies and laws for promoting and regulating artificial intelligence (AI). The regulatory and policy landscape for AI is an emerging issue in jurisdictions worldwide, including for international organizations without direct ...

View Profile β†’ Wikipedia β†—

Large language model

Type of machine learning model

A large language model (LLM) is a language model trained with self-supervised machine learning on a vast amount of text, designed for natural language processing tasks, especially language generation. The largest and most capable LLMs are generative pre-trained transformers (GPTs) that provide the c...

View Profile β†’ Wikipedia β†—

Entity Intersection Graph

Connections for Regulation of artificial intelligence:

🏒 Anthropic 14 shared
🏒 OpenAI 9 shared
🌐 Presidency of Donald Trump 7 shared
🌐 AI safety 6 shared
πŸ‘€ Donald Trump 4 shared
View full profile

Mentioned Entities

Regulation of artificial intelligence

Guidelines and laws to regulate AI

Large language model

Type of machine learning model

}
Original Source
--> Computer Science > Artificial Intelligence arXiv:2603.04837 [Submitted on 5 Mar 2026] Title: Design Behaviour Codes : A Taxonomy-Driven Layered Governance Benchmark for Large Language Models Authors: G. Madan Mohan , Veena Kiran Nambiar , Kiranmayee Janardhan View a PDF of the paper titled Design Behaviour Codes : A Taxonomy-Driven Layered Governance Benchmark for Large Language Models, by G. Madan Mohan and 2 other authors View PDF Abstract: We introduce the Dynamic Behavioral Constraint benchmark, the first empirical framework for evaluating the efficacy of a structured, 150-control behavioral governance layer, the MDBC (Madan DBC) system, applied at inference time to large language models . Unlike training time alignment methods (RLHF, DPO) or post-hoc content moderation APIs, DBCs constitute a system prompt level governance layer that is model-agnostic, jurisdiction-mappable, and auditable. We evaluate the DBC Framework across a 30 domain risk taxonomy organized into six clusters (Hallucination and Calibration, Bias and Fairness, Malicious Use, Privacy and Data Protection, Robustness and Reliability, and Misalignment Agency) using an agentic red-team protocol with five adversarial attack strategies (Direct, Roleplay, Few-Shot, Hypothetical, Authority Spoof) across 3 model families. Our three-arm controlled design (Base, Base plus Moderation, Base plus DBC) enables causal attribution of risk reduction. Key findings: the DBC layer reduces the aggregate Risk Exposure Rate from 7.19 percent to 4.55 percent (Base plus DBC), representing a 36.8 percent relative risk reduction, compared with 0.6 percent for a standard safety moderation prompt. MDBC Adherence Scores improve from 8.6 by 10 to 8.7 by 10 (Base plus DBC). EU AI Act compliance (automated scoring) reaches 8.5by 10 under the DBC layer. A three judge evaluation ensemble yields Fleiss kappa greater than 0.70 (substantial agreement), validating our automated pipeline. Cluster ablation identifies the Integrity...
Read full article at source

Source

arxiv.org

More from USA

News from Other Countries

πŸ‡¬πŸ‡§ United Kingdom

πŸ‡ΊπŸ‡¦ Ukraine