NeSyDebates: Neuro-Symbolic Debates for Safeguarded Generative AI
Role: Co-Investigator (Co-I) Funder: EPSRC (UK) & Japan Science and Technology Agency (JST) — ASPIRE Japan–UK Joint Programme on Advancing Human-Centered AI Period: January 2026 – March 2031 (63 months) UK PI: Prof. Francesca Toni, Department of Computing, Imperial College London Japan PI: Prof. Ken Satoh, Center for Juris-Informatics, Joint Support-Center for Data Science Research, Research Organization of Information and Systems
Overview
With the rapid advancement of generative AI models, challenges remain regarding their safety, security, and alignment with human values. Large language models (LLMs) face risks such as hallucinations that produce false information and prompt misuse that may lead to the unlawful acquisition of personal data.
NeSyDebates develops a Neuro-Symbolic Debate System that:
- Automatically extracts machine-readable normative argument structures from natural-language descriptions of norm violations — covering policies, regulations, and laws.
- Applies these extracted normative arguments to new cases to detect, explain, and prevent potential violations.
The system is validated in two application domains: legal document processing with LLMs, and text-to-image generation AI used in online image creation.
Our Contributions
Within NeSyDebates, our group at the APSS Lab focuses on:
- Characterizing and explaining model behavior — understanding when and why generative models produce policy-violating outputs.
- Robust, interpretable neuro-symbolic methods — designing enforceable safeguards grounded in policies, regulations, and laws.
- Machine unlearning — developing approaches to correct harmful outputs by selectively removing offending knowledge from trained models.
- Evaluation against real-world constraints — benchmarking methods against real policies and legal frameworks.
Context
NeSyDebates was selected as one of four funded projects out of 85 proposals submitted to the joint JST–EPSRC ASPIRE call. The project brings together expertise in computational logic, argumentation, multi-agent systems, AI safety, and law.