About ACQUIRE
Large Language Models are already writing, reviewing, and repairing code, yet the community lacks a rigorous, shared basis for judging whether these AI-produced artifacts are dependable throughout real software lifecycles. Today’s emphasis on benchmark accuracy obscures risks that matter in practice: silent hallucinations, insecure toolchains and prompts, brittle behavior under distribution shift, opaque provenance, and evidence that cannot be audited or reproduced. ACQUIRE’26 responds to this gap by convening AI and Software Engineering researchers and practitioners to refocus the conversation from raw performance to verifiable quality, grounded in auditable taxonomies and metrics, assurance cases and transparent and reproducible evidence. The workshop’s aim is to make AI-for-code not just powerful, but trustworthy and dependable, encompassing all aspects of software quality, including security, maintainability, correctness and performance.
We welcome contributions that address the following areas:
- Quality & assurance: taxonomies, auditable metrics, conformance profiles, and safety/assurance cases for LLMs and agents for code tasks;
- Security of models & supply chain: threat models spanning models, data, prompts, and toolchains, with provenance, SBOM/AI-BOM, signing, and attestation;
- Robustness in practice: hallucination detection/mitigation, shift- and fault-tolerance, vulnerability detection and patch quality, and CI/CD gating with runtime guards;
- Evidence & reproducibility: open benchmarks, standardized reporting for datasets, metrics, prompts, agents, and protocols, and certification-oriented evaluation.
Cross-cutting themes include human-AI collaboration (uncertainty display, attribution), and the impact of AI on maintainability and technical debt. We welcome empirical studies, methods, tools, and experience reports, especially those that deliver auditable evidence, align on taxonomies and reporting schemas, advance provenance and attestation practices, and demonstrate robust, reproducible evaluation under real-world and adversarial conditions.
Submissions reporting negative results or unexpected findings are also welcome, as they offer valuable insights.
ACQUIRE 2026 is co-located with EDCC 2026 — Canterbury, UK • 7-10 April 2026
Submit a Paper Registration Info
Deadlines in AoE Anywhere on Earth
Topics of Interest
This call for papers invites all researchers and practitioners to explore the quality and reliability aspects of AI in the SE field. The workshop will cover a wide range of topics, including but not limited to:
Foundations of Quality & Dependability for Code LLMs
- Taxonomies & reference models
- Measurement frameworks & methodology
- Assurance/safety cases for LLMs & agents
- Open benchmarks & leaderboards; standardized reporting (datasets, metrics, protocols)
Software Quality for AI-Generated Code
- Correctness & robustness; hallucination detection & mitigation
- Maintainability & technical-debt control
- Portability & compatibility
- Efficiency & performance
Security of the LLM Supply Chain
- Threat models (model/data/prompt/toolchain)
- Signing & attestation; SBOMs/AI-BOMs
- Dependency risk in plug-ins/RAG/agents
- Governance, audits & compliance
Attacks & Mitigations on LLMs
- Data poisoning
- Adversarial prompts & jailbreaks
- Prompt-injection & tool-use exploits
- Guardrails, policy-as-code, content filters, red teaming
- Attack/defense benchmarks
Human-AI Collaboration
- Comparative studies (correctness, speed, maintainability, security)
- Pair-programming & review workflows
- Explainability, attribution & uncertainty; bias & fairness
- Cognitive load & UX
Privacy, Licensing, Provenance & Integrity
- Privacy-preserving methods; leakage/memorization controls
- Licensing compliance
- Provenance & traceability at scale
- Legal, ethical & regulatory aspects
Green AI for Software Engineering
- Energy/compute profiling
- Efficiency-aware prompting & inference
- Carbon-aware deployment
- Quality-cost-environment trade-offs
Vulnerability Detection & Patching
- Secure-by-construction patterns
- Ground-truth benchmarks
- Patch quality, regressions & assurance