Skip to main content

Task-Tailored Schema Agents for Reliable LLM Automation

Supervisor

Andrey Kravchenko

Suitable for

MSc in Advanced Computer Science
Computer Science, Part B
Computer Science, Part C

Abstract

Supervised by Andrey Kravchenko andrey.kravchenko@cs.ox.ac.uk

Expected background of students and CS techniques that will be applied.

The student will know how to work or be able to quickly learn how to work with Pytorch, Hugging Face, IPython notebooks. They should also have a solid mathematical base.

Task-Tailored Schema Agents for Reliable LLM Automation

Design an agent that, given a task description and a small sample of target data, automatically proposes a fit-for-purpose schema and compiles a robust extraction/automation pipeline, then validates, executes, and self-improves it. The agent will (i) induce a schema (fields, types, constraints, relations) from task instructions and examples; (ii) compile this schema into a structured-output program with constrained decoding; and (iii) verify—and, when available, link—outputs against a reference ontology or database.  

To harden reliability, add an agentic refinement loop (Self-Refine; Reflexion): when coverage or validity drops, the agent critiques its outputs and edits the schema before re-running. Implement the workflow in DSPy, allowing the compiler to tune prompts, demonstrations, and tool ordering to the target metric (e.g., F1 × validity × latency). Evaluate across 2–3 domains (e.g., scientific IE, event extraction) using schema coverage/minimality, per-field F1, JSON/ontology validity, and robustness to small task shifts.

Outcomes: (1) an open-source agent that induces, enforces, and verifies task-specific schemas from a handful of examples; (2) measurable gains in structural validity and downstream F1 from combining constrained decoding with ontology-based verification; and (3) ablations, an error taxonomy, and a small reusable schema library plus tuned DSPy pipelines.