Hands-on: Build agents for claims processing & underwriting
Foundations
AI Agents vs Agentic AI
What's the difference and when does it matter?
From LLMs to Agents: Why the Shift?
LLMs alone have limitations that matter for insurance operations:
No memory β can't track claim state across interactions or remember policy history
No tools β can't query policy admin systems, call underwriting APIs, or update claims databases
No planning β can't break a claims adjudication into steps and execute them
Reliability drops as input size, task complexity, and output scope increase
The Fallacy of Composition: An LLM that writes great claims summaries can't independently manage an end-to-end claims adjudication. What works for individual parts won't necessarily work for the combined whole.
AI Agent vs Agentic AI
AI Agent
A single AI that can use tools to complete a specific task.
Follows a defined workflow
Uses tools (search, calculate, API calls)
Completes one task at a time
Human defines the steps
Example: A chatbot that looks up policy details and answers policyholder questions
Agentic AI System
Multiple AI agents working together autonomously to achieve a goal.
Plans its own approach
Breaks complex goals into subtasks
Coordinates multiple agents
Reflects and adjusts strategy
Example: A system that receives a claim, extracts medical data, validates against policy benefits, checks for fraud signals, and routes for approval β all autonomously
4 Types of AI Agents
Type
How it works
AnyCompany example
Workflow
Follows a predetermined path like a recipe. Fast, predictable.
Automated policy endorsement β same steps every time
Autonomous
Self-orchestrating with memory and tools. Handles complex, open-ended tasks.
Claims fraud investigation that adapts its approach based on evidence patterns
Hybrid
Routes between workflow and autonomous based on complexity.
Claims triage: simple claims auto-approved, complex claims escalated to adjuster
Multi-Agent
AI team β multiple agents collaborating on a shared goal.
Underwriting pipeline: medical + financial + compliance agents working in parallel
Key insight: Most teams start with workflow agents and graduate to hybrid or multi-agent as they gain confidence.
4 Levels of AI Autonomy
Level
Description
AnyCompany example
L1: Human-Assisted
AI drafts, human reviews and approves everything
AI drafts claims summary, adjuster reviews before settlement
L2: Semi-Autonomous
AI handles routine cases, flags exceptions for humans
Simple outpatient claims auto-approved, hospitalization claims flagged for review
L3: Autonomous
AI runs the full pipeline, alerts humans on exceptions only
Full underwriting pipeline with exception alerts for high-sum-assured cases
L4: Multi-Agent
Multiple AI agents collaborate and adapt dynamically
Dynamic claims processing that coordinates medical, financial, and fraud agents
Where to start: Most insurance teams should begin at L1-L2. Start with AI-assisted drafting, then graduate to semi-autonomous processing as you validate accuracy.
When to Use Which?
Scenario
Agent type
Autonomy
Why
Answer policyholder FAQs
Workflow
L2
Single task, defined scope
Classify incoming claims
Workflow
L2
One-step classification
Process health claim end-to-end
Hybrid
L2-L3
Multi-step with exception handling
Fraud investigation
Autonomous
L2
Adaptive analysis, high stakes
Life insurance underwriting
Hybrid
L2
Check docs, verify, assess, recommend
Cross-product risk monitoring
Multi-Agent
L3-L4
Coordinate Life + Health + General claims
Key Takeaway: AI Agents vs Agentic AI
AI Agent = one skilled worker following instructions. Good for single, well-defined tasks.
Agentic AI = a team that plans, executes, and adapts. Good for multi-step workflows.
Key insight: The loop is what makes agentic AI different from a simple prompt. It can recover from errors and adapt β like a team lead who checks work and adjusts the plan.
Key Takeaway: The Agentic Loop
Every agentic system follows: Observe → Plan → Act → Reflect → Repeat
The "Reflect" step is what makes it intelligent β it can catch its own mistakes and retry
This is the same self-critique technique used in prompt engineering, but automated
When designing workflows, ask: "What happens when a step fails?" β that's where reflection matters
The loop continues until the goal is met or a human is asked to intervene
Workflow Patterns
4 Patterns for Automation
Chaining, Parallelization, Routing, Orchestration
Why Workflow Patterns Matter
Your teams make thousands of complex decisions daily. Human cognitive load limits scalability.
Traditional Automation
Rigid if/then rules
Breaks when inputs vary
Can't handle ambiguity
Expensive to maintain
Agentic Automation
Understands context and intent
Adapts to variations in input
Handles ambiguity with reasoning
Improves with better prompts
The shift: From "program every rule" to "describe the goal and let AI figure out the steps." Well-crafted prompt templates are the instructions that power these patterns.
The 4 Workflow Patterns
Pattern
How it works
Insurance example
Prompt Chaining
Sequential steps β output of one becomes input of next
Claim → Extract medical data → Validate benefits → Generate settlement
Parallelization
Multiple tasks run simultaneously, results combined
Medical + Financial + Compliance review an underwriting application at the same time
Routing
Classify input, send to the right processing path
Classify claim type → route to life, health, motor, or travel workflow
Orchestration
Dynamic coordination β spawns subtasks based on conditions
Claims adjudication: auto-approve <$5K, AI + human for $5K-$50K, human review >$50K
4 layers, each doing one job: Steering sets the rules. Skills define the expertise. Hooks trigger automatically. MCP connects to your data. Together = an autonomous agent.
Steering & Skills: Your Agent's Brain
Steering (Global Rules)
Always-on instructions for every interaction:
Currency defaults (SGD unless specified)
PII handling rules (never expose NRIC or policy numbers)
Insurance terminology and brand voice
Compliance requirements (MAS, PDPA)
File: .kiro/steering/rules.md
Skills (On-Demand Expertise)
Specialized instructions for specific tasks:
Claims triage and adjudication
Underwriting risk assessment
Fraud investigation playbook
Policy servicing automation
File: .kiro/skills/[name]/SKILL.md
Key insight: Prompt templates become SKILL.md files. Same content, new format β now they auto-activate when the right task appears.
Hooks & MCP: Your Agent's Hands
Hooks (Auto-Triggers)
Events that automatically run your skills:
File created: New claim submitted β run triage skill
Databases: Query policy data, claims history, underwriting records
APIs: Call risk scoring engine, AML screening, medical bureau
Files: Process claim documents, read policy wordings
Messaging: Send Slack alerts, email notifications
Think of it this way: Skills = the recipe. MCP = the kitchen equipment. Hooks = the timer that starts cooking automatically. You design the recipe; your tech team sets up the kitchen.
SKILL.md: Best Practices
When converting a prompt template to a SKILL.md file:
Element
What to include
Example
Name
Clear, descriptive skill name
Claims Triage Agent
Description
When this skill should activate
"Activate when processing incoming insurance claims for triage"
Trigger
Keywords or file patterns
"claim", "FNOL", "adjudication"
Instructions
Your prompt template
Role, context, steps, output format, guardrails
Guardrails
What the skill must NOT do
"Never auto-approve claims >$50K", "Always flag pre-existing conditions"
Output format
Expected structure
JSON with claim_decision, confidence, reasoning, escalation_flag
Supervised vs Autopilot: Your Safety Net
Before we start building β here's how to stay safe when AI modifies your files.
π‘οΈ
Supervised Mode
Kiro shows every change before applying
You approve or reject each modification
Can revert individual files after approval
Like "track changes" in Word
Best for: first-time skills, high-stakes outputs, learning
vs
β‘
Autopilot Mode
Kiro works autonomously β no approval needed
Review all changes after completion
Revert any file you don't like
Sub-agents only work in Autopilot
Best for: trusted skills, routine tasks, parallelization
Nothing is irreversible. Both modes track every change. You can always revert individual files. When in doubt, use Supervised mode β you'll see exactly what the AI is about to do before it happens.
Connections
Model Context Protocol (MCP)
How AI connects to your databases, APIs, and tools
Why Does AI Need MCP?
Without MCP, AI can only work with text you paste into it. MCP changes that.
Without MCP
Export CSV β paste into chat β AI analyzes what you gave it
Open 50 invoices one by one
Check compliance in 3 systems manually
Copy transaction data, format it, paste it
With MCP
AI queries your merchant database directly
AI scans the invoice folder automatically
AI checks all 3 systems in one request
AI runs SQL query, gets exactly what it needs
Think of MCP as USB-C for AI. One standard cable connects any device to any peripheral. MCP connects any AI (Kiro, Claude, Cursor) to any data source (databases, APIs, files) with one standard protocol.
MCP Architecture
MCP follows a simple client-server pattern:
3 Core Primitives
Tools β Actions AI can perform: run SQL, call API, send message
Resources β Data AI can read: file contents, database schemas
Prompts β Reusable templates for common interactions
Key Properties
Open standard β works with Claude, Kiro, Cursor, VS Code
Local or remote β runs on your machine or on a server
Pre-built servers β community has built servers for most systems
MCP Servers for Insurance Teams
MCP Server
Connects to
Insurance use case
Setup
SQLite / PostgreSQL
Databases
Query claims data, policy records, underwriting history
Config only
Filesystem
File directories
Process folders of claim documents, scan medical reports
Config only
Slack / Teams
Team messaging
Send alerts when fraud detected or high-value claim submitted
API key
Google Drive / S3
Document storage
Read policy wordings for RAG grounding, store adjudication reports
OAuth / IAM
Custom REST API
Internal services
Call underwriting rules engine, AML screening, medical bureau
Custom build
You design the recipe, your tech team sets up the kitchen. In the Agent Design Canvas, you specify "MCP connections needed: merchant database, Slack for alerts." Your engineering team configures the actual MCP servers. The open-source community has pre-built servers for most common systems β no custom code needed.
MCP Configuration: It's Just JSON
An MCP configuration in Kiro β 7 lines of JSON, no code:
What happens: Kiro launches the server, discovers tools, and you can query the claims database in plain English.
Key Takeaway: MCP
MCP = USB-C for AI (one standard, any connection)
3 primitives: Tools, Resources, Prompts
Pre-built servers for databases, files, APIs, messaging
You specify needs β tech team configures servers
Try the interactive MCP explainer on the workshop site
Advanced
Sub-Agents: AI Delegating to AI
When one agent isn't enough β delegate subtasks to specialized agents
What Are Sub-Agents?
A sub-agent is an AI that gets delegated a task by another AI. The main agent stays focused on the big picture while sub-agents handle specialized subtasks independently.
π§
Main Agent
"Analyze this claim"
β
π
Context Gatherer
Find relevant files & data
β‘
Task Executor
Run analysis independently
π οΈ
Custom Agent
Specialized for your domain
β
π
Combined Result
Synthesized output
This is parallelization in action. The main agent delegates to 3 sub-agents, each runs independently with its own tools, and results are combined. Same pattern from the slides β but happening inside Kiro automatically.
Sub-Agents in Kiro Today
Type
Where it lives
How it activates
Finance example
Context Gatherer(Built-in)
Built into Kiro
Kiro uses it automatically when exploring unfamiliar code or files
"Find all files related to claims processing"
Task Executor(Built-in)
Built into Kiro
Kiro delegates automatically when you ask for parallel tasks
"Analyze this data while I work on the API"
Custom Agent(You create)
.kiro/agents/[name].md
You invoke by name: "use the claims-reviewer agent toβ¦"
"Review this claim and score fraud risk"
Built-in: Automatic
Kiro decides when to delegate β no config needed
Works in Autopilot mode
You can also ask explicitly: "do X and Y in parallel"
Custom: You Create & Invoke
Saved as .kiro/agents/name.md β like skills
Has its own tools (can restrict to read-only)
Version in git, share with team
Invoke: "use the [name] agent to review this file"
Skills vs Custom Agents
π
Skills (.kiro/skills/)
Instructions added to Kiro's context
Uses Kiro's own tools
Auto-activates on keywords/file patterns
Best for: structured output templates
Analogy: a recipe Kiro follows
vs
π€
Custom Agents (.kiro/agents/)
Runs as a separate autonomous agent
Has its own tools (can be restricted)
Invoked by name: "use the [name] agent"
Best for: review, analysis, validation
Analogy: a colleague Kiro delegates to
Try it in Exercise 1 (Bonus Step): Create a finance-reviewer custom agent, then invoke it to review your risk assessment. The agent runs independently, scores your work, and suggests improvements.
Hands-On
Exercises & Workshop
Convert templates to skills, then design your agent workflow
Lab: Build a Claims Triage Skill
Turn a prompt template into a reusable, auto-activating skill (25 min):
Step 1: Create a steering file with AnyCompany Insurance global rules (currency, PII, compliance)
Step 2: Convert a claims triage prompt template to SKILL.md format with frontmatter
Step 3: Add a hook for auto-triggering (e.g., "when new data file is created")
Step 4: Test the full stack β steering + skill + hook working together
What you'll have: A portable, reusable skill that auto-activates when the right task appears β a production-ready automation component.
Lab 10: Design Your Agent Workflow
Pick a real workflow from your team and fill out the Agent Design Canvas (45 min):
The Agent Design Canvas
Agent Role: What does this agent do?
Workflow Steps: What's the sequence?
Pattern: Chain / Parallel / Route / Orchestrate?
Data Inputs: What data does it need?
Data Outputs: What does it produce?
Tools/MCP: What systems does it connect to?
Guardrails: What must it NOT do?
Escalation Rules: When does a human step in?
Autonomy Level: L1 / L2 / L3?
Business Impact: Time saved, errors reduced?
Workshop: Design Your Workflow
In groups of 3-4:
Step 1: Identify (10 min)
What manual process takes the most time?
Where do errors happen most often?
What's repetitive but requires judgment?
Step 2: Map (15 min)
Draw the current process (boxes and arrows)
Label each step with a workflow pattern
Mark decision points and exception paths
Step 3: Assess (10 min)
Feasibility: Can current AI do this?
Impact: How much time/cost saved?
Risk: What if the AI gets it wrong?
Step 4: Present (5 min each)
Share your Agent Design Canvas
Get feedback and suggestions
Identify quick wins vs long-term projects
Key Takeaway: Designing AI Workflows
Best candidates for automation: repetitive, judgment-based, high-volume, error-prone
Production deployment with monitoring and observability
2-4 weeks
5. Scale
Expand to more workflows, add multi-agent coordination
Ongoing
Key benefit: Start fast with Kiro (your prototype), then harden with AgentCore for production. You don't need to choose upfront β the skills you build today are portable.