The AI monitoring assistant exposed a scaling gap in the monitoring tool.
Users weren’t scanning dashboards anymore.
They were asking direct questions.
The system had data. It lacked decision logic to prioritize what mattered.
Before (Dashboard)
Multi-signal visual exploration for expert users
Behavior Shift (Chat)
Became the entry point for non-expert reps
System (What Broke)
Results varied across similar queries
Large queries hit compute limits
+ New users needed monitoring insights before they explored dashboards.
What changed in monitoring chat
Background
AI Chat for an Agriculture Monitoring Platform - conversational interface embedded in the monitoring tool
Narrow decision windows
Weather & disease uncertainty
Field-level variability
Financial stakes and Crop loss
User
One sales representative
manages 50+ farmers making decisions across 30,000+ acres.
This tool enables earlier action.
Reducing crop loss by up to 30%.
Decision logic turned chat into a reliable decision layer.
Deterministic logic was introduced so AI responses could be trusted.
Fewer clarification questions
Users needed fewer follow-ups to act on monitoring insights.
Adopted beyond the original feature
The prioritization logic was adopted by another internal tool.
+ This required system-level decisions, not just better prompts
The strategic shift
Three pressures shaped the problem space.
User needs, infrastructure limits, and AI trust requirements all pulled in different directions.
The real issue wasn’t missing insights. It was the structure between signals and responses.
+ The issue wasn’t missing data. It was missing structure.
The diagnosis
Reframing the problem as system-level pressures
User Clarity Pressure
Chat became the entry point for many new users.
Typical questions:
+ What should I focus on first today?
+ Where is precipitation behind?
Users needed direct answers. Dashboards weren’t the right surface.
Infrastructure Constraint Pressure
Large territory queries stressed the system.
+ Compute truncation
+ Aggregation mismatch across hierarchy levels
Scope was constrained for v1 due to engineering limitations.
AI Trust Pressure
Early responses revealed reliability problems.
+ Inconsistent signal prioritization
+ Hallucinated insights
Leadership escalated the issue to align on safeguards.
The model was deciding things it shouldn’t.
Without ordering logic, priorities changed across similar queries, making responses unreliable.
The system needed a decision layer before an AI layer.
+ The design problem became a ranking problem.
The decision model
Moving decision logic out of the model
Monitoring Signals
GDU, precipitation, growth stage, diseases
Decision Layer
Ranked operations, prioritized fields using decision logic
AI Chat Response
Prevalence, Count, Severity
Scenario A - Isolated Severity
1 severe field
Scenario B - Widespread Exposure
7 moderately elevated fields
Before
Scenario A ranked first
LLM ranked based on the most dramatic signal.
After
Scenario B ranked first
Ranking Logic →
Prevalence
% of fields affected within an operation
Count
Total fields affected
Severity
Intensity of deviation within a field
Scenario A
Low (1/16 fields)
Low (1 field)
High (severe spike)
Scenario B
High (7/16 fields)
High (7 fields)
Medium (moderate deviation)
Why it matters for users -
Widespread impact signals an emerging operation-level problem.
It increases the risk of coordinated crop stress and yield loss, where intervention becomes necessary—not just monitoring.
With limited time, reps prioritize actions that change outcomes at scale.
Deterministic ranking defined what matters first.
+ Exposure outweighs anomaly when prioritizing risk at scale.
Designing under constraints
Signal Reliability
Model outputs varied across regions, making ranking inconsistent.
Data science
Compute limits
Large-scale queries frequently failed with “too much to compute.”
System architecture
New user entry
New users didn’t know what to look for, so they asked generic questions.
User behavior
Leadership direction
Hallucinations raised concerns about trust in AI outputs.
Strategic alignment
Reusable AI pattern
The approach needed to scale beyond a single tool.
Platform thinking
Shipping Reality
The system had to fit existing pipelines and timelines.
Delivery constraints
Translating pressures into system decisions
+ We didn’t design a response. We defined how decisions are made.
View execution
Decision Architecture for Monitoring AI
Query Layer
Shapes the request before computation
Query + Filters: field / operation / agency, growth stage
Decision Layer
Turns signals into ranked priorities
Field-Level Data: Planting date, Weather
Deterministic Classification: Assigns Elevated / Stable / Insufficient states
Operation Ranking: Prevalence → Count → Severity
Response Layer
Turns decisions into explainable output
Structured Response: Formats overview, rankings, signals, top fields
LLM Explanation
User: Which operations need attention?
Before

After

+ Ranked operations with structured signal breakdown.
+ Decision-ready output without additional follow-ups.
Separating decision logic from AI behavior
+ These decisions changed how users reached and acted on insights.
See the impact
Behavior changed. Decisions got faster.
Measured through interaction patterns and post-launch usage shifts.
Fewer follow-ups per query
Users reached decisions with fewer prompts
Most monitoring queries resolved in a single response
Increased monitoring usage in chat
Shift from product questions → decision queries
More users engaging with monitoring insights
+ The system didn’t change the data. It changed how users acted on it.
How this scaled across tools
Where I Added Leverage
Deterministic decision engine
Moved ranking out of the LLM and defined evaluation order for stable, defensible outputs
Replaced anomaly-first ranking with how farming decisions actually scale.
AI reliability & guardrails
Surfaced hallucinations to leadership and restricted responses to verified signals.
Defined compute-aware outputs using “top fields + operations” instead of full datasets.
Negotiated model scope
Defined “good output” with PM, Eng, Data, and Commercial
Pushed back on some data science models due to inconsistent accuracy and to ship weather-first.
Scoped features by feasibility, removing non-critical signals for a stable first release
New user adoption strategy
Shifted output from information → prioritized action (chat as primary decision layer for new users)
Across-verticals alignment
Defined reusable response structure across hierarchy levels enabling cross-team adoption
A reusable decision layer for AI products
Each tool had different needs. The system adapted without redesign.
Designing for AI isn’t about generating answers.
It’s about structuring decisions.
+ The structure stayed constant. The logic adapted.
Query Layer
Decision Layer
Response Layer
Monitoring (Our tool)
Growth stage
View level
Weather check
Risk priority
Overview
Key fields
Signal drivers
Fungicide (closely matching tool)
Crop type
Timing window
Disease risk
Action timing
Next action
Risk reason
Products (different suite)
Product mix
Field group
Not needed
Plan output
Expected impact
+2 tools till now
Custom filters
Custom logic
Structured output


