Natural Language OEE Query — Ask 'Why Is Plate Bay Slow?'

By lamine yamal on May 8, 2026

natural-language-oee-query

The shift supervisor on the rolling mill morning shift has six dashboards open: the MES OEE board, the SCADA HMI, the historian trend client, the quality lab report, the maintenance work order list, and the daily production tracker exported into Excel. To answer the question her boss just asked over WhatsApp — "why was plate bay output 18% under target last shift?" — she has to triangulate across all six. Twenty minutes later she has a hypothesis (a quality reject burst on coil 4477 plus a 22-minute crane wait on bay 2C) but cannot prove it without pulling another dashboard. By that point her boss has stopped waiting and the next shift's standup is starting. Multiply this by every shift, every line, every plant in the company. Most factories already have all the data they need — they just don't have a way to ask the question. Natural Language OEE Query is the AI layer that closes the gap. The supervisor types or speaks the question in plain English (or Hindi, or any plant working language) into a single copilot panel; the copilot classifies the intent, decomposes the answer into Availability × Performance × Quality, attributes the loss to the correct shift, line, or asset, surfaces the most likely root cause from the historian and maintenance data, and suggests the action. Twenty minutes of dashboard digging collapses into 8 seconds of conversation. The lab still owns the certified quality numbers; the MES still owns the timestamps; the maintenance system still owns the work orders. The copilot does not replace any of those. It just lets the supervisor finally ask them all at once. Runs on the iFactory turnkey on-prem AI server stack — RTX PRO 6000 Blackwell Workstation in the control room for single-plant, NVIDIA DGX Station GB300 Ultra Desktop Superchip for corporate fleet rollout. Six-week deployment from PO to live copilot on the floor. Come see us live at SAP Sapphire Orlando, May 11–13 — interact with the copilot hands-on, type your own questions into the panel, watch the answers render against a representative plant model. Register for the booth here.

SAP SAPPHIRE ORLANDO · MAY 11–13, 2026 · INTERACT WITH THE COPILOT HANDS-ON
NATURAL LANGUAGE OEE COPILOT · ASK THE PLANT IN ENGLISH · A×P×Q DECOMPOSITION · ROOT CAUSE

Stop Digging Through Six Dashboards. Ask The Plant.
"Why Is Plate Bay Slow?" — The Copilot Answers In 8 Seconds.

Type or speak the question in plain English. The copilot classifies the intent, decomposes Availability × Performance × Quality, attributes the loss to the right shift / line / asset, surfaces the most likely root cause from your historian and maintenance data, and suggests the next action. Six-week deployment. Runs on the on-prem AI server stack you own outright. Come see us live in Orlando — type your own questions into the copilot, hands-on at the booth.

~8 sec
Question typed to answer rendered · typical
A × P × Q
OEE decomposition shown for every answer
6 weeks
PO to live copilot · not 6 months
100%
On-prem · air-gapped · you own it
The Dashboard Problem

Six Dashboards, One Question, Twenty Minutes — Every Shift, Every Plant

The published OEE benchmark data is unambiguous: most discrete manufacturing plants operate between 40% and 65% OEE; world-class is 85%. The 20-to-45-percentage-point gap is mostly availability and performance losses that the dashboards can show — but only if a person spends the twenty minutes to dig through them. Most days, on most lines, that twenty minutes happens once instead of every shift. The copilot is not a new dashboard; it is the conversational layer that lets the same data finally answer the question the supervisor was always trying to ask. Talk to our deployment lead about the dashboards your supervisors have open today.

SIX DASHBOARDS, MANUAL TRIANGULATION
Twenty minutes per question, often abandoned

Supervisor opens MES OEE board, SCADA HMI, historian trend, quality lab report, maintenance work order list, daily Excel tracker. Triangulates manually. Finds half the answer. Boss has moved on. Next shift starts. The data was always there; the path through it was the bottleneck.

ASK THE PLANT · A×P×Q + ROOT CAUSE
+
8 seconds · English in, decomposed answer out

Supervisor types "why was plate bay output 18% under target last shift". Copilot returns: Availability 78% (target 90%), Performance 88% (target 95%), Quality 96% (target 99.9%); dominant loss = 22-min crane wait on bay 2C plus quality reject burst on coil 4477; suggested action = expedite the work order on crane 2C hoist, route lab feedback to caster.

COPILOT WRITES TO MES / DCS
!
The boundary you don't cross

A copilot that closes maintenance tickets, releases held lots, or adjusts production targets without human review is not a copilot — it is an unvalidated controller. The Natural Language OEE Copilot has no write path to the MES, DCS, or quality release. Recommendations only. Supervisor commits on the existing system under existing authority.

Sample Queries · The Conversation In Action

What The Supervisor Actually Types — And What The Copilot Returns

A walk-through of six representative queries from a typical morning shift on a steel plate mill. Numbers are illustrative. The pattern is what matters: the supervisor asks in the language she would use over WhatsApp, the copilot returns the structured decomposition with the root-cause hint and the suggested action, and the conversation can continue as a follow-up. Different plants ask different questions; the same intent classifier handles all of them.

PLATE MILL COPILOT · SHIFT A · OPERATOR M. KAPUR
6 queries answered in 47 seconds · on-prem RTX PRO 6000 Blackwell · air-gapped
Why is plate bay slow this morning?
06:47 · intent: oee_root_cause · resolved 8.2 sec
OEE66% · vs target 80%
Availability78% · vs target 90% · main loss
Performance88% · vs target 95%
Quality96% · vs target 99.9%
Dominant loss: 22-minute crane wait on bay 2C between 04:18–04:40 plus quality reject burst on coil 4477 (8 plates rolled out of spec, 03:55–04:12). Combined effect on shift OEE: −14 percentage points.
Suggested action: expedite WO-22418 on crane 2C hoist (already drafted in CMMS); route lab feedback on coil 4477 to caster shift lead for slag practice review.
Show me availability by line for last 24 hours
06:49 · intent: availability_breakdown · resolved 4.7 sec
LineAvailabilityvs target 90%
Plate bay 1A91.2%+1.2 pp
Plate bay 2C78.0%−12.0 pp
Coil mill 389.4%−0.6 pp
Galvanising92.6%+2.6 pp
Plate bay 2C is the outlier; the crane 2C hoist event from query 1 accounts for ~10 percentage points of the gap. Recommend prioritising the maintenance window.
What are my top 3 reject reasons this week?
06:51 · intent: quality_top_n · resolved 5.3 sec
1Surface defect · scale pitting42% of rejects · 31 lots
2Dimensional · thickness out of band28% of rejects · 21 lots
3Mechanical · tensile below spec14% of rejects · 10 lots
Scale pitting trend is up 18% week-on-week; correlates with descaler nozzle pressure drift on plate bay 2A. Suggested action: open inspection task on descaler nozzles.
Are we on track for monthly target?
06:53 · intent: production_pace_projection · resolved 6.1 sec
Month-to-date68,420 t · vs plan 72,400 t · 5.5% behind
Days remaining11 · required pace 7,820 t/day
Current pace7,140 t/day · gap of 680 t/day
At current pace, month-end shortfall projects ~7,500 t. Closing crane 2C hoist WO and recovering plate bay 2C availability projects to recover ~5,200 t. Recommend escalation to maintenance head this morning.
Has crane 2C had problems before?
06:55 · intent: asset_history · resolved 7.4 sec
3 incidents in last 90 days. WO-21206 (12 Feb): hoist motor brush replacement, 4-hour planned outage. WO-21847 (8 Mar): trolley drive bearing, 2-hour reactive teardown. WO-22418 (today): hoist current draw drift, drafted but not yet executed. Pattern suggests bearing wear progression; recommend coordinated bearing inspection campaign across both 2C hoist and trolley assemblies.
Suggested action: route this analysis to reliability engineering; convert WO-22418 from corrective to predictive scope.
What did B-shift do differently last Tuesday?
06:57 · intent: shift_comparison · resolved 8.9 sec
B-shift Tuesday OEE 79.4% vs your A-shift average 71.2%. Primary delta: 14 fewer minutes of small stops (less than 5 min each), 6% lower scale-pitting reject rate. Walk-around log shows B-shift ran descaler nozzle inspection at 22:30 hand-over — not in your shift's procedure. Reject-rate delta correlates strongly with this practice.
Suggested action: add descaler nozzle inspection to A-shift hand-over checklist; flag for shift supervisor review at next stand-up.

What you noticed in the conversation: the copilot is not just answering each question in isolation. By query 4, it is using context from query 1. By query 6, it is comparing patterns across shifts and connecting them to walk-around log entries it never saw before this session. That is the difference between a chat-based dashboard reader and a real copilot — conversation memory plus access to the full operational corpus, not just the OEE board.

What OEE Decomposes Into

A × P × Q — Three Numbers, Six Big Losses, One Multiplied Score

Every answer the copilot returns shows the OEE decomposition because that is what makes the answer auditable. Availability times Performance times Quality is the formula Seiichi Nakajima published as part of Total Productive Maintenance — world-class plants run at 85% OEE, most plants run at 40% to 65%. The gap lives in one or two of the three factors; the copilot's job is to surface which one and why. Below is what each factor means and what loss types live inside it.

A
Availability
Run Time ÷ Planned Production Time
World-class target: ≥ 90%
Loss types tracked by the copilot:
  • Planned downtime (changeover, maintenance windows)
  • Unplanned downtime (breakdown, material starvation, operator availability)
  • Setup and adjustment time
P
Performance
(Ideal Cycle Time × Total Parts) ÷ Run Time
World-class target: ≥ 95%
Loss types tracked by the copilot:
  • Small stops (less than 5 min — jams, sensor trips, micro-adjustments)
  • Slow cycles (running below ideal cycle time — wear, settings, material variation)
  • Reduced speed running (operator throttling, tool conditioning)
Q
Quality
Good Count ÷ Total Count
World-class target: ≥ 99.9%
Loss types tracked by the copilot:
  • Production rejects (scrap and rework during stable production)
  • Startup rejects (parts produced during ramp-up not meeting spec)
  • Held lots awaiting disposition

Why showing the decomposition matters: "OEE was 66% last shift" is not actionable. "Availability was 78% because of a 22-minute crane wait on bay 2C" is. The copilot always returns the three component scores alongside the headline OEE so the supervisor knows which lever to pull.

How Root Cause Is Inferred

Intent Classifier · OEE Decomposition · Attribution · Action Hint

The copilot's pipeline is built around the published LLM + RAG architecture pattern for manufacturing copilots, applied specifically to OEE inference. Five stages, each with a specific job. The whole pipeline runs on the on-prem RTX PRO 6000 Blackwell server — sub-second from question typed to answer streaming.

01
INTENT CLASSIFIER
What is the supervisor actually asking?

Fine-tuned classifier on top of the on-prem LLM. Maps the natural-language question to one of roughly 40 intent categories: oee_root_cause, availability_breakdown, performance_loss_attribution, quality_top_n, asset_history, shift_comparison, production_pace_projection, downtime_explainer, and so on. Multilingual — English, Hindi, Spanish, Mandarin, Portuguese, Tamil, Bahasa, German.

02
OEE DECOMPOSITION
A × P × Q computed against your data

For the time window the question implied (last shift, last 24h, week-to-date, MTD), the copilot pulls the relevant historian and MES data, computes Availability, Performance, and Quality, and identifies which of the three is the dominant gap against world-class or your plant's targets. This is the structured number that anchors every answer.

03
ATTRIBUTION
Where in the plant does the loss live?

Once the dominant factor is identified, attribution splits the loss across shifts, lines, assets, products, and time slices. SHAP-style feature attribution on the structured time-series; for the "why" questions, the LLM augments the structured attribution with the maintenance log entries, walk-around notes, and quality-team notes from the same window.

04
ROOT-CAUSE HINT
Most likely cause · with confidence

RAG-grounded against your CMMS work order history, walk-around logs, quality lab notes, and historical incident reports for similar patterns. Returns the most likely root cause with a confidence score and a citation pointer back to the source records the supervisor can open if she wants to verify. Multiple causes ranked when the data does not support a single dominant explanation.

05
ACTION HINT
What can the supervisor do about it?

Suggested action grounded against your standard operating procedures and your existing maintenance and quality workflows. Not a write — the supervisor commits the action on the existing system (CMMS work order, MES adjustment, shift handover note) under her own authority. The copilot proposes; the supervisor disposes.

Where The Data Comes From

Sources Already Sitting In Your Plant — Nothing New To Install

The copilot does not require a new data layer. It connects through standard read-only integrations to the systems your plant already runs — the same systems the supervisor has open across her six dashboards today. The integrations are part of the six-week deployment; iFactory's enterprise team owns them.

MES & PRODUCTION
SAP Manufacturing Werum PAS-X Körber Vibe Rockwell FactoryTalk Siemens Opcenter Custom MES via SQL

Production schedules, batch records, OEE timestamps, shift assignments, product hierarchy.

HISTORIAN & SCADA
Aveva PI System Inductive Ignition GE Proficy Wonderware OPC-UA direct InfluxDB / TimescaleDB

Time-series sensor data, run states, downtime reasons, cycle times, throughput counters — the structured layer the OEE decomposition computes against.

QUALITY & LIMS
LabWare LIMS SAP QM Veeva QualityOne MasterControl Custom LIMS via SQL

Lab samples, certified results, hold and release decisions, deviation records, CAPA history.

CMMS & MAINTENANCE
OXMaint IBM Maximo SAP PM IFS Cloud Infor EAM eMaint

Work orders, asset history, planned maintenance schedules, spare parts movements, technician notes.

UNSTRUCTURED PLANT KNOWLEDGE
SOPs & work instructions Walk-around logs Shift handover notes Incident reports OEM equipment manuals

The corpus the LLM grounds against for root-cause inference. Indexed once during deployment, refreshed continuously.

SHIFT & HR
Shift roster systems Skills matrix Training records Operator certification

Used for shift comparison queries and skills-gap inference where applicable.

Read-only by default. The copilot pulls data from these systems — it does not write back to them. Any action the copilot recommends gets committed by the supervisor on the originating system under her existing authority. No data leaves your perimeter; the corpus indexes and the model both live on the on-prem appliance.

6-Week Deployment

PO Today · Live Copilot On The Floor In Six Weeks

The Natural Language OEE Copilot is the iFactory deployment with the fastest path to live. No new instrumentation. No data lake to build. The connectors to MES, historian, LIMS, and CMMS are already developed and run on the on-prem appliance from day one. The six-week schedule below is what your plant lead should expect — and what we Gantt at PO.

WEEK 1
Hardware ship + connector setup

RTX PRO 6000 Blackwell Workstation arrives at your control room. iFactory field engineer racks, connects to your network, begins read-only connectors to MES, historian, LIMS, CMMS. Site survey for any non-standard data sources.

WEEK 2
Data corpus indexing

Historical OEE data, work orders, walk-around logs, SOPs, shift handover notes, incident reports indexed onto the appliance. Typically 6 to 18 months of history depending on what your systems retain. Quality validation against your existing OEE board.

WEEK 3
Intent classifier fine-tuning

Intent classifier fine-tuned on your plant's terminology — line names, asset names, product codes, shift labels, local language patterns. The "plate bay 2C" your supervisor types should resolve to the same line the historian calls "PB-2C-MAIN". This week is where the calibration happens.

WEEK 4
Shadow mode · supervisor review

Copilot answers questions in shadow mode. Two or three supervisors from your shift roster ask their actual end-of-shift questions; the answers are reviewed against the answers they would have given themselves. Disagreements logged as training data for week 5.

WEEK 5
Refinement + integration tests

Disagreement patterns from week 4 used to refine the intent classifier and the attribution logic. End-to-end integration tested against your existing CMMS, MES, and quality systems. Operator card UI styling matched to your existing dashboard standards.

WEEK 6
Go-live + handover

Copilot goes live across all shifts on the pilot line. Three days of on-site iFactory support — supervisor training, walk-around with each shift, last-mile fixes. By Friday of week 6, your plant is asking the copilot questions instead of digging through dashboards. Six weeks. PO to live.

The Hardware Underneath

RTX PRO 6000 Blackwell Workstation — The Box That Runs The Copilot

For a single-plant deployment, the workstation below is the standard. It sits in your control room, runs every iFactory AI you turn on (the copilot, the digital twin, the predictive maintenance models), and is owned outright by you. For corporate fleet rollouts where one node serves multiple plants, the NVIDIA DGX Station GB300 Ultra Desktop Superchip tier is the right answer — up to 1 trillion parameters of local inference, $85K to $100K, more capacity than your fleet will use in five years.

SINGLE-PLANT · THE COPILOT'S DEFAULT BOX
RTX PRO 6000 Blackwell Workstation — Digital Twin Server
GPUNVIDIA RTX PRO 6000 Blackwell · 96 GB GDDR7 ECC
CUDA cores24,064 · 5th-gen Tensor cores · FP4 support
CPUAMD Ryzen 9 9900X · 12-core
Memory128 GB DDR5
Storage2 TB NVMe SSD
OS & stackUbuntu 24.04 + NVIDIA Omniverse pre-loaded
Air-gappedBy default · no public-internet dependency
OwnershipCapEx · you own outright

What this gives the plant: 96 GB of GPU memory is enough to run the copilot's intent classifier plus a 70B-parameter LLM plus the digital twin plus the OEE inference pipeline simultaneously. 24,064 CUDA cores keep the response time under 10 seconds even under multi-supervisor concurrent load. FP4 support means the LLM runs efficiently at production throughput. Sits next to the DCS rack; air-gapped by default; you own the box and the trained models.

CORPORATE FLEET · HEADQUARTERS-SIZED · UP TO 1 TRILLION PARAMETERS
NVIDIA DGX Station GB300 Ultra Desktop Superchip

Imagine a sleek workstation about the size of a desktop briefcase sitting at corporate. NVIDIA's DGX Station with the GB300 Grace Blackwell Ultra Desktop Superchip — 768 GB unified coherent memory, 20 petaFLOPS of AI compute, dual 400 GbE LAN. NVIDIA states the platform supports models up to 1 trillion parameters running locally without cloud infrastructure. For a multi-plant company, this is the box at headquarters that runs every plant's copilot, every plant's digital twin, the cross-plant fleet analytics, and the corporate operator copilot — all simultaneously. Roughly $85K to $100K outright, or rented through our cloud / OXMaint Server Farm. Most customers start with the RTX PRO 6000 Workstation per plant and add the DGX Station GB300 Ultra at corporate when they're ready to roll out across the fleet.

Come see us live in Orlando — interact with both servers hands-on. Touch the RTX PRO 6000 Blackwell Workstation. Touch the NVIDIA DGX Station GB300 Ultra Desktop Superchip. Type your own questions into the copilot running on a representative plate-mill model. Watch the answer render against the OEE decomposition you would actually see on your floor. Register for the booth — SAP Sapphire Orlando, May 11–13.

FAQ

What Shift Supervisors & Plant Managers Ask First

Does the copilot work in our local language?

Yes. English by default; Hindi, Spanish, Mandarin, Portuguese, Tamil, Bahasa, German, French, Japanese, Korean, Vietnamese supported out of the box. Other languages added during Phase 1 calibration on request — the underlying LLM handles them; the intent classifier fine-tunes on your plant's terminology in the chosen language. Multilingual queries within the same conversation are supported.

Does the copilot write to our MES, DCS, or CMMS?

No, by architecture. All connectors are read-only. The copilot recommends actions; the supervisor commits them on the originating system. Work orders go into your CMMS as drafts only when the supervisor explicitly asks the copilot to draft one — and even then, the draft requires supervisor sign-off in the CMMS before it becomes active. There is no autonomous write path in the tool surface.

How accurate are the answers?

Two layers of accuracy. The OEE decomposition (A × P × Q numbers) is computed from your historian and MES data — if the underlying data is correct, the decomposition is correct. The root-cause hint is a probabilistic inference and is presented with a confidence score and citation pointers back to the source records. Below 70% confidence the copilot says so explicitly. Supervisors learn quickly which questions return high-confidence answers and which need a manual review.

What about privacy — does our shop floor data leave the building?

No. The copilot runs entirely on-prem on the RTX PRO 6000 Blackwell Workstation (or DGX Station GB300 Ultra at corporate). Air-gapped from public internet by default. The LLM weights are on the appliance; the corpus indexes are on the appliance; the audit trail is on the appliance. We do not stream prompts to a cloud LLM. We do not collect telemetry. Your plant data and your operating recipes stay on the box you own.

What if our supervisors don't trust the copilot?

That is the entire purpose of the citation pointers and the confidence scores. Every answer surfaces the source records (work orders, historian trends, lab samples, walk-around logs) the copilot used — one click opens the underlying record so the supervisor can verify. Most supervisors start sceptical and use the citations heavily for the first two weeks; by week three they trust the answers but keep the citations as audit material.

Six weeks — is that realistic?

For the standard deployment (one plant, standard MES / historian / CMMS / LIMS, no exotic data sources), yes. The connectors are pre-built; the only plant-specific work is the intent-classifier fine-tuning on your terminology and the corpus indexing of your historical records. Plants with non-standard or heavily customised systems can extend by 2 to 4 weeks — we Gantt the schedule at PO and tell you up front if your environment will push past six weeks.

Should we choose RTX PRO 6000 or DGX Station GB300 Ultra?

For one plant or single-site fleet, the RTX PRO 6000 Blackwell Workstation is the right answer. For corporate-fleet rollouts where one node at headquarters serves multiple plants and runs cross-plant analytics, the NVIDIA DGX Station GB300 Ultra is the right answer — its trillion-parameter capacity covers fleet-wide copilots plus enterprise LLM plus cross-plant digital twins on a single node. Most customers start on the per-plant tier and add the corporate tier when ready to scale.

What happens if we don't renew support after year one?

The copilot keeps running. You own the appliance, the trained intent classifier, the indexed corpus, the audit logs. Renew support and quarterly retraining annually, run it in-house with our handover docs, or do a mix. No kill switch. The intent classifier gets sharper with continued retraining as your terminology and product mix evolve; if you stop, it freezes at the last-trained state and continues running.

SAP SAPPHIRE ORLANDO · MAY 11–13, 2026 · INTERACT WITH THE COPILOT HANDS-ON

Come See Us Live In Orlando — Type Your Own Questions Into The Copilot

Two ways forward. First · come see us live at SAP Sapphire Orlando, May 11–13. Walk the iFactory booth and interact with the on-prem AI server stack hands-on — touch the RTX PRO 6000 Blackwell Workstation, touch the NVIDIA DGX Station GB300 Ultra Desktop Superchip, type your own questions into the copilot running on a representative plate-mill model, watch the OEE decomposition and the root-cause hint render in real time on the screen in front of you. Bring scenarios from your floor and our plant lead will run them through the copilot at the booth. Second: a 30-minute working session with our deployment lead — bring 90 days of your plant's OEE history, a typical week of walk-around logs, and a list of the questions your supervisors actually ask. We'll project what the six-week deployment looks like for your environment.

~8 sec
Question to answer

A × P × Q
Decomposition for every answer

6 weeks
PO to live copilot

100%
On-prem · you own it

Share This Story, Choose Your Platform!