November 2, 2025
November 2, 2025
According to Dr. Hamza Akta, Field CTO for Manufacturing at MongoDB, most manufacturers focus on collecting more data and training better models while ignoring the critical layer in between: how your AI agents actually access and understand your operational data.
Traditional databases weren't designed for the way AI agents work. They store data in tables and columns optimized for transactional processing. But AI agents need semantic understanding, the ability to find similar failures even when the exact keywords don't match, retrieve relevant context from unstructured documents, and reason across time-series data, maintenance logs, and technical manuals simultaneously.
This requires vector databases with semantic search capabilities. Here's why this matters now and how to architect data infrastructure that actually enables intelligent agents rather than constraining them.
Before diving into vectors and databases, understand what makes AI applications work in manufacturing. Akta breaks it down to three essential components:
The large language model: GPT-4, Claude, LLaMA, or specialized models. This provides reasoning and natural language capabilities. Everyone focuses here first.
The embedding model: Converts your data (text, images, time-series signals) into numerical representations that capture semantic meaning. Most people ignore this until it becomes a problem.
The database: Stores all your operational data—structured, unstructured, time-series—and makes it accessible to AI with the right context. This is where most implementations fail.
Most manufacturers get component #1 right. They struggle with #2. They completely underestimate #3.
The pattern Akta sees repeatedly: Companies deploy powerful LLMs, wonder why responses are inaccurate or irrelevant, then blame the AI model. The actual problem? The LLM never received the right context because the database architecture can't deliver it.
Your relational databases work great for transactions—order processing, inventory management, production tracking. They're terrible for the way AI agents need to work.
Problem 1: Keyword-only search doesn't understand meaning. An operator searches for "spindle vibration issues" in your maintenance system. The database returns exact matches. But previous failures logged as "excessive bearing noise" or "high-frequency oscillation" never appear, even though they describe the same problem. The AI agent only sees partial information and makes suboptimal recommendations.
Problem 2: Context lives in silos. Equipment manuals in PDF form. Maintenance logs in your CMMS. Time-series data in historians. Production context in your MES. Each sits in a separate system. AI agents need to reason across all of them simultaneously, but your database architecture forces them to query each system independently and somehow piece together meaning.
Problem 3: Structured and unstructured data don't mix. Your bill of materials is structured data. Your work instructions are unstructured text. Your failure analysis reports combine both. Traditional databases handle one or the other, never both elegantly. AI agents need seamless access to everything.
Problem 4: No semantic understanding of relationships. Two different cutting machines have similar failure modes, but they're different models with different part numbers. A relational database sees them as completely unrelated because the keys don't match. An AI agent should understand they're similar and apply lessons learned from one to the other.
These aren't edge cases. They're fundamental mismatches between how databases traditionally organize information and how AI needs to access it.
Everyone talks about vectors and embeddings without explaining what they are or why they matter. Here's the practical explanation:
Vectors are numerical representations of meaning. Take any piece of data—a maintenance log entry, an equipment manual page, a sensor time-series pattern, an image of a defect—and convert it into an array of numbers (typically 768 or 1536 dimensions).
These numbers aren't random. They're coordinates in high-dimensional space where semantically similar items cluster together.
The practical impact: "Bearing failure due to lack of lubrication" and "insufficient grease causing bearing wear" use completely different words. Keyword search sees them as unrelated. Vector embeddings recognize they describe the same problem because the semantic meaning is similar.
This enables the search patterns AI agents actually need:
Similarity search: Find all previous failures similar to this current issue, even when described differently.
Semantic retrieval: Get relevant context for a query based on meaning, not just keyword matching.
Multi-modal search: Find equipment based on a photo, audio sample, or sensor pattern—not just text descriptions.
Cross-domain connections: Link failures in one machine type to similar patterns in other equipment, even when they're catalogued completely differently.
Akta describes a client in insurance using vector search on images of broken windshields to find similar previous claims and their resolution paths. The same concept applies to manufacturing: search through images of defects, audio signatures of failing motors, or vibration patterns to find semantically similar cases.
Most AI implementations end up with multiple databases: one for time-series data, another for structured data, a separate vector database for embeddings, maybe a graph database for relationships. Managing five different databases for one AI application creates massive complexity.
MongoDB's approach: store everything in one flexible document database with native capabilities for all data types.
Time-series data: Specialized collections optimized for sensor data, production metrics, and operational telemetry. Fast reads, fast writes, high compression. Your equipment data flows from MQTT brokers directly into time-series collections stored as JSON documents.
Structured data: Bills of materials, work orders, production schedules, quality records. Also JSON documents in regular collections. No rigid schema—add fields as your data model evolves without migrations.
Vector embeddings: Arrays of numbers stored in the same database alongside your operational data. Create indexes on these vectors to enable semantic search without moving data to a separate system.
Unstructured documents: PDFs, manuals, images converted to vectors and stored with metadata. Search across them semantically while maintaining connections to structured context.
Graph relationships: Traverse connections between parts, processes, equipment, and failures using MongoDB's graph lookup capabilities within the same database.
The key advantage: Your AI agent queries one database and gets everything it needs—time-series context, maintenance history, equipment manuals, similar previous failures, and bill of materials—in one operation.
Large language models out of the box know general manufacturing concepts. They don't know your specific equipment, processes, quality standards, or operational constraints.
That's where Retrieval Augmented Generation (RAG) becomes critical. Instead of sending raw questions to an LLM, you:
The critical insight: RAG isn't optional for manufacturing AI. It's the only way to ground LLM reasoning in your operational reality.
MongoDB's vector search enables this pattern natively. You're not integrating separate vector databases—everything happens in one query using their aggregation framework.
Akta's predictive maintenance use case illustrates how vector databases transform AI applications from alert systems to decision support:
Traditional approach: Sensors detect anomalies. System alerts operators. They manually figure out what to do next.
LLM-powered approach enabled by vectors:
Phase 1 - Criticality Analysis: Before even deploying sensors, use RAG to query structured maintenance data and unstructured institutional knowledge. Ask "Which equipment should I prioritize?" The system analyzes failure frequencies, maintenance costs, downtime impact, and tribal knowledge from maintenance logs to recommend where sensors deliver maximum ROI.
Phase 2 - Anomaly Detection: Deploy sensors, stream time-series data to MongoDB, run anomaly detection models. Store alerts and patterns.
Phase 3 - Intelligent Scheduling: When anomalies occur, use vector search to find similar previous failures and their resolutions. Pull relevant equipment manuals, SOPs, and work orders. Query inventory systems for parts availability. Ask LLM to "generate a repair plan based on this context" including who should do the work and when.
The difference: You go from "something's wrong" to "here's exactly what's wrong, here's how we fixed it last time, here are the parts we need, here's who should do it and when" automatically generated with factory-specific context.
This only works if your database can seamlessly combine time-series anomaly data, unstructured maintenance logs, structured inventory records, and equipment documentation in one semantic search.
LLMs answer questions. AI agents complete tasks. The difference matters for manufacturing applications.
An AI agent has three core capabilities beyond basic LLM interaction:
Memory (both short-term and long-term): Remembers what happened in previous steps of a task. Recalls information from previous sessions. Learns from past decisions.
Tool access: Can query databases, call APIs, invoke other models, trigger actions in operational systems. Not just generating text—actually doing work.
Orchestration: Breaks complex tasks into steps, creates plans, uses tools in sequence, adapts based on results.
Example task: "Optimize production parameters for Line 3 while maintaining quality."
A basic LLM can't do this. It would generate generic advice.
An AI agent:
MongoDB's role: Provides the memory layer (storing short and long-term context), the data layer (time-series, structured, unstructured), and the tool layer (vector search, graph traversal, aggregation framework).
Akta points out that MongoDB integrates with popular AI frameworks (LangChain, LangGraph, LlamaIndex) as a memory provider and data source. Your agents don't directly manage database connections—the framework handles it using MongoDB as the backend.
For complex manufacturing tasks, single agents struggle. Multi-agent systems where specialized agents collaborate work better.
Akta's machine optimization example:
Predictive Maintenance Agent: Monitors sensor data and anomaly detection outputs. Task: predict failures and recommend actions to extend equipment lifespan. When vibration spikes, recommends "slow spindle speed to reduce stress."
Process Optimization Agent: Receives the maintenance recommendation. Task: maximize production efficiency. Runs simulation to assess impact on cycle time and throughput. Determines optimal speed reduction that balances reliability with production targets.
Quality Assurance Agent: Receives the optimized parameters. Task: ensure product quality isn't compromised. Analyzes how parameter changes affect part quality based on historical data. Provides feedback on acceptable parameter ranges.
Human Oversight: Reviews agent recommendations before final parameter adjustments.
The agents collaborate by passing data and recommendations between each other, each focusing on their specialized domain. None can solve the problem alone—it requires coordination.
Why this architecture works: Each agent has clear constraints and rules of engagement. They don't try to reason about everything. They combine specialized perspectives to reach better decisions than any single agent could.
MongoDB's role: Provides shared context all agents access. Stores intermediate decisions. Tracks the reasoning chain. Enables each agent to query relevant data for their domain.
Two semantic technologies get conflated: vector databases and graph databases. They solve different problems.
Vector databases excel at: Finding similar items based on semantic meaning. "Show me all failures similar to this current issue." Works great for unstructured data—text, images, audio, sensor patterns.
Graph databases excel at: Modeling explicit relationships and dependencies. "Trace the failure path from this component through the entire assembly." Works great for structured connections—parts hierarchies, process flows, supply chains.
When you need both: Many manufacturing use cases benefit from combining them:
MongoDB handles both patterns. Graph traversal through their aggregation framework. Vector search through native indexing. No separate databases to manage.
For manufacturing leaders planning LLM and agent deployments:
Phase 1 - Data Foundation (Weeks 1-4):
Phase 2 - RAG Implementation (Weeks 5-8):
Phase 3 - Single Agent Deployment (Weeks 9-12):
Phase 4 - Multi-Agent System (Months 4-6):
Phase 5 - Scale and Optimize (Months 6-12):
Akta's advice: Don't start with the hardest problems. Start with high-ROI use cases that prove value quickly:
Operator assistance: Chatbots that help operators find information in manuals, SOPs, and previous work orders. Saves 10+ minutes per lookup. Immediate productivity gain.
Maintenance scheduling: Automated generation of repair plans based on anomaly type, parts availability, technician skills, and production schedule. Reduces planning time by 70%+.
Shift handover: Automatic categorization and summarization of what happened during shifts—alarms, quality issues, production delays. Eliminates manual report writing.
Root cause analysis: Vector search through similar previous failures to identify common patterns and proven solutions. Reduces troubleshooting time dramatically.
Document search: Semantic search across all technical documentation to find relevant information based on natural language queries, not keywords. Makes institutional knowledge accessible.
These use cases share characteristics that make them ideal starting points:
The technical architecture is straightforward. The integration challenges are real.
The gap: AI agents typically run in IT environments (cloud, data centers). They need data from OT environments (shop floor, control systems). Those environments use different protocols (Modbus, Profinet vs. REST APIs), different security models (air-gapped vs. cloud-connected), and different uptime requirements (99.9% vs. 99.999%).
Akta's recommendations:
Use industrial IoT platforms as the bridge. Don't try to connect AI agents directly to PLCs. Connect to a platform (MQTT brokers, edge gateways) that speaks both languages.
Store OT data in cloud-accessible databases. Stream time-series data from shop floor to MongoDB Atlas (their cloud offering). AI agents query the database, not the shop floor directly.
Implement proper security: Separate operational control networks from data networks. AI agents read operational data but don't control equipment directly without human approval.
Plan for brownfield integration: Most manufacturing is brownfield with decades-old equipment. Budget for connectivity solutions that work with legacy systems, not just greenfield deployments.
Design for high availability: Manufacturing doesn't tolerate downtime. Your database and AI infrastructure needs redundancy, failover, and uptime guarantees that match operational requirements.
The companies succeeding with AI agents treat integration as a first-class concern from day one, not an afterthought after building the AI.
Every previous blog post in this series—edge AI, software-defined control, knowledge graphs, time-series databases, visual inspection—assumed you'd figure out the database layer somehow.
But the database layer determines what your AI can actually do:
Without vector search: Your agents can only find exact keyword matches. They miss similar failures logged with different terminology. They can't learn from analogous situations.
Without unified data platform: Your agents query five different systems, piece together context manually, introduce latency, and struggle with data consistency.
Without proper memory architecture: Your agents forget previous interactions. They can't learn from past decisions. They can't maintain context across multi-step tasks.
Without semantic understanding: Your agents treat your 30 years of maintenance logs as noise instead of institutional knowledge to learn from.
The companies deploying AI successfully aren't using better LLMs than everyone else. They're using better data infrastructure—specifically, vector databases that enable semantic search, unified platforms that eliminate data silos, and memory architectures that let agents actually learn.
MongoDB isn't the only option, but the pattern matters: Your AI is only as smart as your database architecture allows it to be.
If you're leading AI initiatives in manufacturing:
Audit your current data architecture: Can you retrieve semantically similar items? Can you combine time-series, structured, and unstructured data in one query? Can you traverse relationships between equipment and failures? If not, you're building AI on a weak foundation.
Start with one RAG use case: Pick operator assistance or maintenance scheduling. Convert your manuals and logs to vectors. Implement basic retrieval augmented generation. Measure the accuracy improvement over raw LLM responses.
Choose databases that support multiple data types: Time-series, documents, vectors, and graph capabilities in one platform. Avoiding multiple specialized databases reduces complexity dramatically.
Design memory into your AI agents: Don't treat agents like stateless services. Give them memory—both short-term (within tasks) and long-term (across sessions). This transforms chatbots into intelligent assistants.
Plan for high availability: Your AI infrastructure needs the same uptime guarantees as your production systems. Cloud databases with built-in redundancy, automatic failover, and geographic distribution.
Focus on ROI, not technology: Start with problems that cost you money today. Wasted troubleshooting time. Manual scheduling. Difficult information access. Prove value in weeks, then expand.
The future of manufacturing AI isn't about bigger models. It's about smarter data infrastructure that makes your operational knowledge accessible to AI in ways that actually matter—semantic search, unified context, continuous learning, and autonomous action.
Kudzai Manditereza is an Industry4.0 technology evangelist and creator of Industry40.tv, an independent media and education platform focused on industrial data and AI for smart manufacturing. He specializes in Industrial AI, IIoT, Unified Namespace, Digital Twins, and Industrial DataOps, helping digital manufacturing leaders implement and scale AI initiatives.
Kudzai hosts the AI in Manufacturing podcast and writes the Smart Factory Playbook newsletter, where he shares practical guidance on building the data backbone that makes industrial AI work in real-world manufacturing environments. He currently serves as Senior Industry Solutions Advocate at HiveMQ.