Skip to content
Platform ExplainerSep 202510 min read

ArgusAI: Why We Built a Fully Offline, On-Premises AI Layer for Industrial IoT

ArgusAI
platform-explainerargusaion-premises-aiair-gappedprivate-inferenceregulated-industriesllmmcp-serverera-3

The AI Capability Gap in Regulated Industries

AI tools have become genuinely useful for industrial operations. Natural language queries across operational data. Pattern recognition across maintenance histories. Anomaly detection that improves with accumulated history. Early warning systems that surface failure signatures before they become failures.

But the organizations that most need these capabilities — defense manufacturers, classified shipyards, critical infrastructure operators, pharmaceutical plants, utilities running operational technology on isolated networks — are precisely the organizations that can’t use cloud AI platforms.

It isn’t that they don’t want to. It’s that their data cannot legally, contractually, or operationally be transmitted to external servers.

ArgusAI exists because these organizations need AI and the cloud AI ecosystem can’t serve them.


What ArgusAI Is

ArgusAI is a fully offline, on-premises AI deployment framework for industrial environments. It runs on infrastructure you control — your data center, your secure network, your air-gapped enclave — and never establishes outbound connections to external AI services.

The three components of an ArgusAI deployment:

On-premises LLM hosting: ArgusAI packages and deploys large language models on your infrastructure using optimized inference engines. Supported model classes include instruction-tuned LLMs in the 7B–70B parameter range, running on GPU inference hardware with quantization for hardware-constrained environments. The model runs on your servers. Inference runs on your hardware. Your data never reaches an external endpoint.

MCP servers: Model Context Protocol servers are the bridge between the LLM and live operational data. MCP servers run inside your network, connect to ArgusIQ’s data model (or other operational data sources), and provide the LLM with structured access to current and historical operational data. When you ask “which motors on Press Line 2 are showing early-warning patterns?”, the MCP server is what translates that question into a structured query, retrieves the relevant data, and provides it to the model as context.

Private inference pipeline: The complete request-response cycle — from user query to answer — runs entirely within your environment. Query enters the system. MCP server retrieves relevant context from operational data. LLM processes the query with context. Answer returned to user. Nothing leaves the network perimeter.

graph LR A[User Query] --> B[ArgusAI Inference] B --> C[MCP Server] C --> D[ArgusIQ Data] D -->|context| B B --> E[Answer]
Scroll to see full diagram

What ArgusAI Can Answer

ArgusAI’s capabilities depend on two things: the quality of the LLM and the richness of the operational data the MCP servers can access. In a full ArgusIQ deployment, ArgusAI has access to:

  • Live and historical sensor telemetry for every connected asset
  • Digital twin records: asset identity, specifications, operational baseline, health score
  • Complete maintenance history: every work order, every repair finding, every parts record
  • Alert history: every condition detected, every acknowledgment, every escalation
  • Spatial data: asset locations, zone assignments, floor plan layouts
  • Work order status: open, in-progress, completed, overdue

With that data model as context, ArgusAI can answer operational questions that would otherwise require a data analyst to write a query:

“Show me every asset in Building A that has had the same failure mode more than twice in the past 18 months.”

“What is the average time between alert detection and work order closure for the maintenance team on first shift vs. second shift?”

“Which assets currently show the vibration signature pattern that preceded bearing failures in the last two quarters?”

“Generate a compliance summary for the cold chain monitoring in Zone 3 for the past 30 days.”

“What happened with the compressed air system at Plant B last Thursday between 6 PM and midnight?”

These aren’t searches or dashboard filters. They’re natural language questions answered by an AI that has read the full context of the operational record and can reason across it.


The Regulated Industry Use Cases

Defense Manufacturing and Shipbuilding

Defense contractors under ITAR restrictions, companies holding facility clearances, shipyards building classified vessels — these organizations operate on networks designed to prevent data exfiltration. Cloud AI isn’t just a risk — it’s a policy violation.

ArgusAI deployed on a classified or ITAR-restricted network brings AI capability to production status reporting, maintenance analytics, Government Furnished Equipment accountability, and production scheduling — all within the security perimeter.

A question like “what is the current completion status of all structural modules for Hull 127, and which modules are behind schedule based on historical production rates?” gets answered inside the network, from data that never left the building.

Critical Infrastructure and Utilities

Water utilities, electrical grid operators, natural gas pipeline companies — these organizations run operational technology on networks intentionally isolated from the internet. The air gap is a security design, not a legacy limitation.

ArgusAI runs inside that air gap. Natural language queries against sensor data, maintenance history, and compliance records — without requiring an internet connection that the security architecture was designed to prevent.

Pharmaceutical and Regulated Manufacturing

FDA 21 CFR Part 11 compliance, GMP documentation requirements, batch record integrity — pharmaceutical and biotech manufacturing operates under regulatory frameworks that impose strict controls on where data goes and who can access it.

ArgusAI’s on-premises deployment means AI-assisted compliance documentation, batch record queries, and process analytics without transmitting regulated data to external servers.


How ArgusAI Differs From Cloud AI + VPN

A common question: why not just use a cloud AI service behind a VPN, routing requests through your network?

The answer is architectural. When a query is sent through a VPN to a cloud AI service:

  • The data leaves your network, encrypted in transit, and reaches an external server
  • The inference happens on hardware you don’t control
  • The cloud provider’s terms of service govern what they do with the data
  • The external service can be unavailable due to outages, API rate limits, or connectivity issues
  • Your classified or regulated data has been transmitted beyond your security perimeter

ArgusAI inverts this:

  • The model runs on your hardware, inside your network
  • Inference happens in your data center or secure enclave
  • No data is transmitted to any external service
  • Availability is governed by your infrastructure, not a third-party’s SLA
  • The security perimeter is never breached

Hardware Requirements

ArgusAI’s hardware requirements depend on the model size and inference volume:

Entry-level deployment (7B model, moderate query volume):

  • Single GPU server with 24–48 GB VRAM (NVIDIA A10/A30 class)
  • Adequate for a single facility with dozens of users and hundreds of queries per day
  • Smallest footprint for classified environments with restricted hardware budgets

Production deployment (13B–34B model, high query volume):

  • Dual GPU server with 80 GB+ VRAM (NVIDIA A100/H100 class) or multi-GPU configuration
  • Handles facility-wide deployment with concurrent users and real-time operational queries
  • Recommended for large facilities and multi-site operations

High-performance deployment (70B model, complex reasoning tasks):

  • Multi-GPU cluster with NVLink interconnect
  • Supports complex multi-step reasoning, report generation, and multi-document synthesis
  • Appropriate for enterprise-wide deployment with demanding analytical use cases

Quantized model variants are available for hardware-constrained deployments — 4-bit and 8-bit quantization reduces memory requirements at a modest accuracy tradeoff appropriate for most operational query workloads.


The MCP Server Architecture

Model Context Protocol servers are what make ArgusAI operational — not just conversational. Without MCP servers, an on-premises LLM can answer general knowledge questions but has no access to your operational data.

MCP servers for ArgusAI connect to:

  • ArgusIQ Asset Hub: Digital twin records, baselines, health scores, sensor telemetry
  • ArgusIQ CMMS: Work orders, maintenance history, PM schedules, parts inventory
  • ArgusIQ Alarm Engine: Alert history, current active alerts, escalation status
  • ArgusIQ Space Hub: Asset locations, zone assignments, floor plan data
  • External data sources: ERP systems, document repositories, specification databases

Each MCP server handles its domain’s data access, schema translation, and context formatting. The LLM receives structured context assembled by the MCP servers in response to each query — so every answer is grounded in the actual operational data, not the model’s training knowledge.


What ArgusAI Is Not

ArgusAI is not:

  • A replacement for domain expertise. It surfaces patterns, summarizes data, and answers operational questions. The interpretation and decision still belong to experienced operations personnel.
  • A control system. ArgusAI does not send commands to equipment, adjust setpoints, or initiate automated responses. It informs decisions; it doesn’t make them.
  • An instant deployment. Running an on-premises LLM requires GPU infrastructure, model deployment, MCP server configuration, and integration testing. The deployment timeline is measured in weeks, not hours.
  • Infallible. LLMs produce incorrect answers. ArgusAI includes citation support — every answer includes the source data records it drew from — so users can verify the answer against the underlying operational data.

ArgusAI and ArgusIQ: The Relationship

ArgusAI is designed to run alongside ArgusIQ but doesn’t require it. Organizations with other operational data systems can deploy ArgusAI with MCP servers configured to connect to those systems.

In a full ArgusIQ + ArgusAI deployment, Ask Argus — the natural language assistant built into ArgusIQ — is powered by ArgusAI when on-premises inference is required. Cloud deployments of ArgusIQ can use external AI services for Ask Argus; regulated deployments can route Ask Argus through ArgusAI running on-premises.

The architecture is the same either way. The inference engine changes based on the security requirements of the deployment environment.


The Decision to Build This

We built ArgusAI because our customers needed AI capability in environments where cloud AI couldn’t go. The organizations doing the most operationally complex, regulation-intensive, security-sensitive work in the economy — defense manufacturing, critical infrastructure, regulated processing — were being left behind by an AI ecosystem that assumed internet connectivity was a given.

It isn’t a given for these organizations. It’s a security boundary.

ArgusAI is what AI looks like when you build it to work inside that boundary rather than around it.


Talk to our team about ArgusAI for your regulated or air-gapped environment.

Ready to see how this applies to your operations?

Every article describes real capabilities you can deploy today.