If you’ve spent any time on a shop floor, you know the truth: raw sensor data is a lie. Between PLC noise, intermittent Wi-Fi in the heat-treat zone, and the perennial "midnight sync error" between your MES and ERP, your data pipeline is likely a graveyard of missing timestamps and ghost readings. If your goal is true Industry 4.0 maturity, you can't just dump raw telemetry into a bucket and pray for insights.
I see vendor decks every week that promise "real-time analytics" without showing me a single line of Kafka or a mention of circuit breaker patterns. That doesn’t fly. If you want a data platform that actually works, you need to handle data quality at the edge, in the pipe, and in the lake. Here is how you build for reality.
The "Dirty" Reality: IT/OT Integration Challenges
The divide between IT and OT isn't just cultural; it’s architectural. Your PLC data is high-frequency and binary-heavy, while your ERP data (like SAP or Oracle) is transactional and relational. Trying to join these without a proper semantic layer is why your Power BI dashboards are showing negative cycle times. When I look at partners like STX Next, NTT DATA, or Addepto, the first question I ask is: How fast can you start and what do I get in week 2? If they can’t show me a raw-to-refined pipeline in 14 days, they aren't ready for a shop floor.
The Proof Point Checklist
Before you commit to a stack, build your baseline. If you aren't tracking these, you're flying blind:

Metric Industry Standard Why it Matters Records per Day 5M+ Validates ingestion throughput limits. Latency (OT to ERP) < 500ms Critical for real-time alerting. Downtime Percentage < 0.5% Measures data pipeline availability. False Negative Rate < 2% Indicates efficacy of outlier detection.
Platform Selection: Azure vs. AWS vs. The Modern Lakehouse
Whether you’re on Azure (Fabric/ADF) or AWS (Kinesis/Glue), the platform choice is secondary to your processing framework. I see too many teams get bogged down in "cloud religion." Focus on the tooling that bridges the gap between streaming and batch.
- Databricks/Spark: Non-negotiable for large-scale sensor aggregation. Use Delta tables to enforce schema constraints at the bronze-to-silver transition. Snowflake: Excellent for the ERP integration side, especially with Snowpipe for continuous ingestion. Azure Fabric: Great for unified governance, but you need to pair it with proper Airflow orchestration to ensure your batch jobs don't starve your streaming telemetry.
The Three Pillars of Sensor Data Quality
You cannot fix bad data once it's in the warehouse. You have to handle it where it lives.
1. Pipeline Validation (The "Guardrail" Approach)
You need to implement schema enforcement. If a sensor reports a temperature of 5000°C because of a bad bridge, your pipeline should move that record to a "quarantine" table. Use dbt (data build tool) to run tests against your silver layer. If your cycle time average spikes by 300% in a 10-minute window, the pipeline should trigger an alert before the dashboard ever updates.
2. Handling Outliers in Streaming Pipelines
Don't just use simple averages. Sensor data is noisy. Implement Z-score or Interquartile Range (IQR) filtering in your streaming jobs—ideally using Kafka Streams or Spark Structured Streaming. You https://dailyemerald.com/182801/promotedposts/top-5-data-engineering-companies-for-manufacturing-2026-rankings/ need to distinguish between "process drift" (a machine tool wearing down) and "sensor noise" (a loose connection on an I/O card).
3. Contextualization (The MES/ERP bridge)
Data without context is noise. Your sensor data is useless unless it is joined with the work order (MES) and the batch material information (ERP). Use a "Golden Record" approach where the telemetry is stamped with the current product ID and shift lead. This is where NTT DATA and other system integrators often prove their value—integrating the disparate schemas of legacy PLC registers with modern cloud metadata.
Batch vs. Streaming: Stop Choosing One
The "real-time" buzzword annoys me because most manufacturing use cases don't actually need sub-millisecond updates—they need sub-second visibility. Use a Lambda or Kappa architecture:
Streaming Path: Use for immediate alerting (e.g., "Machine X is about to overheat"). Batch Path: Use for complex trend analysis, dbt transformations, and long-term storage in your data lake.By splitting the pipelines, you protect your compute costs while ensuring your machine learning models (for predictive maintenance) have the historical context they need. If you’re using Addepto for AI/ML work, ensure they are working on top of cleaned, validated silver tables, not your raw ingestion bucket.
Final Thoughts: The Week 2 Reality Check
When you sit down with your consultants or your internal engineering leads, cut through the fluff. Ask them: "Which orchestrator are we using to move data from the edge gateway? What is the backfill strategy if the MQTT broker drops packets? How are we testing for sensor drift in the silver layer?"

Manufacturing data is messy. It’s supposed to be. But if you handle the quality at the ingestion gate using rigorous schema validation and outlier detection, you stop being a data janitor and start being a data engineer. Build the platform to be resilient to the mess, because the factory floor isn't going to get any quieter.
Recommended Reading for Your Next Sprint:
- Implementing circuit breakers in Kafka for high-frequency IoT. Using dbt tests to catch "frozen" sensor data in production. Comparing Delta Lake vs. Apache Iceberg for manufacturing time-series storage.