back to overview

A historian without structure is a filing cabinet without an index

Historical data

CONTENT

  • The quiet pride of many factories
  • What a historian actually does
  • When storage does not create structure
  • The difference between data and context
  • Analysis as archive research
  • Why more data does not solve the problem
  • From time series to events
  • When historical data becomes truly usable
  • The role of Capture

The quiet pride of many factories

Ask in any manufacturing company where all process data is stored, and the answer usually comes quickly: the historian. In many plants, that system has been running quietly in the background for years, capturing almost everything that happens on the factory floor. Temperatures, pressures, speeds, vibrations, flow rates, energy consumption, cycle times. Every sensor leaves a trace.

From a technical perspective, that is impressive. Some historians store millions of data points per day. Years of process behavior are preserved. When a problem occurs, there is a reassuring sense that the data must be available somewhere.

And yet a strange contrast often appears. Despite that vast volume of historical data, analysis still turns out to be surprisingly difficult.

The historian contains everything. But it rarely tells you where to begin.

What a historian actually does

To understand that problem, it helps to look first at the original role of a historian. Historian systems are designed to store time series. They record sensor values together with a timestamp and retain that data efficiently over long periods.

That design is particularly well suited to industrial environments. Processes generate continuous streams of signals, and historians are optimized to store those streams reliably and at scale.

In that sense, the architecture of a historian focuses on three core questions:

  • how often a signal should be stored
  • how that can be done efficiently
  • how the data can be retrieved later

What a historian does not automatically solve is another question that matters just as much for analysis: how are all those signals connected?

When storage does not create structure

A historian can perfectly record that temperature rises, pressure drops, and a motor starts drawing more current. But without additional context, it remains unclear how those events relate to one another.

Imagine a production line suddenly experiences a short stop. The historian may show a clear drop in speed, followed by a peak in motor current and a small fluctuation in temperature. From that perspective, the data looks complete.

But analysis is trying to answer a different question. Not just what happened, but why it happened.

And that is where the classic historian architecture runs into a limitation. Sensor values are stored as individual time series. The relationships between events remain implicit.

A data point rarely knows which production order it belongs to, which batch was active, or which other systems were involved at that moment.

The difference between data and context

That difference may seem small, but it changes the way analysis has to be done. When an engineer investigates a problem, they rarely look at a single sensor in isolation. They are trying to understand how different events together influenced the process.

A small temperature deviation may be linked to a change in recipe settings. A rise in motor current may follow a switch in product type. A short alarm may be related to an operator intervention.

Those events usually exist across different systems. The historian contains process data. The MES system knows the production orders. Maintenance systems record interventions. Quality measurements live in other databases.

When those datasets are not explicitly linked, analysis starts to look more like puzzle-solving than investigation.

Engineers gather fragments from different systems and try to reconstruct them afterward in a single timeline.

Analysis as archive research

This is the point where the filing cabinet comparison becomes relevant. A filing cabinet can contain thousands of documents. Each one may hold valuable information. But without an index or a clear structure, it becomes difficult to find the right document quickly.

Many historians work in much the same way. They contain huge volumes of process data, but without a clear index of events.

So when a problem needs to be investigated, the work often begins with searching for relevant signals. Engineers export graphs, compare timelines, and try to reconstruct which events together caused a deviation.

That process can be extremely time-consuming, especially when several systems are involved.

In a complex root cause analysis, it is easy to spend hours bringing together datasets from different sources before the actual analysis can even begin.

Why more data does not solve the problem

A natural response to this situation is to collect even more data. More sensors. More logging. More historical storage. The assumption is that a larger dataset will automatically produce more insight.

But without structure, more data changes very little.

A bigger historian simply means a bigger filing cabinet. The number of documents increases, but the question of how to find the right ones quickly remains exactly the same.

In some cases, additional data even makes analysis more difficult. More signals mean more possible correlations, more graphs, and more hypotheses.

Without a structure that connects events to one another, it becomes harder to determine which signals are actually relevant.

From time series to events

What many industrial data architectures are missing is a layer in which events are described explicitly. Instead of storing sensor values alone, data is organized around events that carry meaning within the production process.

A machine starts. A batch begins. A recipe is adjusted. An operator makes an intervention. An alarm is triggered. Each of those events creates a context in which sensor values gain meaning.

When process data is linked to events like these, the role of historical data begins to change.

The historian remains an important part of the infrastructure, but it is complemented by a structure in which events and assets remain explicitly connected.

Analysis then shifts from searching for signals to understanding relationships between events.

When historical data becomes truly usable

In that kind of architecture, historical data is no longer just a storage location. It becomes a navigation system. An engineer can, for example, select a specific stoppage and immediately see which process parameters changed just before it happened, which production order was active, and which other systems registered anomalies at the same time.

The data no longer needs to be stitched together manually.

The structure in which it is stored already makes those relationships visible.

That difference may seem technically subtle, but it fundamentally changes how fast and how reliably analysis can happen.

The role of Capture

Capture is built on exactly that idea. Instead of storing industrial data purely as time series, the platform organizes data around a semantic structure in which assets, processes, and events remain connected.

Historian data is still collected, but it is immediately placed within a context that makes clear where a signal comes from and which events it relates to.

As a result, historical data changes from an archive of isolated signals into a network of events that actively supports analysis.

The historian remains an essential part of the data landscape. It simply loses its role as an isolated filing cabinet.

And once that shift happens, historical data becomes truly navigable for the first time.