I.II - The Imperative for Assurance: Criticality and Confidence LevelsThe deployment of digital twins in capital-intensive industries has moved beyond visualisation and basic monitoring to become integral to high-consequence decision-making. As these virtual representations inform decisions affecting safety, environmental integrity, and financial performance, the need for a structured assurance framework becomes paramount. The DNV-RP-A204 "Assurance of digital twins" standard provides such a framework, establishing a systematic process for developing and assuring trustworthy digital twin outputs.
4At the
core of the
DNV methodology is a
risk-based approach that links the
potential impact of a decision to the
level of assurance required for the digital twin component supporting it. This component, termed a "Functional Element" (FE), is a discrete module of a digital twin designed to support a specific "
key decision".
4 The standard defines a "
Criticality" level for each key decision, which is determined by evaluating the potential
consequence of a wrong decision and the
likelihood of making it. This Criticality, in turn, dictates the required "Confidence Level" (CL) for the FE, with three levels defined
4:
- Confidence Level 1 (CL1): For decisions with low consequences.
- Confidence Level 2 (CL2): For decisions with moderate consequences, where the FE is one of several sources of information and decision-making is not time-constrained.
- Confidence Level 3 (CL3): For decisions with high potential consequences that could cause major failures, accidents, or environmental impact, or where the FE is the primary source of information or the decision is time-constrained.
For applications with high criticality, corresponding to CL2 and particularly CL3, the requirements for data assurance become significantly more stringent. A simple validation that data is present or syntactically correct is insufficient. DNV-RP-A204 mandates a continuous, evidence-based assurance process that demonstrates the data is trustworthy, accurate, and representative of the physical asset throughout its entire lifecycle.
4 This requirement for continuous assurance is the primary driver for the technical capabilities assessed in this report.
I.II - The Quality Indicator (QI): A Cornerstone of Operational TrustA central concept within the DNV-RP-A204 framework is the "Quality Indicator" (QI). The QI is defined as a diagnostic indicator that reports the trustworthiness of the results provided by a Functional Element.4 It is not merely a backend data quality score; it is a crucial, user-facing tool designed to be presented alongside the FE's output, typically in a user interface or dashboard. Its purpose is to provide the end-user—the operator or engineer making the key decision—with immediate and unambiguous insight into the reliability of the information they are consuming.
4The standard recommends a "traffic light" visualisation (Green, Yellow, Red) where the criteria for each state are clearly defined, allowing a user to understand not only the current quality status but also the underlying reasons for any degradation.
4 This concept is built upon two distinct but integrated assessment processes: a continuous automated assessment and a periodic manual assessment.
4The combination of these two assessment types reveals a foundational principle of the DNV framework. The Quality Indicator is not purely a technical data validation metric; it is a socio-technical construct designed to build and maintain human trust in a complex digital system. The automated component provides constant, real-time vigilance over the data streams and computational models, ensuring the system's internal health. The manual component provides the essential layer of governance, accountability, and verification against physical reality. It assures the user that the digital twin is not an unmanaged "black box" but is actively maintained, calibrated, and kept in sync with the asset it represents. This principle dictates that any platform aiming for DNV compliance must provide robust mechanisms not only for automated data processing but also for formalising and integrating these critical human-in-the-loop workflows.
I.II.I - Continuous Automated AssessmentThe continuous assessment component of the QI is an automated process, typically executed by an algorithm, that continuously monitors factors that can degrade the trustworthiness of an FE's output in real time.
4 According to DNV-RP-A204, requirements 5.2.2-1 and 5.2.2-8, this automated monitoring must address a reduction in quality across several domains:
- Input Data Quality: This involves monitoring incoming data streams for a wide range of issues, including missing data, timeliness violations (latency), syntactic errors, and semantic inconsistencies such as values falling outside of expected physical ranges. The system must be able to detect these issues as they occur.4
- Computation Model Quality: Many FEs rely on computation models (from simple physics-based equations to complex machine learning models) to transform input data into decision-support information. The continuous assessment must monitor the quality of these models, such as detecting higher uncertainty in their output, instability, or performance degradation.4
- Digital Twin Infrastructure Health: The assessment must also cover the health of the underlying infrastructure. This includes detecting fault states in sensor systems, communication networks, or the digital twin platform itself that could compromise the integrity of the data or the FE's results.4
I.II.II Periodic Manual AssessmentThe periodic assessment component addresses critical factors that cannot be reliably or completely automated. It constitutes a formal, scheduled process of manual or semi-manual checks to ensure the digital twin remains a faithful representation of its physical counterpart over time.
4 DNV-RP-A204 requirements 5.2.2-10 and 5.2.2-11 specify that this process should include, but is not limited to, the assessment of:
- Physical and Digital Asset Modifications: Any change, planned or unplanned, to the physical asset (e.g., equipment replacement, process modification) or the digital asset (e.g., software updates) must be assessed for its impact on the FE. The periodic assessment verifies that these changes are correctly reflected and that the digital representation remains valid.4
- Data Quality Not Continuously Checkable: This category includes crucial maintenance activities that directly impact semantic data quality. A primary example is sensor calibration, which cannot be verified automatically. Other examples include changes to master data systems (e.g., updating equipment tags in an ERP system) that provide essential context to raw data streams.4
- Physical Inspections: Data from physical inspections of the asset can provide new information or detect failure modes (e.g., corrosion) that are not captured by online sensors. The results of these inspections must be incorporated into the quality assessment.4
- Computation Model Performance: The performance of computation models must be periodically re-validated against real-world data to ensure they continue to represent reality accurately and have not drifted over time.4
The results of these periodic assessments must be documented and used to update the state of the Quality Indicator, ensuring that the user-facing trustworthiness score reflects both the real-time data health and the long-term governance status.
4I.III - Foundational Requirements: Data Profiling, Cleansing, and GovernanceUnderpinning the Quality Indicator are foundational data management capabilities that DNV-RP-A204, Section 10, identifies as prerequisites for building a trustworthy digital twin system.
4- Data Quality Profiling: The standard requires the capability to perform data profiling, which is the process of analyzing datasets to understand their statistical characteristics, structure, and overall quality. This is a proactive measure to identify potential data quality issues before the data is consumed by critical applications. The goal is to understand the syntactic, semantic, and pragmatic quality of a dataset to determine its fitness for a specific purpose.4
- Data Cleansing: For FEs with high confidence levels, the system must possess robust and automated mechanisms for data cleansing. This is the process of detecting and correcting or removing defects and errors from data streams in real time. The objective is to improve the data quality to the required level of "pragmatic quality"—the degree to which the data is suitable and useful for its intended purpose.4 This is not an optional post-processing step but a required capability for robust, operational FEs.4
- Data Governance: A formal data governance framework is required to exercise authority and control over data assets. This includes defining clear policies, processes, and roles (such as data owners and stewards) to manage data quality throughout its lifecycle. Key aspects include ensuring accountability for data quality, transparency in how quality is managed and documented, and robust processes for controlling changes to data and applications.4