Data Quality Score

Definition

A quantitative measure of data fitness for its intended use, typically assessed across dimensions including accuracy, completeness, consistency, timeliness, uniqueness, and validity. Data quality scores enable organisations to monitor and improve the reliability of their data assets, prioritise remediation efforts, and establish trust in analytical outputs. High data quality is a prerequisite for effective AI and machine learning, and poor data quality is estimated to cost organisations 15-25% of revenue through flawed decision-making.

Complementary Terms

Concepts that frequently appear alongside Data Quality Score in practice.

ESG Score

A quantitative rating assessing a company's performance and risk exposure across environmental, social, and governance criteria, typically assigned by specialist rating agencies such as MSCI, Sustainalytics, and S&P Global. ESG scores increasingly influence investment decisions, cost of capital, and regulatory compliance, and are becoming a material factor in business valuations and due diligence.

Data Pipeline

An automated sequence of data processing steps that extracts, transforms, and loads data from source systems into target systems for analysis, reporting, or machine learning model training. Well-architected data pipelines are critical infrastructure assets that enable data-driven decision-making and AI deployment, and their reliability directly impacts downstream business processes.

Master Data Management (MDM)

The processes, governance, policies, and technology used to ensure that an organisation's critical shared data entities — such as customers, products, suppliers, and accounts — are accurate, consistent, and controlled across all systems and business units. MDM creates a single trusted source of master data, reducing duplication, resolving conflicts, and enabling reliable reporting and analytics.

Quality of Earnings (QoE) Report

A detailed financial analysis, typically prepared by an accounting firm on behalf of a buyer or lender, that assesses the sustainability, accuracy, and adjustability of a target company's reported earnings. A QoE report examines revenue recognition policies, non-recurring items, related-party transactions, working capital normalisation, pro forma adjustments, and the bridge from reported EBITDA to adjusted EBITDA.

Net Promoter Score (NPS)

A customer loyalty metric derived from a single survey question asking respondents how likely they are to recommend a company, product, or service on a scale of zero to ten. NPS is widely used as a proxy for customer relationship quality and brand strength, both of which are critical intangible assets influencing long-term enterprise value.

Synthetic Data

Artificially generated data that mimics the statistical properties of real-world datasets, used to train machine learning models when actual data is scarce, sensitive, or expensive to obtain. Synthetic data enables AI development in privacy-constrained domains such as healthcare and finance, while reducing data acquisition costs and regulatory exposure.

Training Data

The dataset used to train a machine learning model, comprising examples from which the model learns patterns, relationships, and decision boundaries. High-quality, proprietary training data is a significant competitive advantage and intangible asset, particularly in regulated industries where data scarcity creates barriers to entry.

Data Lake

A centralised repository that stores large volumes of raw data in its native format — structured, semi-structured, and unstructured — until it is needed for analysis. Unlike data warehouses, which store data in predefined schemas, data lakes use a schema-on-read approach that provides flexibility for diverse analytical workloads including machine learning, real-time analytics, and ad hoc exploration.

Put this knowledge to work

Use Opagio's free tools to measure and grow the intangible assets that drive your business value.