Technology that is used to monitor an organization's application layer for API failures and potential compute issues. Typically, APM tools are not built to monitor data or infrastructure layers nor do they validate the quality of data pipelines.
Process of monitoring an organization's application layer within an enterprise infrastructure for potential issues and downtime.
Data that has not been used recently (or ever) by an organization. Creates internal confusion and leads to higher data storage costs.
Centralized inventory of data assets across technologies and environments. Enables users to easily navigate and search for data assets.
Process of identifying and categorizing similar data assets in order to provide greater context about an organization's data. For example, classifying sensitive data or business data.
Problems that arise from an organization's expanding data volumes and related processes, technology, users, and use cases. Failure to mitigate data complexity can lead to inefficient data operations, inflated costs, unreliable data, and system downtime.
Process of finding, exploring, and validating data. Data discovery may involve the use of a data catalog, especially for organizations with many data assets and data sources.
Unexpected content changes that can negatively impact an organization's processes, especially its AI/ML workloads. A consistent increase in missing values within a data set, for example, may indicate data drift.
Utilizing data and data infrastructure in a productive way that aligns with the organization's specific needs. Achieved by minimizing redundant and underutilized data, ineffective data management practices, and resource contention.
Software application that leverages analytics and ML/AI to improve reliability, scalability, and costs across an organization's data, pipelines, and workloads. Provides visibility into the health and performance of all aspects of enterprise data systems.
Period of downtime during which users and/or downstream applications are unable to access certain data assets. May be caused by a variety of situations, such as resource contention, structural changes, or system health issues.
End-to-end visibility into the flow and cost of data across an organization's interconnected systems.
Processes and technology used to ingest data from source systems into an organization's data ecosystem.
Process of crawling, analyzing, and summarizing data in a way that helps users understand the organization's data.
Measurement of a data set's overall health for its intended use. Minimizing or eliminating missing and incorrect data are key aspects of ensuring data quality.
Process of ensuring that data has arrived as expected during its movement from point A to point B.
Ensuring the dependable delivery of quality data in an uninterrupted, on-time schedule. Data reliability is essential for building trust with business users.
Return on investment realized by leveraging an organization's data. Calculated by subtracting data's estimated return minus its total cost, which includes the sum of costs pertaining to data storage, compute, pipelines, and related systems. The net return is then divided by the total cost, multiplied by 100 and expressed as a percentage.
Effective use of data to achieve an organization's business goals and support its use cases, such as BI reporting, data applications, embedded analytics, AI/ML workloads.
Derogatory term used to describe an organization's data when it is siloed or generally disorganized. Guaranteeing data quality and accessibility becomes difficult—if not impossible—when data swamps exist, thereby eroding end user confidence in organizational data and leading to an even murkier data swamp.
Process of ensuring that data follows and conforms to the schema definition, follows business rules, and is accurate and usable.
Tangible and intangible costs that can be attributed to a company's inefficient storage and utilization of data.
A historical accounting of data's journey from its original data source to present day usage, including any dependencies and connected assets.
An abbreviation for "mean time to resolution," a metric that measures a team's responsiveness to resolving issues. (Sometimes referred to as "mean time to recovery.") Calculated by dividing the sum of all time required to resolve issues during a given period by the total number of incidents during the same period. Data teams should strive for low MTTRs.
Data that helps an organization understand the reliability, scale, and cost of its data, processing, and pipelines. Used to predict, prescribe, prevent, troubleshoot, optimize, and contextualize.
Acquiring or deploying more of a particular resource (storage, compute, etc.) than what is actually necessary to support an organization's current needs. Often occurs as a safeguard to protect against unexpected changes in demand.
Structural changes to schemas and tables, such as the addition or deletion of a column, that can break pipelines or impact downstream applications.
KPIs that measure a service provider's adherence to targets set forth in a company's SLA. Examples of data-related SLIs include data pipeline uptime percentage and average response time.
Specific targets that are defined by an SLA and agreed to by key stakeholders within an organization. Data-related SLOs commonly relate to system availability and service provider responsiveness.