Available Monitor Types
Comprehensive guide to all monitor types available in Decube's data quality platform.
Decube offers comprehensive monitoring capabilities to ensure data quality across your entire data infrastructure. Choose the right monitor type based on your specific data quality requirements.
Quick Selection Guide
Monitor Type
Best For
Setup Time
Technical Level
Schema Drift
Preventing pipeline breaks
Auto-enabled
Beginner
Freshness
Data dependencies
5 minutes
Beginner
Volume
Data load validation
5 minutes
Beginner
Field Health
Column-level data quality
10 minutes
Intermediate
Custom SQL
Complex business rules
15 minutes
Advanced
Job Failure
ETL pipeline monitoring
Auto-enabled
Beginner
Table-Level Monitors
Schema Drift β‘ Auto-Enabled
Purpose: Automatically detects structural changes to your tables and columns Value: Prevents downstream application failures and pipeline breaks
This monitor is enabled automatically for all tables when you connect a data source. It detects:
Table or column additions/deletions
Data type changes
Schema modifications that may cause compatibility issues
Use Cases:
π¨ Alert when critical table structures change
π‘οΈ Protect downstream applications from schema breaks
π Track data warehouse evolution over time
Freshness π
Purpose: Tracks when tables were last updated to ensure data timeliness Value: Ensures dashboards and reports have current data
Our ML-powered freshness monitors learn your table update patterns and alert when data becomes stale based on historical patterns.
Use Cases:
β° Critical for real-time dashboards
π Essential for daily business reports
π Monitor ETL pipeline completion
Volume π
Purpose: Monitors row count changes and detects data load anomalies Value: Catches missing or incomplete data loads before they impact business
Volume monitors establish baselines for expected row counts and alert when insertions fall below or exceed normal ranges.
Use Cases:
π Detect missing data loads
π Identify unexpected data spikes
π Monitor ETL pipeline data volumes
Column-Level Monitors
Field Health π
Purpose: Validates data quality at the column level with comprehensive tests Value: Ensures clean, consistent data for analytics and ML models
Available Field Health Tests:
Data Completeness Tests
Null Checks
Null: Monitors null values with configurable thresholdsAbsolute mode: Set threshold by row count (e.g., "fail when > 100 null rows")
Percentage mode: Set threshold by percentage (e.g., "fail when > 5% nulls")
Auto mode: Let ML learn acceptable null patterns (scheduled only)
Not Null: Validates no null values exist (fails if any nulls found)
Use Case: Essential for mandatory fields and columns with dependencies
Uniqueness Validation
Unique: Monitors duplicate values with configurable thresholdsAbsolute mode: Set threshold by row count (e.g., "fail when > 10 duplicate rows")
Percentage mode: Set threshold by percentage (e.g., "fail when < 95% unique")
Auto mode: Let ML learn acceptable uniqueness patterns (scheduled only)
Is Unique: Ensures all values are distinct (fails if any duplicates found)
Use Case: Critical for primary keys and unique constraints
Data Range & Distribution Tests
Statistical Validation
Average: Monitors average values against expected rangesMin/Max: Validates minimum and maximum value boundariesCardinality: Tracks distinct value counts (high/medium/low classification)
Use Case: Detect outliers and validate business rule compliance
String Validation
String length: Validates minimum/maximum string lengths using range thresholdsEmail: Monitors invalid email addresses with configurable thresholdsAbsolute mode: Set threshold by row count (e.g., "fail when > 50 invalid emails")
Percentage mode: Set threshold by percentage (e.g., "fail when > 2% invalid")
Auto mode: Let ML learn acceptable patterns (scheduled only)
UUID: Monitors invalid UUIDs with configurable thresholds (supports Absolute, Percentage, Auto modes)Regex Match: Custom pattern matching validation with configurable thresholdsPattern is specified separately from threshold
Threshold controls how many non-matching values trigger an incident
Supports Absolute, Percentage, and Auto modes
Advanced Monitoring
Custom SQL π§
Purpose: Create custom validation logic for complex business rules Value: Monitor sophisticated data relationships and business-specific requirements
Write custom SQL queries to validate complex business logic. Incidents trigger when your query returns any rows (row_count > 0).
Common Use Cases:
π° Revenue reconciliation between systems
π Cross-table data consistency checks
π Complex business rule validation
βοΈ Data quality SLA monitoring
Example Applications:
Set Up Custom SQL MonitorsJob Failure Monitoring βοΈ Auto-Enabled
Purpose: Automatically monitors ETL pipeline and data job execution Value: Ensures data transformation processes complete successfully
Job Failure monitors are automatically created when you connect ETL-type sources. They track failed jobs to identify processing issues.
Key Features:
π Auto-configured for dbt, Airflow, and other ETL tools
π§ Customizable alert channels
π Historical failure tracking
β‘ Real-time failure detection
Enhanced Monitoring Features
Grouped-By Monitoring π―
Purpose: Segment monitoring by dimension values for granular insights Value: Monitor data quality at the business logic level
Track data quality metrics segmented by specific column values (e.g., by region, customer type, or product category).
Applications:
π Monitor data quality by geographic region
π₯ Track customer data completeness by segment
π¦ Validate product data by category
π’ Ensure compliance by business unit
Getting Started Recommendations
For Data Engineers π¨βπ»
Start with Freshness & Volume for critical tables
Add Custom SQL for complex validation logic
Use Field Health for data pipeline validation
For Governance Teams π‘οΈ
Begin with Schema Drift monitoring (auto-enabled)
Implement Field Health for compliance requirements
Set up Grouped-By monitoring for business segments
For Platform Admins π€
Configure Job Failure monitoring for all ETL processes
Set up Volume monitoring for data load validation
Establish Freshness SLAs for business-critical data
Next Steps: Ready to get started? Visit our Enable Asset Monitoring guide to begin setting up your first monitors.
Last updated