Decube
Try for free
  • 🚀Overview
    • Welcome to decube
    • Getting started
      • How to connect data sources
    • Changelog
    • Public Roadmap
  • 💻Security & Infrastructure
    • Overview
    • Deployment Methods
      • SaaS (Multi-Tenant)
      • SaaS (Single-Tenant)
      • Bring-Your-Own-Cloud (BYOC)
    • Data Policy
  • 🔌Data Warehouses
    • Snowflake
    • Redshift
    • Google Bigquery
    • Databricks
    • Azure Synapse
  • 🔌Relational Databases
    • PostgreSQL
    • MySQL
    • SingleStore
    • Microsoft SQL Server
    • Oracle
  • 🔌Transformation Tools
    • dbt (Cloud Version)
    • dbt Core
    • Fivetran
    • Airflow
    • AWS Glue
    • Azure Data Factory
    • Apache Spark
      • Apache Spark in Azure Synapse
    • OpenLineage (BETA)
    • Additional configurations
  • 🔌Business Intelligence
    • Tableau
    • Looker
    • PowerBI
  • 🔌Data Lake
    • AWS S3
    • Azure Data Lake Storage (ADLS)
      • Azure Function for Metadata
    • Google Cloud Storage (GCS)
  • 🔌Ticketing and Collaboration
    • ServiceNow
    • Jira
  • 🔒Security and Connectivity
    • Enabling VPC Access
    • IP Whitelisting
    • SSH Tunneling
    • AWS Identities
  • ✅Data Quality
    • Incidents Overview
    • Incident model feedback
    • Enable asset monitoring
    • Available Monitor Types
    • Available Monitor Modes
    • Catalog: Add/Modify Monitor
    • Set Up Freshness & Volume Monitors
    • Set Up Field Health Monitors
    • Set Up Custom SQL Monitors
    • Grouped-by Monitors
    • Modify Schema Drift Monitors
    • Modify Job Failure Monitors (Data Job)
    • Custom Scheduling For Monitors
    • Config Settings
  • 📖Catalog
    • Overview of Asset Types
    • Assets Catalog
    • Asset Overview
    • Automated Lineage
      • Lineage Relationship
      • Supported Data Sources and Lineage Types
    • Add lineage relationships manually
    • Add tags and classifications to fields
    • Field Statistcs
    • Preview sample data
  • 📚Glossary
    • Glossary, Category and Terms
    • Adding a new glossary
    • Adding Terms and Linked Assets
  • Moving Terms to Glossary/Category
  • AI Copilot
    • Copilot's Autocomplete
  • 🤝Collaboration
    • Ask Questions
    • Rate an asset
  • 🌐Data Mesh [BETA]
    • Overview on Data Mesh [BETA]
    • Creating and Managing Domains/Sub-domains
    • Adding members to Domain/Sub-domain
    • Linking Entities to Domains/Sub-domains
    • Adding Data Products to Domains/Subdomains
    • Creating a draft Data Asset
    • Adding a Data Contract - Default Settings
    • Adding a Data Contract - Freshness Test
    • Adding a Data Contract - Column Tests
    • Publishing the Data Asset
  • 🏛️Governance
    • Governance module
    • Classification Policies
    • Auto-classify data assets
  • ☑️Approval Workflow
    • What are Change Requests?
    • Initiate a change request
    • What are Access Requests?
    • Initiate an Access Request
  • 📋Reports
    • Overview of Reports
    • Supported sources for Reports
    • Asset Report: Data Quality Scorecard
  • 📊Dashboard
    • Dashboard Overview
    • Incidents
    • Quality
  • ⏰Alert Notifications
    • Get alerts on email
    • Connect your Slack channels
    • Connect to Microsoft Teams
    • Webhooks integration
  • 🏛️Manage Access
    • User Management - Overview
    • Invite users
    • Deactivate or re-activate users
    • Revoke a user invite
  • 🔐Group-based Access Controls
    • Groups Management - Overview
    • Create Groups & Assign Policies
    • Source-based Policies
    • Administrative-based Policies
    • Module-based Policies
    • What is the "Owners" group?
  • 🗄️Org Settings
    • Multi-factor authentication
    • Single Sign-On (SSO) with Microsoft
    • Single Sign-On (SSO) with JumpCloud
  • Export/Import
    • Export/Import Overview
  • Export for Editing existing objects
  • Export for Creating new objects
  • CSV Template Structure (Edit existing items)
  • CSV Template Structure (Add new items)
  • Importing Data (Edit existing items & Add new items)
  • History
  • ❓Support
    • Supported Features by Integration
    • Frequently Asked Questions
    • Supported Browsers and System Requirements
  • Public API (BETA)
    • Overview
      • Data API
        • Glossary
        • Lineage
        • ACL
          • Group
      • Control API
        • Users
    • API Keys
Powered by GitBook
On this page
  • Purpose
  • Accessing the Import Tab
  • Selecting Operation
  • Uploading a CSV
  • General Guidelines:
  • Important Warnings:
  • Smart Validation & Error Handling
  • Validation in Progress
  • Success
  • Fail

Importing Data (Edit existing items & Add new items)

Learn more about the Import functionality to quickly update metadata across the platform.

PreviousCSV Template Structure (Add new items)NextHistory

Last updated 2 days ago

Purpose

The Import feature allows users to bulk update existing metadata or add new glossary or policy items by uploading structured CSV files. This enables faster changes, improves data quality, and minimizes manual edits across Decube.

Accessing the Import Tab

Once import is submitted, changes cannot be undone.

Please verify all values and ensure you are uploading to the correct workspace/module.

To begin, navigate to the Import tab from the side navigation in Export/Import tab.

Selecting Operation

Each import operation begins by selecting:

  • Import Type: Choose between Edit Existing Items or Add New Items

  • Operation Type: Depending on your selection, the following options are supported

Import Type

Supported Operation Types

Edit Existing Items

Catalog (Dataset / Non-Dataset), Glossary, Policies

Add New Items

Glossary, Policies

Ensure that you select the specific object type and click on Upload CSV button.

After clicking on Upload CSV button, you wil be brought to the next page to upload your CSV.

Uploading a CSV

You can either:

  • Drag and drop your file into the upload area, or

  • Click on Upload file, to browse from your system.

Decube supports only .csv files. Ensure your file follows the required format outlined in the CSV Format Reference.

General Guidelines:

  • Each file should only contain data related to a single logical group of operation. For example:

    • One Glossary import (including glossary, categories, terms)

    • One Catalog source (e.g., Redshift → schema, tables, columns)

    • One object type group (e.g., all Dashboards or all Charts)

  • The file must contain the identifier columns for the selected operation type. Rows missing identifiers will be skipped or rejected.

Important Warnings:

These warnings will be displayed on Uploading a CSV screen according to the selected operation type.

Go through the warnings and make sure you are uploading the valid file.

Catalog Objects (Dataset & Non-Dataset)

  • File size must be under 100 MB and not exceed 10,000 rows.

  • Only one Data Owner allowed per object.

  • Maximum 3 tags per object.

  • Description length should not exceed 8,000 characters.

Glossary Objects (Edit)

  • File size must be under 100 MB and not exceed 10,000 rows.

  • Renaming of glossary, category, or term names is not allowed. Do it via UI.

  • Descriptions are mandatory for all glossary-related objects.

  • Each glossary can have up to 3 Data Owners.

  • Description limit: 8,000 characters.

Policies (Edit)

  • File size must be under 100 MB and not exceed 10,000 rows.

  • Do not rename existing policies. Use the platform UI instead.

  • Policy Classification Name must:

    • Be mandatory

    • Be unique

    • Be 5 characters or fewer

Glossary Objects (Add New Items)

  • File size must be under 100 MB and not exceed 10,000 rows.

  • Glossary names must be unique.

  • Category names must be unique within a glossary.

  • Term names must be unique within a category or glossary.

  • Descriptions are mandatory.

  • Character limits:

    • Name: 100 characters

    • Description: 8,000 characters

  • Maximum of 3 Data Owners per glossary.

Policies (Add New Items)

  • File size must be under 100 MB and not exceed 10,000 rows.

  • Policy Name and Policy Classification Name are required and must be unique.

  • Classification Name: max 5 characters.

Once file is uploaded, next will be Smart Validation & Error Handling.

Smart Validation & Error Handling

Once uploaded, your CSV file undergoes validation for:

  • File format

  • Required headers & structure

  • Object-level constraints (e.g., tag limits, owner limits, character limits)

  • Data type and identifier validation

Common validation errors include:

  • Missing required fields

  • Unknown or duplicate attributes

  • Unsupported object type or row structure

Validation in Progress

Once you upload a file the system begins validating your input.

What you will see:

  • A loader with Message: “We’re validating your file…”

  • Do not refresh or leave the page while validation is in progress.

  • Leaving the page may interrupt the upload process and you may need to restart the operation.

  • This process may take up to a minute depending on the file size and number of rows.

After validation finishes, there can be two scenarios

  • Success: If the file passes all checks, you’ll be directed to the success confirmation screen.

  • Fail: If any rows or structure within the file fail validation, you will see the error screen.

Success

What’s included:

  • Success message: “Your file has been validated and is ready for import”

  • Summary box: shows the summary of changes that will be executed on the platform.

  • Name the CSV: you can name the csv as per your preference.

  • Checkbox: You must acknowledge a disclaimer confirming the changes are irreversible.

  • Import button: Once confirmed, you can proceed with importing the validated metadata.

  • Cancel Upload: If you do not agree with change summary and you want to make changes to the file, you can cancel upload.

  • Once import is successful changes will be executed on the platform.

Before finalizing the import, users are required to acknowledge a critical confirmation that import operations are irreversible.

This step is designed to prevent accidental overwrites or updates, especially in environments where metadata quality and governance are crucial.

Fail

What’s included:

  • Error message: “There were issues with your file.”

  • Error summary box: will show the all the errors encountered while validation. For example:

    • Number of rows that failed validation

    • Types of errors detected (e.g., missing identifiers, invalid attribute values)

  • Retry button: Allows you to re-upload a corrected file.

Note: If validation fails during the initial upload, the import will not start, and no email will be sent. The user must correct the file based on the error summary shown on-screen and reinitiate the import manually.

  • Retry returns the user to the import screen to re-upload a corrected CSV file based on the issues shown in the error summary.

After a CSV file is validated and the user confirms the import, the system processes the file in the background. Upon completion, users receive an email notification informing them whether the import succeeded or failed.

If your email shows an error:

  • Trigger Condition: Sent when the CSV passes validation, but the import process fails due to issues such as system-level errors, unexpected attribute values, or conflicts during metadata update.

  • You will need to download the error report from the History tab.

About the Import Error Report:

  • No changes will be executed on the platform in import fail.

  • The error report is the same CSV file the user uploaded, but it includes inline error messages for each failed row.

  • Example: If row 10 had an invalid attribute, that row will be marked with the corresponding error.

  • Users only need to fix the rows with errors—they do not need to recreate the entire file.