Skip to main content

Data Quality Dashboard

The Catalog Data Quality Dashboard gives you a centralized view of data quality test results across your data assets. You can see which tables pass or fail tests, how much of your estate has coverage, and which tools are running those tests.

Catalog Does Not Run Tests

Catalog displays results from your existing data quality tools. It does not run tests itself. Configure integrations such as dbt, Soda, other Quality integrations, or the Quality API to push test results into Catalog.

Accessing the Dashboard

Open the Data Quality Dashboard at https://app.castordoc.com/data-quality.

Tabs and Filters

The dashboard has two tabs: Overview and Result Details. Use the filters at the top to narrow the data:

  • Warehouse: Filter by data warehouse
  • Database: Filter by database
  • Schema: Filter by schema
  • Table Tag: Filter by table tag
  • Owner: Filter by table owner
  • Result Status: Filter by outcome using pass, fail, or warning. Those map to the same states you see in the tables: pass is SUCCESS, fail is ALERT, and warning is WARNING.

Overview Tab

The Overview tab shows a high-level summary of data quality across your tables.

Data Quality Dashboard Overview tab with summary metrics, coverage charts, and table details.
Overview tab with summary metrics, coverage, and table-level test results

Tables Overall Status

This section summarizes the health of your tables:

  • Tables with all tests passed: Percentage of tables where every test is passing. For example, 60.7% means most of your tables are in good shape.
  • Tables by Status: A breakdown of tables by their current status:
    • ALERT: Tables with one or more failing tests
    • WARNING: Tables with tests in a warning state
    • SUCCESS: Tables where all tests pass
  • UTC Time - Latest tests: Timestamp of the most recent test run. Use this to see how fresh the data is.

Tables Coverage

The Tables Coverage chart shows what percentage of your tables have quality tests:

  • Has tests: Tables that have at least one quality test
  • No test: Tables with no quality tests

Use this to identify gaps in coverage and prioritize where to add tests.

Number of Tests by Technology

This chart shows how tests are distributed across the tools that run them. You will often see labels such as:

  • API: Tests pushed through the Catalog Quality API
  • Coalesce Quality: Tests from Coalesce Transform (built-in data quality)
  • dbt: Tests from dbt
  • Soda: Tests from Soda
  • Monte Carlo: Tests from Monte Carlo
  • Sifflet: Tests from Sifflet

Tables Details

The Tables Details table lists individual tables and their test performance. Columns include:

  • Database . Schema: Location of the table (for example, NY_TAXI . GOLD)
  • Table: Table name
  • Popularity: Usage score based on query activity
  • Tests Failing: Count of failed tests (highlighted in red)
  • Tests Warning: Count of tests in warning state (highlighted in yellow)
  • Tests Passing: Count of passed tests (highlighted in green)

Use this table to identify tables that need attention and drill into specific assets.

Result Details Tab

The Result Details tab shows a granular view of individual test results. Use it to see which specific tests failed, when they ran, and what the outcome was. This helps you investigate issues and track resolution over time.

Data Quality Dashboard Result Details tab with Test Result Details table showing database, scope, test name, run time, and status.
Result Details tab with individual test results

Test Result Details Table

The Test Result Details table lists each test run with the following columns:

  • Database . Schema . Table: Full path of the data asset being tested (for example, AUSTIN_TEST_DB . PROD . EDW_PROD . WRK_CUSTOMER_LOYALTY)
  • Scope: Whether the test applies to a specific column (for example, Column: E_MAIL) or the entire table (Table)
  • Name: The name of the data quality rule or test (for example, E_MAIL: Unique, quantity > 0, Includes Today's Data)
  • Test run time: When the test was last executed
  • Status: Outcome of the test: ALERT (failed), WARNING, or SUCCESS (passed)

Use the cloud icon or the three-dot menu at the top right of the table to export or download the results.

Data Quality at the Asset Level

You can also see data quality when browsing individual tables. Click the table name to view the Data Quality.

  • Overall quality status for that table
  • Individual test results (passed, failed, warning)
  • Test execution timestamps
  • Historical test performance

The Overview and Result Details tabs summarize estate-wide coverage and recent runs. They are not week-over-week trend charts. For longer historical context on one table, click on the table and go to the Data Quality tab.

Setup and Configuration

Enabling Data Quality Integration

To surface test results in the dashboard, connect the tool that runs your tests. For Coalesce-built pipelines, work through the following in order.

  1. Coalesce Transform and Coalesce Quality: Configure a Warehouse integration in Catalog first. Without it, the first Coalesce Transform ingestion cannot complete. Then create a Coalesce Quality source.
  2. External tools: Use the Quality API to push test results to Catalog.
  3. dbt, Soda, or other partners: Configure the matching integration under Quality integrations.

Visibility

Data quality visibility follows asset visibility. If a table or schema is hidden in Catalog, its quality tests will not appear in the dashboard. Quality information uses the same permission model as data assets.

Troubleshooting

Quality Tests Not Appearing

  • Verify the integration is configured correctly (dbt, Soda, API, Coalesce Transform plus Coalesce Quality source, or another quality tool).
  • Check that the source tables and schemas are visible in Catalog.
  • Ensure API credentials have the right permissions if you use the Quality API.

Missing Test Results

  • Confirm tests are running in your quality tool.
  • Check API integration logs for errors.
  • Verify test metadata is being sent correctly to Catalog.

What's Next?