I. The New Analytics Engineering
1.1 Demonstrating Value and ROI
1.2 Rising Influence of Data Leaders
1.3 Unleashing the Power of AI
Users of AI-powered productivity features save $50k on average
III. Analytics Engineering Benchmarks IRL
3.1 Frequency of Data Pipeline Runs
Nearly 75% of projects run production jobs on a daily frequency
- 50% of data pipeline runs are scheduled to run daily
- 22% run several times a day
- 11% run monthly
- 7% run weekly
- 6% run several times a week
- 2% run hourly
- 2% are high frequency, i.e., scheduled in real-time or near real-time (e.g., every few minutes)
3.2 Testing
- Overall test coverage is low, with an average of only 8.7% of models across projects having at least one test.
- Test coverage — or the percentage of models with at least one test — is inversely related to project size:
In smaller projects with 1 to 50 models, an average 18.5% of those models, or nearly 1 in 5, had a test.
In medium-sized projects with 51 to 200 models, that average falls to 10%.
In projects with 201 to 500 models, coverage continues to drop to an average of 8.8%.
In projects with 500+ models, the average is 8.2%.
The same relationship holds for other test metrics, including percentage of models with tests other than non-null tests and percentage of columns with tests. Coverage drops in a stair-step fashion as project size increases, with the biggest drop-off in coverage coming once projects grow larger than 50 models.
Test coverage for fewer than 10% of models across projects
Across all projects in Paradime, only 8.7% of models have at least one test.
Another insight is that If we look for tests beyond fundamental non-null tests, the proportion of test adoption drops further. For example:
In small projects with between 1 to 50 models, an average of 18.5% of models had at least one test, but that average dropped to 13.9% for models with at least one test that was not a “not-null” test. That’s a significant drop, meaning that only roughly 14 in 100 models, on average, had testing beyond non-null tests.
Among projects with 201 to 500 models, an average 8.8% of models had at least one test, but an average 7.5% had at least one test that was not a “not-null” test.
Not-null tests, which gauge the proportion of data in specified columns that are not nulls (vs. null values), are often the first type of test implemented in a data project. They serve as a starting point for more comprehensive testing. As with overall coverage, testing beyond non nulls declines with project size, ranging from an average 14% of models in small projects to only 7% in extra-large projects.
Moreover, when looking at average column coverage, coverage also drops significantly as project size increases. In small projects, only 7% of columns have tests. In medium to extra-large projects, an average of 3.5% or fewer columns are tested.
Testing coverage declines with projects size, with a steep drop-off beyond 50 models
Data underscores the need for test generation assistance, particularly as projects grow in size and complexity.
Testing coverage declines steadily with project size, with only 7% of projects with more than 500 models including tests other than non-null tests.
- On average, a low percentage of models have any testing coverage at all, and the larger the project is, the lower the average test penetration (e.g., in projects with 500+ models, an average of only 8.2% of models have tests).
- Among models with tests, a significant proportion are running only non-null tests.
- Column coverage averages between 7% for small projects (those with up to 50 models) and 2% for extra-large projects (which have 500 or more models). This means that even in models with tests, only a fraction of the data is being validated.
3.3 Materialization Strategies
Our analysis of Paradime usage data reveals interesting patterns in how organizations leverage different materialization strategies. These strategies play a crucial role in optimizing data pipeline performance and resource utilization.
3.3.1 Incremental Models: Balancing Efficiency and Complexity
Incremental models update only new data rather than performing full rebuilds, resulting in meaningfully shorter runtimes. The strategy is particularly impactful with resource-intensive transformations since it allows for the transformation to focus only on new records.
Adoption of this strategy shows a nuanced pattern:
Small projects (1-50 models): 5% of models
Medium projects (51-200 models): 11% of models
Large projects (201-500 models): 14% of models
Extra-large projects (500+ models): 5% of models
Adoption of advanced materialization strategies increases with indicators of project maturity and complexity, like number of models – except in very large projects
In dbt™ projects in Paradime that have 201 to 500 models, ~14% of models are incremental
Incremental materialization is used in roughly 9% of all models across projects but in a greater proportion (14%) in large projects with between 201 and 500 models.
The pattern holds when employing a different indicator of project maturity and complexity — the number of maintainers on a project.
Adoption of advanced materialization strategies tends to increase with indicators of project maturity and complexity, like number of maintainers
In dbt™ projects in Paradime that have 11 to 15 maintainers, ~14% of models are incremental
Incremental materialization usage grows with indicators of project maturity and complexity, like the number of maintainers — up to a point, dipping lower for projects with the most maintainers.
This “inverted U” pattern reveals that while incremental models become more prevalent as projects grow, there’s a slight decrease in adoption for the largest projects.
This could be due to:
Complexity threshold: At extreme scales, the overhead of configuring and managing many incremental models might outweigh their benefits.
Alternative strategies: Very large projects might employ different optimization techniques or architectural patterns to manage costs and drive efficiency.
Diverse use cases: Extremely large projects might serve a wider variety of needs, some of which may not benefit from incremental models.
3.3.2 Snapshot Models: Tracking Historical Changes
- Small projects (1-50 models): 0.35% of models
- Medium projects (51-200 models): 0.85% of models
- Large projects (201-500 models): 1.27% of models
- Extra-large projects (500+ models): 1.85% of models
Snapshot usage increases with dbt™ project size
Nearly 2% of models in “extra-large” projects with 500 or more models use Snapshot
Snapshot usage is positively correlated with the number of models in a project.
3.3.3 Ephemeral Models
- Small projects (1-50 models): 0% of models
- Medium projects (51-200 models): 2.03% of models
- Large projects (201-500 models): 6.82% of models
- Extra-large projects (500+ models): 0.97% of models
- Average across projects: 2.6% of models
3.4 Optimizing Production Jobs
dbt run
as a whole with no selectors and no way of refining the set of commands they want to run,” says Alexandre Carvalho at Stenn. “That’s actually pretty bad practice. The implication is that you’re rerunning the whole project, and there are significant costs and time associated with that.”- The exclude flag, on its own, is essentially unused across projects with up to 500 models, and appears in only 4% of commands in extra-large projects with more than 500 models.
- In large projects, 1 in 10 commands are running without either flag. In extra-large projects with more than 500 models, 23% of jobs are running with neither flag.
Snapshot usage increases with dbt™ project size
Nearly 2% of models in “extra-large” projects with 500 or more models use Snapshot
Optimization tactics, like use of the select and exclude flags in commands, are more widely used in larger projects, but many jobs still run with neither flag.
- Small projects (1-50 models): 1.8 commands
- Medium projects (51-200 models): 1.2 commands
- Large projects (201-500 models): 3.3 commands
- Extra-large projects (500+ models): 2.4 commands
3.5 Terminal Commands
dbt run
, in particular, but also dbt compile
, dbt build
, and dbt test
. Together, these four terminal commands account for more than 90% of executed commands.dbt run-operation
, which invokes macros, only accounts for 1.6% of terminal commands. It’s difficult not to suggest that analytics engineers could be exploring less common features more aggressively.3.6 dbt™ Bloopers
dbt runoperation
(missing the dash in “run-operation”) and dbt seeds
(not plural!). Others less so: dbt json
does not closely resemble a command in dbt™ (or any other technology) we can think of. What’s surprising is the frequency with which some users are typing these commands. While dbt runoperation
generates an error, it has been tried more than 6,000 times. The mysterious dbt json? Nearly 2,000 times.Unrecognized command!?
dbt runoperation
over 6,000 times.