Paradime is the ultimate workspace for analytics on top of dbt. Get all your analytics work done from a single place.
Today we are lifting the curtains and introducing Paradime, the operating system for analytics, that we have been building for the past 2 years in stealth.
Building a startup in stealth is hard. It's harder than you can imagine. Hiring is hard, prospecting is hard. Talking to people about your product is hard. But it does help build and iterate a product without distractions alongside our design partners to firm up the proposition. So that's what we've been up to.
In the modern data stack, over the last few years, the proliferation of new point solutions has led to tool chaos. There are so many tools out there that it's impossible to keep track of them all. Data leaders are spending more time than they should trying to manage their tools, costs and vendors. They are tired of making round-trips with procurement. They're also sick of dealing with data sprawl across multiple vendors, which increases the risk of data breaches.
For the analytics engineer, the tool chaos has led to cognitive overload and a drop in productivity. Life has become harder than ever before.
As Benn summarized in his post The powder keg of the modern data stack that:
The biggest looming battle, however, will be over a different territory: The brain—or operating system
Back in 2018-19, at Octopus, my team was rebuilding the entire data stack from scratch. We ripped apart a brittle legacy stack consisting of SSIS, bespoke ingestion code written in .NET, SQL Server, Qlik etc.
The final solution that we ended up with after 6 months of grueling work looked something like below:
The community had not yet coined the term **Modern Data Stack.** It used to be called the **ELT** framework 😀.
On top of the tools architecture, when we added the people/team layer, we saw that:
Since 2019, the explosion of the Modern Data Stack has also exploded the tools that work on top of the warehouse. As these tools came into being, the stack became even more fragmented. The people and their attention connected to the tools also became fragmented and unbundled.
Every data analyst would spend most of their days bouncing between low-code/some-code/open-source / commercial apps, fighting fires and drowning in data requests when all they should be doing is generating RoI for the business.
An average Series C+ organization would have 70 employees in business functions per data analyst, which means on any given day these 70 employees would be going nuts that they don't get answers to their questions or requests and the poor analyst would be screaming in his head that he wanted to build valuable insights not respond to Slack all day long. This deadlock we saw first hand at Octopus, then at the Guardian, then at Revolut, Hubspot, Carta, and the list literally goes on.
We heard time and again from data analysts and analytics engineers that while dbt™* had liberated them from SQL hell, the explosion of tools have driven them back into hell-fire.
On the other side, we saw that business functions hunger for data and making data-driven business decisions that increased exponentially. Slack without data context was no longer fit for purpose for the data-driven enterprise.
The essential human to human conversation layer around data was clunky, time consuming and devoid of context.
We realized that the new world of dbt™* + the modern data stack needs a new category of tools to work with so people can work faster, smarter, and a lot less stressed.
The data analytics discipline has evolved massively in the last few years. There is a movement to bring software engineering principles to analytics. Analytics as a discipline is different from software engineering. There is code context, data context, and people context. Yet, as analysts, we are stuck with tools used by software engineers.
We are changing this status quo by bringing to the market the following:
If we take the blueprint for modern business intelligence from @A16Z below, each of the boxes in the diagram represent a category of tools of the Modern Data Stack. Each tool represents a source of compute, where something happens with the data. But then there are arrows connecting these boxes. These arrows represent people, processes, and productivity sinks.
It's like having really powerful processors on the motherboard while the bus system between them is limited in throughput.
Paradime is built to super-charge those analytics workflows, which today are either non-existent in most orgs or only present in businesses with significant resources to build internal tooling.
To that effect, today we are announcing 5 components of that workflow as explained below.
Shifting away from setting up and managing dbt™* workspaces on individual laptops, in Paradime you can onboard analysts in less than 3 mins once the account is set up by the admin. Admin account setup takes less than 30mins and does not require any engineering support. There is no 3 month implementation and professional services costs.
We support connecting to dbt™* repository on Github, BitBucket, and Gitlab. We support connecting to Redshift, BigQuery, Snowflake and Firebolt and more.
The Code IDE is the crown jewel of the Paradime experience.
The Paradime IDE brings best-in-class desktop IDE experience for analytics to the cloud. It’s fast, performant and has the widest coverage of features. It’s purpose built for analytics workflows compared to general purpose cloud IDEs such as Gitpod, AWS Cloud9, Stackblitz, which are more suited for software engineering.
It comes with all the ergonomics that developers expect from a desktop IDE, but some of the notable features include:
If you ever felt stuck with a dbt Cloud™* IDE or intimidated by the complexity of local setup, now you have a choice. We are taking the IDE experience to a whole new level with plenty of ground breaking features coming next year.
The problem we wanted to solve here was:
So we built the Paradime Graph lineage that updates with every commit in your dbt™* repo in real-time. It spans from your data sources all the way to your dashboards and reports. We support both Looker and Tableau with more coming on the way.
For Looker, we provide lineage across Views, Explores, Looks, Dashboards, and Schedules, providing you with an end-to-end view of your dbt+Looker lineage. Similarly, for Tableau we can link Data Sources, Worksheets, and Dashboards.
Teams are moving beyond dbt exposures , which are hard to maintain and does not provide any visibility on the components of the BI layer between dbt™* tables and BI dashboards.
The problem we wanted to solve here was:
In Paradime, you can author workflows using a simple YAML format that is git-tacked. You can set up notifications across Slack and email. Finally, you can view results from production runs in the UI. We also have an integration with AWS S3 so we can pipe all your data back to your own S3 bucket and you stay in control of your data.
We also have APIs to trigger your dbt™* schedules from Airflow, Dagster or Prefect and receive alerts when they are complete. The API gives platform teams more control to manage dependencies upstream and downstream of dbt™* schedules.
And did I tell you - we also have a one-click importer for all your dbt™* jobs from dbt Cloud™* to Paradime, so migrating does not feel like a crazy adventure.
Building a dbt™* model is most times the last step in the modeling process. Analysts spend significant time exploring data in the warehouse and fiddling with raw and compiled SQL of existing models to run their daily work. There are numerous back and forth between the IDE and SQL editor, copying and pasting, editing and replacing table names with refs.
The problem we wanted to solve here was:
The focus here being on flow and productivity.
During the evolution of OS for Apple, there were two schools of thought - Wozniak believed that it should be an open system so hobbyists and tinkerers could play with it and Steve Jobs, thought it should be a system that just works. Today, we love MacOS, because it just works.
During our research, we found that there are similarly two main personas of analysts / analytics engineers in the world today: