Governed data democracy for your Databricks cloud
dlt is the most popular open-source Python library for data movement. With dlt you empower your Pythonic data platform teams to...
- add custom data sources,
- enhance your Iceberg/Delta data lakes and Unity Catalog,
- augment your AI workflows and AI tooling

Bring long tail of custom sources to Databricks
Empower your team to easily get data from the long tail of custom sources to Databricks. You can run dlt pipeline directly from Databricks notebooks. No need to specify additional credentials. Just load the data within a couple of minutes and use them in ML workflows right away.
Governance via the Unity Catalog integration
Effortlessly integrate dlt with Databricks Unity Catalog for unified data governance and simplified access control. dlt works in harmony with Unity Catalog as well as Delta Lake and Iceberg to ensure secure and organized data workflows.

Unlocked AI innovation with great data
Unlock AI innovation by adding dlt to internal and external agentic workflows. Have your engineers build custom sources with dlt and AI code editors such as Cursor or Continue. Augment your data engineering team with our dlt+ agents.
Our tests have proven that dlt meets our requirements regarding performance, customization and data privacy. It fits easily into our existing hosting and security infrastructure making the production rollout cost effective. The team behind dlt has a culture of supporting enterprise customers and we were able to get help and advice quickly.
- Erling Brandvik
- Sparebank 1 SR-Bank
Data democracy for our product, business, and operation teams means that they can independently satisfy a majority of their data needs through no-code self-service. If the teams want to build anything else to push the company forward, they don't need to wait for permission or data access to do it. All kinds of new reporting is being done that wasn't possible before.
- Alex Butler
- Harness
UPGRADE YOUR TEAM PRODUCTIVITY
Try our framework for running dlt pipelines in production
Provide high-quality data for your LLMs
Accelerate your AI workflows with dlt+, delivering refined, business-level data. dlt+ brings industry standards to your data pipelines, providing robust data lineage, data contracts, and data governance.
Deployed anywhere Python runs, dlt+ integrates seamlessly with AI code agents, making it easier than ever to build, refine, and monitor your data flows. With dlt+, these agents gain full understanding of your business-level data, helping you achieve deeper insights faster.

Case Studies
See all
How Flatiron Health used dlt to accelerate privacy-enhancing data processing
Learn how Flatiron Health cut 50% of their cost of ingestion and transformation pipelines using dlt.
How insurance company Dentolo democratizes data access using dlt
How Dentolo transformed its data ingestion process by adopting dlt empowering the team with a composable data stack and democratizing data access across the organization.
PostHog leverages dlt to offer their users a scalable and inexpensive one-click data warehouse
How PostHog adopted dlt to build a scalable, customizable data warehouse that seamlessly handles large datasets, empowering their team to deliver a flexible and high-performing solution for users.
Augment the AI workflows and tools of your data engineering team
Utilize dlt with AI code editors to create dlt pipelines seamlessly
Empower your team to create customisable production-grade data pipelines within minutes.

Use the dlt+ Transform Agent to generate trusted business-level data
The agent helps your team to automatically generate dbt and sqlmesh models based on dlt pipelines.

Run 100s of dlt pipelines with our dlt+ DataOps Agent
The agent utilizes OpenTelemetry tracing data to dramatically reduce the costs of running pipelines.
