dltHub

Build with dlt.

Ship with dltHub Pro.

dlt is the open-source Python library 10,000s of developers use to build data pipelines. dltHub Pro is the agentic platform that deploys, monitors, and scales them. One command. No manual environment setup. No silent failures.

OPEN SOURCE

dlt

The AI-native Python library for data movement. Write any pipeline, run anywhere, no backend needed.

$pip install dlt
Docs9,700+ sources · Agent-friendly docs

IN PUBLIC PREVIEW

Launch May 2026

dltHub Pro

Agents build your dlt pipelines from a prompt. Pro deploys them to production with scheduling, alerting, and observability — one command, zero manual setup.

DocsAgent-friendly docs
dltHub Pro

"What I didn't expect is how much it unblocks the team. A mid-level engineer can spin up a prototype, browse the raw data in dltHub Pro's local DuckDB workspace, validate the SQL schema - all without pulling in a senior. That loop of prototype, inspect, fix, re-run - that's the real unlock."

Marcello Victorino

Marcello Victorino

Staff Data Engineer, Tasman Analytics

Agentic Workflows

Complete agentic workflows for every phase of data engineering

Not autocomplete, not a chatbot on a dashboard. A guided sequence of skills, commands, rules, and MCP - with guardrails agents can't skip. Maintained by dltHub, controlling the infrastructure agents and pipelines operate on.

REST API Pipelinedlt
dltHub RuntimePRO
Data Explorationdlt
TransformationsPRO
Foundationdlt

Agentic Workflows in Detail

Discover individual skills per agentic workflow

See how each workflow guides your agent - step by step, from first prompt to production deployment.

REST API Pipeline1/7

Find a dlt source for a given API or data provider. Use when the user asks about a source, wants to find a connector, or asks to implement a pipeline for a specific data source.

Sonnet 4.6 · REST API Pipeline · ~/pipelines

? for shortcuts

dlt is the open-source Python library 10M+ developers use to build data pipelines

10M+

PyPI downloads per month

8,000+

Companies running dlt in production

800+

Snowflake customers in production

Agent-deployed pipelines are growing 10x year over year

Pipelines built and shipped by AI agents are on an exponential trajectory. Human-authored pipelines still grow steadily — but agents are scaling production workloads 10x faster.

By agents
By humans
Early 2025
2026
10x more pipelines deployed by agents
Quotation mark icon{testimonial.author?.name}

The current machine learning revolution has been enabled by the Cambrian explosion of Python open-source tools that have become so accessible that a wide range of practitioners can use them. As a simple-to-use Python library, dlt is the first tool that this new wave of people can use. By leveraging this library, we can extend the machine learning revolution into enterprise data.

Quotation mark icon
Julien Chaumond
CTO/Co-Founder at Hugging Face
Quotation mark icon{testimonial.author?.name}

Python and machine learning under security constraints are key to our success. We found that our cloud ETL provider could not meet our needs. dlt is a lightweight yet powerful open source tool we can run together with Snowflake. Our event streaming and batch data loading performs at scale and low cost. Now anyone who knows Python can self-serve to fulfil their data needs.

Quotation mark icon
Maximilian Eber
CPTO & Co-Founder at Taktile

Learn agentic data engineering

Free, self-paced course. From first prompt to production deployment.

An image with the command "pip install "dlt[hub]" in the middle, and logos of REST API sources around it
DLTHUB CONTEXT

Made for LLMs: Data source to Live Reports in Python

dltHub Context is a hub of AI-native context assets, including skills, commands, hooks, AGENT.md, coding files and more, allowing you and an LLM to code any dlt pipeline from any REST API to any dlt destination - within minutes.

We already support more than 10,100 sources, and see a clear path toward hundreds of thousands. Go from writing pipeline code to ingesting data and delivering reports via Notebooks, all in one flow, with outputs tailored to data users.

Frequently Asked Questions

How is dltHub Pro different from a Claude skill or tools like Replit?

Tools like Claude skills or Replit are great for writing and running code. But they are not built for data engineering workflows end to end.

dltHub Pro gives your team complete agentic workflows that cover every phase: coding, running, deploying, and debugging pipelines, on infrastructure you control. Not just a skill, not just an editor, but a guided workflow from first line to production.

How is dlt different from Fivetran or a Python script that uses the request library?

dlt is the perfect match between standardization and customization. You get the automation that matters: schema inference, incremental state, normalization, and loading, while keeping the full flexibility and portability of plain Python.

And with agentic dltHub workflows, your team can code, run, deploy, and debug pipelines faster, with the reliability you can trust at every step.

What is dltHub?

dltHub is the managed platform for deploying and operating data pipelines built with dlt. It provides a runtime, observability, data quality checks, and collaboration features so teams can go from development to production with one command.

What is dlt?

dlt (data load tool) is an open-source Python library for building data pipelines. It lets you write any connector, run anywhere, and requires no backend. dlt is Apache 2.0 licensed and always free to use.