You help build data ingestion pipelines using dlt.
Use this when the user wants to extract data from APIs, databases, or files and load it into a warehouse or lakehouse.
dlt init <source_name> <destination_name>
Example:
dlt init sql_database duckdb
python <pipeline_script>.py
dlt pipeline <pipeline_name> info
dlt pipeline <pipeline_name> show
import dlt
@dlt.source
def my_api_source(api_key=dlt.secrets.value):
@dlt.resource(write_disposition="replace")
def customers():
response = requests.get("https://api.example.com/customers",
headers={"Authorization": f"Bearer {api_key}"})
yield response.json()
return customers
pipeline = dlt.pipeline(
pipeline_name="my_api",
destination="duckdb",
dataset_name="raw",
)
load_info = pipeline.run(my_api_source())
print(load_info)
@dlt.resource(write_disposition="merge", primary_key="id")
def orders(updated_at=dlt.sources.incremental("updated_at")):
params = {"since": updated_at.last_value}
response = requests.get("https://api.example.com/orders", params=params)
yield response.json()
from dlt.sources.sql_database import sql_database
source
## When to use it
When users need help creating, running, or managing Dlt pipelines and related tasks.
## What's included
- Scripts: none
- References: none
## Compatible agents
Likely compatible with general coding assistants and CLI-capable agents (Copilot, Codex, Gemini).
This skill has not been reviewed by our automated audit pipeline yet.