Loading Data from Oracle Database
to BigQuery
Using dlt
in Python
Join our Slack community or book a call with our support engineer Violetta.
Oracle Database
is the first database designed for enterprise grid computing, offering a flexible and cost-effective way to manage information and applications. BigQuery
is a serverless, cost-effective enterprise data warehouse that operates across clouds and scales with your data. This guide explains how to load data from Oracle Database
to BigQuery
using the open-source Python library dlt
. The dlt
library simplifies the process of data extraction, transformation, and loading, making it easier to integrate these two powerful platforms. For more information on Oracle Database
, visit Oracle's official page.
dlt
Key Features
- Governance Support:
dlt
pipelines offer robust governance support through metadata utilization, schema enforcement, and schema change alerts. Read more - Transform the data using SQL: Use the
dlt
SQL client to query and transform data with SQL statements. Learn more - Using dbt for transformations: Integrate dbt into your pipeline to structure transformations into DAGs with cross-database compatibility. Find out more
- Using Pandas for transformations: Fetch query results as Pandas data frames and perform transformations using Pandas functionalities. Explore this feature
- Scaling and Finetuning:
dlt
offers several mechanisms and configuration options to scale up and finetune pipelines, including parallel execution and memory buffer adjustments. Read more
Getting started with your pipeline locally
0. Prerequisites
dlt
requires Python 3.8 or higher. Additionally, you need to have the pip
package manager installed, and we recommend using a virtual environment to manage your dependencies. You can learn more about preparing your computer for dlt in our installation reference.