Tutorial: Cloning data from Oracle Database to BigQuery
Objectives
Create and run a Cloning job
Configure the target
Start the job
Monitor the job
View the results
Clean up
This guide will walk you through deploying a job that continuously clones data changes from an Oracle database to a BigQuery dataset using our Cloning Service.
Objectives
In this tutorial, you will:
Configure your Oracle database to allow supplemental logging.
Create and execute a Cloning job.
View the results in BigQuery.
Create and run a Cloning Service job
Create the job In the Algoreus UI, click the menu and navigate to the Clone page.
On the Create new Cloning job page, specify a Cloning job Name. Optionally, provide a description.
Click Next.
Configure the source:
Select Oracle as the source.
For Connectivity Method, if your Oracle server allows incoming traffic from public IPs, choose IP allowlisting. Otherwise, for Private Connection Name choose Private connectivity (VPC peering) and enter the VPC peering name.
In the credentials section, enter your username and password for accessing the Oracle Server.
Leave all other properties as is.
Click Next. If the connection is successful, a list of tables under sample schema will appear. For this tutorial, select all of the tables and events to clone (including Insert, Update, and Delete events).
Click Next.
Configure the target
Select the BigQuery target.
The Project ID and Service Account Key are automatically detected. Keep the default values as is.
(Optional) In the Advanced section, you can configure the staging bucket's name and location, the load interval, the staging table prefix, and the behavior when tables or databases are dropped.
Click Next.
(Optional) Configure the advanced properties. For this tutorial, you can use the default settings.
Click Next.
On the Review assessment page, click View mappings by any of the tables to get an assessment of schema issues, missing features, or connectivity issues that might occur during Cloning. If any issues occur, they must be resolved before you can proceed. For this tutorial, if any of the tables have issues, go back to the step where you selected tables and instead select a table or event (Inserts, Updates, or Deletes) without issues.
Click Next.
Review the summary Cloning job details, and then click Deploy Cloning job.
Start the job
On the Cloning job details page, click Start.
The Cloning job transitions from Provisioning to Starting to Running state. In the running state, the Cloning job loads an initial snapshot of the table data that you selected into BigQuery. In this state, the state of the table is listed as Snapshotting. After the initial snapshot is loaded into BigQuery, any changes made to the table are cloned to BigQuery, and the state of the table is listed as Cloning.
Monitor the job
You can start and stop the Cloning job, review its configuration and logs, and monitor your Cloning job.
You can monitor Cloning job activities from the Cloning job details page.
From the Cloning page, click the desired Cloning job Name.
Click Monitoring.
View the results
The Cloning job creates a replicated dataset and table in BigQuery, with names inherited from the corresponding Oracle database and table names.
If your source database name does not comply with the BigQuery dataset naming conventions, or if you have naming conflicts when you Clone multiple datasets into a single BigQuery dataset, specify aliases in the BigQuery target properties.
Open BigQuery in the console.
In the left panel, click on your project name to expand a list of datasets.
Select the dataset, and then select a table to view.
Clean up
To avoid incurring charges to your Cloud account for the resources used in this tutorial, either delete the project that contains the resources, or keep the project and delete the individual resources.
Last updated