Video Tutorial - Integrate Google BigQuery data in SSIS
This video covers following and more so watch carefully. After watching this video follow the steps described in this article.
- How to download / install required driver for
Google BigQuery integration in SSIS - How to configure connection for
Google BigQuery - Features about
API Source (Authentication / Query Language / Examples / Driver UI) - Using
Google BigQuery Connection in SSIS
Prerequisites
Before we perform steps listed in this article, you will need to make sure following prerequisites are met:- SSIS designer installed. Sometimes it is referred as BIDS or SSDT (download it from Microsoft site).
- Basic knowledge of SSIS package development using Microsoft SQL Server Integration Services.
- Make sure ZappySys SSIS PowerPack is downloaded and installed (download it). Check Getting started section for more information.
- Optional (If you want to Deploy and Schedule ) - Deploy and Schedule SSIS Packages
How to read data from Google BigQuery in SSIS (Export data)
In this section we will learn how to configure and use Google BigQuery Connector in API Source to extract data from Google BigQuery.
-
Begin with opening Visual Studio and Create a New Project.
Select Integration Service Project and in new project window set the appropriate name and location for project. And click OK.
-
In the new SSIS project screen you will find the following:
- SSIS ToolBox on left side bar
- Solution Explorer and Property Window on right bar
- Control flow, data flow, event Handlers, Package Explorer in tab windows
- Connection Manager Window in the bottom
Note: If you don't see ZappySys SSIS PowerPack Task or Components in SSIS Toolbox, please refer to this help link. -
Now, Drag and Drop SSIS Data Flow Task from SSIS Toolbox. Double click on the Data Flow Task to see Data Flow designer.
-
From the SSIS toolbox drag and API Source (Predefined Templates) on the data flow designer surface, and double click on it to edit it:
-
Select New Connection to create a new connection:
-
Use a preinstalled Google BigQuery Connector from Popular Connector List or press Search Online radio button to download Google BigQuery Connector. Once downloaded simply use it in the configuration:
Google BigQuery -
Proceed with selecting the desired Authentication Type. Then select API Base URL (in most cases default one is the right one). Finally, fill in all the required parameters and set optional parameters if needed. You may press a link Steps to Configure which will help set certain parameters. More info is available in Authentication section.
Steps to get Google BigQuery Credentials
This connection can be configured using two ways. Use Default App (Created by ZappySys) OR Use Custom App created by you.
To use minimum settings you can start with ZappySys created App. Just change UseCustomApp=false on the properties grid so you dont need ClientID / Secret. When you click Generate Token you might see warning about App is not trusted (Simply Click Advanced Link to expand hidden section and then click Go to App link to Proceed). To register custom App, perform the following steps (Detailed steps found in the help link at the end)- Go to Google API Console
- From the Project Dropdown (usually found at the top bar) click Select Project
- On Project Propup click CREATE PROJECT
- Once project is created you can click Select Project to switch the context (You can click on Notification link or Choose from Top Dropdown)
- Click ENABLE APIS AND SERVICES
- Now we need to Enable two APIs one by one (BigQuery API and Cloud Resource Manager API).
- Search BigQuery API. Select and click ENABLE
- Search Cloud Resource Manager API. Select and click ENABLE
- Go to back to main screen of Google API Console
Click OAuth consent screen Tab. Enter necessary details and Save.
- Choose Testing as Publishing status
- Set application User type to Internal, if possible
- If MAKE INTERNAL option is disabled, then add a user in Test users section, which you will use in authentication process when generating Access and Refresh tokens
- Click Credentials Tab
- Click CREATE CREDENTIALS (some where in topbar) and select OAuth Client ID option.
- When prompted Select Application Type as Desktop App and click Create to receive your ClientID and Secret. Later on you can use this information now to configure Connection with UseCustomApp=true.
- Go to OAuth Consent Screen tab. Under Publishing Status click PUBLISH APP to ensure your refresh token doesnt expire often. If you planning to use App for Private use then do not have to worry about Verification Status after Publish.
Google BigQueryUser Account [OAuth]https://www.googleapis.com/bigquery/v2Required Parameters UseCustomApp Fill in the parameter... ProjectId (Choose after [Generate Token] clicked) Fill in the parameter... DatasetId (Choose after [Generate Token] clicked and ProjectId selected) Fill in the parameter... Optional Parameters ClientId Fill in the parameter... ClientSecret Fill in the parameter... Scope Fill in the parameter... RetryMode Fill in the parameter... RetryStatusCodeList Fill in the parameter... RetryCountMax Fill in the parameter... RetryMultiplyWaitTime Fill in the parameter... Job Location Fill in the parameter... Redirect URL (Only for Web App) Fill in the parameter... Steps to get Google BigQuery Credentials
Use these steps to authenticate as service account rather than Google / GSuite User. Learn more about service account here Basically to call Google API as Service account we need to perform following steps listed in 3 sections (Detailed steps found in the help link at the end)Create Project
First thing is create a Project so we can call Google API. Skip this section if you already have Project (Go to next section)- Go to Google API Console
- From the Project Dropdown (usually found at the top bar) click Select Project
- On Project Propup click CREATE PROJECT
- Once project is created you can click Select Project to switch the context (You can click on Notification link or Choose from Top Dropdown)
- Click ENABLE APIS AND SERVICES
- Now we need to Enable two APIs one by one (BigQuery API and Cloud Resource Manager API).
- Search BigQuery API. Select and click ENABLE
- Search Cloud Resource Manager API. Select and click ENABLE
Create Service Account
Once Project is created and APIs are enabled we can now create a service account under that project. Service account has its ID which looks like some email ID (not to confuse with Google /Gmail email ID)- Go to Create Service Account
- From the Project Dropdown (usually found at the top bar) click Select Project
- Enter Service account name and Service account description
- Click on Create. Now you should see an option to assign Service Account permissions (See Next Section).
Give Permission to Service Account
By default service account cant access BigQuery data or List BigQuery Projects so we need to give that permission using below steps.- After you Create Service Account look for Permission drop down in the Wizard.
- Choose BigQuery -> BigQuery Admin role so we can read/write data. (NOTE: If you just need read only access then you can choose BigQuery Data Viewer)
- Now choose one more Project -> Viewer and add that role so we can query Project Ids.
- Click on Continue. Now you should see an option to Create Key (See Next Section).
Create Key (P12)
Once service account is created and Permission is assigned we need to create key file.- In the Cloud Console, click the email address for the service account that you created.
- Click Keys.
- Click Add key, then click Create new key.
- Click Create and select P12 format. A P12 key file is downloaded to your computer. We will use this file in our API connection.
- Click Close.
- Now you may use downloaded *.p12 key file as secret file and Service Account Email as Client ID (e.g. some_name@some_name.iam.gserviceaccount.com).
Manage Permissions / Give Access to Other Projects
We saw how to add permissions for Service Account during Account Creation Wizard but if you ever wish to edit after its created or you wish to give permission for other projects then perform forllowing steps.- From the top Select Project for which you like to edit Permission.
- Go to IAM Menu option (here)
Link to IAM: https://console.cloud.google.com/iam-admin/iam - Goto Permissions tab. Over there you will find ADD button.
- Enter Service account email for which you like to grant permission. Select role you wish to assign.
Google BigQueryService Account (Using Private Key File) [OAuth]https://www.googleapis.com/bigquery/v2Required Parameters Service Account Email Fill in the parameter... P12 Service Account Private Key Path (i.e. *.p12) Fill in the parameter... ProjectId Fill in the parameter... DatasetId (Choose after ProjectId) Fill in the parameter... Optional Parameters Scope Fill in the parameter... RetryMode Fill in the parameter... RetryStatusCodeList Fill in the parameter... RetryCountMax Fill in the parameter... RetryMultiplyWaitTime Fill in the parameter... Job Location Fill in the parameter... -
Select the desired endpoint, change/pass the properties values, and click on Preview Data button to make the API call.
API Source - Google BigQueryRead / write Google BigQuery data inside your app without coding using easy to use high performance API Connector -
That's it! We are done! Just in a few clicks we configured the call to Google BigQuery using Google BigQuery Connector.
You can load the source data into your desired destination using the Upsert Destination, which supports SQL Server, PostgreSQL, and Amazon Redshift. We also offer other destinations such as CSV, Excel, Azure Table, Salesforce, and more. You can check out our SSIS PowerPack Tasks and components for more options. (*loaded in Trash Destination)
Load Google BigQuery data in SQL Server using Upsert Destination (Insert or Update)
Once you read data from the desired source, now let's see how to Load Google BigQuery data in SQL Server using Upsert Destination. Upsert Destination can Merge/Synchronize data from source to target for Microsoft SQL Server, PostgreSql and Redshift. It supports very fast Bulk Upsert (Update or Insert) operation along and Bulk delete.
-
From the SSIS toolbox drag and drop Upsert Destination on the dataflow designer surface
Connect our Source component to Upsert Destination
-
Double click on Upsert Destination component to configure it.
-
Select the desired Microsoft SQL Server/PostgreSql/Redshift Target Connection or click NEW to create new connection. Select Target Table or click NEW to create new table based on source columns.
Configure SSIS Upsert Destination Connection - Loading data (REST / SOAP / JSON / XML /CSV) into SQL Server or other target using SSIS -
Set Action to Upsert => (insert if not matching in target else update). Select Target Connection and Target Table. Check on Insert and Update. Click on Map All to Mappings all columns and check on Only Primary Key columns.
-
Click on OK to save Upsert Destination settings UI.
-
That's it. Run the SSIS Package and it will read the data from the Google BigQuery and load the data in the SQL Server/PostgreSql/Redshift using Upsert Destination.
Video Tutorial - Write or lookup data to Google BigQuery using SSIS
This video covers following and more so watch carefully. After watching this video follow the steps described in this article.
- How to download SSIS PowerPack for
Google BigQuery integration in SSIS - How to configure connection for
Google BigQuery - How to Write or lookup data to
Google BigQuery - Features about SSIS API Destination
- Using
Google BigQuery Connector in SSIS
How to write or lookup data to Google BigQuery in SSIS (Import data)
In upper section we learned how to read data, now in this section we will learn how to configure Google BigQuery in the API Source to Post data to the Google BigQuery.
-
Read the data from the source, being any desired source component. In example we will use ZappySys Dummy Data Source component.
-
From the SSIS Toolbox drag and drop API Destination (Predefined Templates) on the Data Flow Designer surface and connect source component with it, and double click to edit it.
-
Select New Connection to create a new connection:
API Destination - Google BigQueryRead / write Google BigQuery data inside your app without coding using easy to use high performance API Connector -
Use a preinstalled Google BigQuery Connector from Popular Connector List or press Search Online radio button to download Google BigQuery Connector. Once downloaded simply use it in the configuration:
Google BigQuery -
Proceed with selecting the desired Authentication Type. Then select API Base URL (in most cases default one is the right one). Finally, fill in all the required parameters and set optional parameters if needed. You may press a link Steps to Configure which will help set certain parameters. More info is available in Authentication section.
Steps to get Google BigQuery Credentials
This connection can be configured using two ways. Use Default App (Created by ZappySys) OR Use Custom App created by you.
To use minimum settings you can start with ZappySys created App. Just change UseCustomApp=false on the properties grid so you dont need ClientID / Secret. When you click Generate Token you might see warning about App is not trusted (Simply Click Advanced Link to expand hidden section and then click Go to App link to Proceed). To register custom App, perform the following steps (Detailed steps found in the help link at the end)- Go to Google API Console
- From the Project Dropdown (usually found at the top bar) click Select Project
- On Project Propup click CREATE PROJECT
- Once project is created you can click Select Project to switch the context (You can click on Notification link or Choose from Top Dropdown)
- Click ENABLE APIS AND SERVICES
- Now we need to Enable two APIs one by one (BigQuery API and Cloud Resource Manager API).
- Search BigQuery API. Select and click ENABLE
- Search Cloud Resource Manager API. Select and click ENABLE
- Go to back to main screen of Google API Console
Click OAuth consent screen Tab. Enter necessary details and Save.
- Choose Testing as Publishing status
- Set application User type to Internal, if possible
- If MAKE INTERNAL option is disabled, then add a user in Test users section, which you will use in authentication process when generating Access and Refresh tokens
- Click Credentials Tab
- Click CREATE CREDENTIALS (some where in topbar) and select OAuth Client ID option.
- When prompted Select Application Type as Desktop App and click Create to receive your ClientID and Secret. Later on you can use this information now to configure Connection with UseCustomApp=true.
- Go to OAuth Consent Screen tab. Under Publishing Status click PUBLISH APP to ensure your refresh token doesnt expire often. If you planning to use App for Private use then do not have to worry about Verification Status after Publish.
Google BigQueryUser Account [OAuth]https://www.googleapis.com/bigquery/v2Required Parameters UseCustomApp Fill in the parameter... ProjectId (Choose after [Generate Token] clicked) Fill in the parameter... DatasetId (Choose after [Generate Token] clicked and ProjectId selected) Fill in the parameter... Optional Parameters ClientId Fill in the parameter... ClientSecret Fill in the parameter... Scope Fill in the parameter... RetryMode Fill in the parameter... RetryStatusCodeList Fill in the parameter... RetryCountMax Fill in the parameter... RetryMultiplyWaitTime Fill in the parameter... Job Location Fill in the parameter... Redirect URL (Only for Web App) Fill in the parameter... Steps to get Google BigQuery Credentials
Use these steps to authenticate as service account rather than Google / GSuite User. Learn more about service account here Basically to call Google API as Service account we need to perform following steps listed in 3 sections (Detailed steps found in the help link at the end)Create Project
First thing is create a Project so we can call Google API. Skip this section if you already have Project (Go to next section)- Go to Google API Console
- From the Project Dropdown (usually found at the top bar) click Select Project
- On Project Propup click CREATE PROJECT
- Once project is created you can click Select Project to switch the context (You can click on Notification link or Choose from Top Dropdown)
- Click ENABLE APIS AND SERVICES
- Now we need to Enable two APIs one by one (BigQuery API and Cloud Resource Manager API).
- Search BigQuery API. Select and click ENABLE
- Search Cloud Resource Manager API. Select and click ENABLE
Create Service Account
Once Project is created and APIs are enabled we can now create a service account under that project. Service account has its ID which looks like some email ID (not to confuse with Google /Gmail email ID)- Go to Create Service Account
- From the Project Dropdown (usually found at the top bar) click Select Project
- Enter Service account name and Service account description
- Click on Create. Now you should see an option to assign Service Account permissions (See Next Section).
Give Permission to Service Account
By default service account cant access BigQuery data or List BigQuery Projects so we need to give that permission using below steps.- After you Create Service Account look for Permission drop down in the Wizard.
- Choose BigQuery -> BigQuery Admin role so we can read/write data. (NOTE: If you just need read only access then you can choose BigQuery Data Viewer)
- Now choose one more Project -> Viewer and add that role so we can query Project Ids.
- Click on Continue. Now you should see an option to Create Key (See Next Section).
Create Key (P12)
Once service account is created and Permission is assigned we need to create key file.- In the Cloud Console, click the email address for the service account that you created.
- Click Keys.
- Click Add key, then click Create new key.
- Click Create and select P12 format. A P12 key file is downloaded to your computer. We will use this file in our API connection.
- Click Close.
- Now you may use downloaded *.p12 key file as secret file and Service Account Email as Client ID (e.g. some_name@some_name.iam.gserviceaccount.com).
Manage Permissions / Give Access to Other Projects
We saw how to add permissions for Service Account during Account Creation Wizard but if you ever wish to edit after its created or you wish to give permission for other projects then perform forllowing steps.- From the top Select Project for which you like to edit Permission.
- Go to IAM Menu option (here)
Link to IAM: https://console.cloud.google.com/iam-admin/iam - Goto Permissions tab. Over there you will find ADD button.
- Enter Service account email for which you like to grant permission. Select role you wish to assign.
Google BigQueryService Account (Using Private Key File) [OAuth]https://www.googleapis.com/bigquery/v2Required Parameters Service Account Email Fill in the parameter... P12 Service Account Private Key Path (i.e. *.p12) Fill in the parameter... ProjectId Fill in the parameter... DatasetId (Choose after ProjectId) Fill in the parameter... Optional Parameters Scope Fill in the parameter... RetryMode Fill in the parameter... RetryStatusCodeList Fill in the parameter... RetryCountMax Fill in the parameter... RetryMultiplyWaitTime Fill in the parameter... Job Location Fill in the parameter... -
Select the desired endpoint, change/pass the properties values, and go to the Mappings tab to map the columns.
API Destination - Google BigQueryRead / write Google BigQuery data inside your app without coding using easy to use high performance API Connector -
Finally, map the desired columns:
API Destination - Google BigQueryRead / write Google BigQuery data inside your app without coding using easy to use high performance API Connector -
That's it; we successfully configured the POST API Call. In a few clicks we configured the Google BigQuery API call using ZappySys Google BigQuery Connector
Advanced topics
Actions supported by Google BigQuery Connector
Google BigQuery Connector support following actions for REST API integration. If some actions are not listed below then you can easily edit Connector file and enhance out of the box functionality.Parameter | Description | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
SQL Statement (i.e. SELECT / DROP / CREATE) |
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Use Legacy SQL Syntax? |
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
timeout (Milliseconds) |
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Job Location |
|
Parameter | Description |
---|---|
ProjectId |
|
DatasetId |
|
TableId |
|
Parameter | Description |
---|
Parameter | Description |
---|---|
SearchFilter |
|
Parameter | Description | ||||||
---|---|---|---|---|---|---|---|
ProjectId |
|
||||||
SearchFilter |
|
||||||
all |
|
Parameter | Description |
---|---|
ProjectId |
|
Dataset Name |
|
Description |
|
Parameter | Description | ||||||
---|---|---|---|---|---|---|---|
ProjectId |
|
||||||
DatasetId |
|
||||||
Delete All Tables |
|
Parameter | Description |
---|---|
ProjectId |
|
DatasetId |
|
TableId |
|
Parameter | Description |
---|---|
ProjectId |
|
DatasetId |
|
Parameter | Description | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
SQL Query |
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Filter |
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Use Legacy SQL Syntax? |
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
timeout (Milliseconds) |
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Job Location |
|
Parameter | Description |
---|---|
DatasetId |
|
TableId |
|
Filter |
|
Parameter | Description |
---|---|
ProjectId |
|
DatasetId |
|
TableId |
|
Parameter | Description |
---|---|
Url |
|
Body |
|
IsMultiPart |
|
Filter |
|
Headers |
|
Conclusion
In this article we discussed how to connect to Google BigQuery in SSIS and integrate data without any coding. Click here to Download Google BigQuery Connector for SSIS and try yourself see how easy it is. If you still have any question(s) then ask here or simply click on live chat icon below and ask our expert (see bottom-right corner of this page).
Download Google BigQuery Connector for SSIS
Documentation
More integrations
Other application integration scenarios for Google BigQuery
Other connectors for SSIS
Download Google BigQuery Connector for SSIS
Documentation
How to connect Google BigQuery in SSIS?
How to get Google BigQuery data in SSIS?
How to read Google BigQuery data in SSIS?
How to load Google BigQuery data in SSIS?
How to import Google BigQuery data in SSIS?
How to pull Google BigQuery data in SSIS?
How to push data to Google BigQuery in SSIS?
How to write data to Google BigQuery in SSIS?
How to POST data to Google BigQuery in SSIS?
Call Google BigQuery API in SSIS
Consume Google BigQuery API in SSIS
Google BigQuery SSIS Automate
Google BigQuery SSIS Integration
Integration Google BigQuery in SSIS
Consume real-time Google BigQuery data in SSIS
Consume real-time Google BigQuery API data in SSIS
Google BigQuery ODBC Driver | ODBC Driver for Google BigQuery | ODBC Google BigQuery Driver | SSIS Google BigQuery Source | SSIS Google BigQuery Destination
Connect Google BigQuery in SSIS
Load Google BigQuery in SSIS
Load Google BigQuery data in SSIS
Read Google BigQuery data in SSIS
Google BigQuery API Call in SSIS