Google BigQuery Connector for TableauRead / write Google BigQuery data inside your app without coding using easy to use high performance API Connector In this article you will learn how to quickly and efficiently integrate Google BigQuery data in Tableau without coding. We will use high-performance Google BigQuery Connector to easily connect to Google BigQuery and then access the data inside Tableau. Let's follow the steps below to see how we can accomplish that! Google BigQuery Connector for Tableau is based on ZappySys API Driver which is part of ODBC PowerPack. It is a collection of high-performance ODBC drivers that enable you to integrate data in SQL Server, SSIS, a programming language, or any other ODBC-compatible application. ODBC PowerPack supports various file formats, sources and destinations, including REST/SOAP API, SFTP/FTP, storage services, and plain files, to mention a few. |
Connect to Google BigQuery in other apps
|
Create Data Source in ZappySys Data Gateway based on API Driver
-
Download and install ODBC PowerPack.
-
Search for gateway in start menu and Open ZappySys Data Gateway:
-
Go to Users Tab to add our first Gateway user. Click Add; we will give it a name tdsuser and enter password you like to give. Check Admin option and click OK to save. We will use these details later when we create linked server:
-
Now we are ready to add a data source. Click Add, give data source a name (Copy this name somewhere, we will need it later) and then select Native - ZappySys API Driver. Finally, click OK. And it will create the Data Set for it and open the ZS driver UI.
GoogleBigqueryDSN
-
When the Configuration window appears give your data source a name if you haven't done that already, then select "Google BigQuery" from the list of Popular Connectors. If "Google BigQuery" is not present in the list, then click "Search Online" and download it. Then set the path to the location where you downloaded it. Finally, click Continue >> to proceed with configuring the DSN:
GoogleBigqueryDSNGoogle BigQuery -
Now it's time to configure the Connection Manager. Select Authentication Type, e.g. Token Authentication. Then select API Base URL (in most cases, the default one is the right one). More info is available in the Authentication section.
User accounts represent a developer, administrator, or any other person who interacts with Google APIs and services. User accounts are managed as Google Accounts, either with Google Workspace or Cloud Identity. They can also be user accounts that are managed by a third-party identity provider and federated with Workforce Identity Federation. [API reference]
Steps how to get and use Google BigQuery credentials
Follow these steps on how to create Client Credentials (User Account principle) to authenticate and access BigQuery API in SSIS package or ODBC data source:
WARNING: If you are planning to automate processes, we recommend that you use a Service Account authentication method. In case, you still need to use User Account, then make sure you use a system/generic account (e.g.automation@my-company.com
). When you use a personal account which is tied to a specific employee profile and that employee leaves the company, the token may become invalid and any automated processes using that token will start to fail.Step-1: Create project
This step is optional, if you already have a project in Google Cloud and can use it. However, if you don't, proceed with these simple steps to create one:
-
First of all, go to Google API Console.
-
Then click Select a project button and then click NEW PROJECT button:
-
Name your project and click CREATE button:
-
Wait until the project is created:
- Done! Let's proceed to the next step.
Step-2: Enable Google Cloud APIs
In this step we will enable BigQuery API and Cloud Resource Manager API:
-
Select your project on the top bar:
-
Then click the "hamburger" icon on the top left and access APIs & Services:
-
Now let's enable several APIs by clicking ENABLE APIS AND SERVICES button:
-
In the search bar search for
bigquery api
and then locate and select BigQuery API: -
If BigQuery API is not enabled, enable it:
-
Then repeat the step and enable Cloud Resource Manager API as well:
- Done! Let's proceed to the next step.
Step-3: Create OAuth application
-
First of all, click the "hamburger" icon on the top left and then hit VIEW ALL PRODUCTS:
-
Then access Google Auth Platform to start creating an OAuth application:
-
Start by pressing GET STARTED button:
-
Next, continue by filling in App name and User support email fields:
-
Choose Internal option, if it's enabled, otherwise select External:
-
Optional step if you used
Internal
option in the previous step. Nevertheless, if you had to useExternal
option, then click ADD USERS to add a user: -
Then add your contact Email address:
-
Finally, check the checkbox and click CREATE button:
- Done! Let's create Client Credentials in the next step.
Step-4: Create Client Credentials
-
In Google Auth Platform, select Clients menu item and click CREATE CLIENT button:
-
Choose
Desktop app
as Application type and name your credentials: -
Continue by opening the created credentials:
-
Finally, copy Client ID and Client secret for the later step:
-
Done! We have all the data needed for authentication, let's proceed to the last step!
Step-5: Configure connection
-
Now go to SSIS package or ODBC data source and use previously copied values in User Account authentication configuration:
- In the ClientId field paste the Client ID value.
- In the ClientSecret field paste the Client secret value.
-
Press Generate Token button to generate Access and Refresh Tokens.
-
Then choose ProjectId from the drop down menu.
-
Continue by choosing DatasetId from the drop down menu.
-
Finally, click Test Connection to confirm the connection is working.
-
Done! Now you are ready to use Google BigQuery Connector!
Fill in all required parameters and set optional parameters if needed:
GoogleBigqueryDSNGoogle BigQueryUser Account [OAuth]https://www.googleapis.com/bigquery/v2Required Parameters UseCustomApp Fill-in the parameter... ProjectId (Choose after [Generate Token] clicked) Fill-in the parameter... DatasetId (Choose after [Generate Token] clicked and ProjectId selected) Fill-in the parameter... Optional Parameters ClientId ClientSecret Scope https://www.googleapis.com/auth/bigquery https://www.googleapis.com/auth/bigquery.insertdata https://www.googleapis.com/auth/cloud-platform https://www.googleapis.com/auth/cloud-platform.read-only https://www.googleapis.com/auth/devstorage.full_control https://www.googleapis.com/auth/devstorage.read_only https://www.googleapis.com/auth/devstorage.read_write RetryMode RetryWhenStatusCodeMatch RetryStatusCodeList 429|503 RetryCountMax 5 RetryMultiplyWaitTime True Job Location Redirect URL (Only for Web App) Service accounts are accounts that do not represent a human user. They provide a way to manage authentication and authorization when a human is not directly involved, such as when an application needs to access Google Cloud resources. Service accounts are managed by IAM. [API reference]
Steps how to get and use Google BigQuery credentials
Follow these steps on how to create Service Account to authenticate and access BigQuery API in SSIS package or ODBC data source:
Step-1: Create project
This step is optional, if you already have a project in Google Cloud and can use it. However, if you don't, proceed with these simple steps to create one:
-
First of all, go to Google API Console.
-
Then click Select a project button and then click NEW PROJECT button:
-
Name your project and click CREATE button:
-
Wait until the project is created:
- Done! Let's proceed to the next step.
Step-2: Enable Google Cloud APIs
In this step we will enable BigQuery API and Cloud Resource Manager API:
-
Select your project on the top bar:
-
Then click the "hamburger" icon on the top left and access APIs & Services:
-
Now let's enable several APIs by clicking ENABLE APIS AND SERVICES button:
-
In the search bar search for
bigquery api
and then locate and select BigQuery API: -
If BigQuery API is not enabled, enable it:
-
Then repeat the step and enable Cloud Resource Manager API as well:
- Done! Let's proceed to the next step and create a service account.
Step-3: Create Service Account
Use the steps below to create a Service Account in Google Cloud:
-
First of all, go to IAM & Admin in Google Cloud console:
-
Once you do that, click Service Accounts on the left side and click CREATE SERVICE ACCOUNT button:
-
Then name your service account and click CREATE AND CONTINUE button:
-
Continue by clicking Select a role dropdown and start granting service account BigQuery Admin and Project Viewer roles:
-
Find BigQuery group on the left and then click on BigQuery Admin role on the right:
-
Then click ADD ANOTHER ROLE button, find Project group and select Viewer role:
-
Finish adding roles by clicking CONTINUE button:
You can always add or modify permissions later in IAM & Admin. -
Finally, in the last step, just click button DONE:
-
Done! We are ready to add a Key to this service account in the next step.
Step-4: Add Key to Service Account
We are ready to add a Key (JSON or P12 key file) to the created Service Account:
-
In Service Accounts open newly created service account:
-
Next, copy email address of your service account for the later step:
-
Continue by selecting KEYS tab, then press ADD KEY dropdown, and click Create new key menu item:
-
Finally, select JSON (Engine v19+) or P12 option and hit CREATE button:
- Key file downloads into your machine. We have all the data needed for authentication, let's proceed to the last step!
Step-5: Configure connection
-
Now go to SSIS package or ODBC data source and configure these fields in Service Account authentication configuration:
- In the Service Account Email field paste the service account Email address value you copied in the previous step.
- In the Service Account Private Key Path (i.e. *.json OR *.p12) field use downloaded certificate's file path.
- Done! Now you are ready to use Google BigQuery Connector!
Fill in all required parameters and set optional parameters if needed:
GoogleBigqueryDSNGoogle BigQueryService Account [OAuth]https://www.googleapis.com/bigquery/v2Required Parameters Service Account Email Fill-in the parameter... Service Account Private Key Path (i.e. *.json OR *.p12) Fill-in the parameter... ProjectId Fill-in the parameter... DatasetId (Choose after ProjectId) Fill-in the parameter... Optional Parameters Scope https://www.googleapis.com/auth/bigquery https://www.googleapis.com/auth/bigquery.insertdata https://www.googleapis.com/auth/cloud-platform https://www.googleapis.com/auth/cloud-platform.read-only https://www.googleapis.com/auth/devstorage.full_control https://www.googleapis.com/auth/devstorage.read_only https://www.googleapis.com/auth/devstorage.read_write RetryMode RetryWhenStatusCodeMatch RetryStatusCodeList 429 RetryCountMax 5 RetryMultiplyWaitTime True Job Location Impersonate As (Enter Email Id) -
-
Once the data source connection has been configured, it's time to configure the SQL query. Select the Preview tab and then click Query Builder button to configure the SQL query:
ZappySys API Driver - Google BigQueryRead / write Google BigQuery data inside your app without coding using easy to use high performance API ConnectorGoogleBigqueryDSN -
Start by selecting the Table or Endpoint you are interested in and then configure the parameters. This will generate a query that we will use in Tableau to retrieve data from Google BigQuery. Hit OK button to use this query in the next step.
#DirectSQL SELECT * FROM bigquery-public-data.samples.wikipedia LIMIT 1000 /* try your own dataset or Some FREE dataset like nyc-tlc.yellow.trips -- 3 parts ([Project.]Dataset.Table) */
Some parameters configured in this window will be passed to the Google BigQuery API, e.g. filtering parameters. It means that filtering will be done on the server side (instead of the client side), enabling you to get only the meaningful datamuch faster . -
Now hit Preview Data button to preview the data using the generated SQL query. If you are satisfied with the result, use this query in Tableau:
ZappySys API Driver - Google BigQueryRead / write Google BigQuery data inside your app without coding using easy to use high performance API ConnectorGoogleBigqueryDSN#DirectSQL SELECT * FROM bigquery-public-data.samples.wikipedia LIMIT 1000 /* try your own dataset or Some FREE dataset like nyc-tlc.yellow.trips -- 3 parts ([Project.]Dataset.Table) */
You can also access data quickly from the tables dropdown by selecting <Select table>.AWHERE
clause,LIMIT
keyword will be performed on the client side, meaning that thewhole result set will be retrieved from the Google BigQuery API first, and only then the filtering will be applied to the data. If possible, it is recommended to use parameters in Query Builder to filter the data on the server side (in Google BigQuery servers). -
Click OK to finish creating the data source.
Read data in SQL Server using ZappySys Data Gateway
To read the data in SQL Server, the first thing you have to do is create a Linked Server:
-
First, let's open SQL Server Management Studio, create a new Linked Server, and start configuring it:
LS_TO_GOOGLE_BIGQUERY_IN_GATEWAYMicrosoft OLE DB Driver for SQL Serverlocalhost,5000GoogleBigqueryDSNGoogleBigqueryDSNChoose SQL Server Native Client 11.0 as Provider if you don't see the option shown above. -
Then click on Security option and configure username we created in ZappySys Data Gateway in one of the previous steps:
-
Optional step. Under the Server Options, Enable RPC and RPC Out and Disable Promotion of Distributed Transactions(MSDTC).
You need to enable RPC Out if you plan to use
EXEC(...) AT [LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY]
rather than OPENQUERY.
If don't enabled it, you will encounter theServer 'LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY' is not configured for RPC
error.Query Example:
EXEC('#DirectSQL SELECT * FROM bigquery-public-data.samples.wikipedia LIMIT 1000 /* try your own dataset or Some FREE dataset like nyc-tlc.yellow.trips -- 3 parts ([Project.]Dataset.Table) */') AT [LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY]
If you plan to use
'INSERT INTO <TABLE> EXEC(...) AT [LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY]'
in that case you need to Disable Promotion of Distributed Transactions(MSDTC).
If don't disabled it, you will encounter theThe operation could not be performed because OLE DB provider "SQLNCLI11" for linked server "MY_LINKED_SERVER_NAME" was unable to begin a distributed transaction.
error.Query Example:
INSERT INTO dbo.Products EXEC('#DirectSQL SELECT * FROM bigquery-public-data.samples.wikipedia LIMIT 1000 /* try your own dataset or Some FREE dataset like nyc-tlc.yellow.trips -- 3 parts ([Project.]Dataset.Table) */') AT [LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY]
-
Finally, open a new query and execute a query we saved in one of the previous steps:
SELECT * FROM OPENQUERY([LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY], '#DirectSQL SELECT * FROM bigquery-public-data.samples.wikipedia LIMIT 1000 /* try your own dataset or Some FREE dataset like nyc-tlc.yellow.trips -- 3 parts ([Project.]Dataset.Table) */')
SELECT * FROM OPENQUERY([LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY], '#DirectSQL SELECT * FROM bigquery-public-data.samples.wikipedia LIMIT 1000 /* try your own dataset or Some FREE dataset like nyc-tlc.yellow.trips -- 3 parts ([Project.]Dataset.Table) */')
Create Linked Server using Code
In previous section you saw how to create a Linked Server from UI. You can do similar action by code too (see below). Run below script after changing necessary parameters. Assuming your Data Source name on ZappySys Data Gateway UI is 'GoogleBigqueryDSN'USE [master]
GO
--///////////////////////////////////////////////////////////////////////////////////////
--Run below code in SSMS to create Linked Server and use ZappySys Drivers in SQL Server
--///////////////////////////////////////////////////////////////////////////////////////
-- Replace YOUR_GATEWAY_USER, YOUR_GATEWAY_PASSWORD
-- Replace localhost with IP/Machine name if ZappySys Gateway Running on different machine other than SQL Server
-- Replace Port 5000 if you configured gateway on a different port
--1. Configure your gateway service as per this article https://zappysys.com/links?id=10036
--2. Make sure you have SQL Server Installed. You can download FREE SQL Server Express Edition from here if you dont want to buy Paid version https://www.microsoft.com/en-us/sql-server/sql-server-editions-express
--Uncomment below if you like to drop linked server if it already exists
--EXEC master.dbo.sp_dropserver @server=N'LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY', @droplogins='droplogins'
--3. Create new linked server
EXEC master.dbo.sp_addlinkedserver
@server = N'LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY' --Linked server name (this will be used in OPENQUERY sql
, @srvproduct=N''
---- For MSSQL 2012,2014,2016 and 2019 use below (SQL Server Native Client 11.0)---
, @provider=N'SQLNCLI11'
---- For MSSQL 2022 or higher use below (Microsoft OLE DB Driver for SQL Server)---
--, @provider=N'MSOLEDBSQL'
, @datasrc=N'localhost,5000' --//Machine / Port where Gateway service is running
, @provstr=N'Network Library=DBMSSOCN;'
, @catalog=N'GoogleBigqueryDSN' --Data source name you gave on Gateway service settings
--4. Attach gateway login with linked server
EXEC master.dbo.sp_addlinkedsrvlogin
@rmtsrvname=N'LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY' --linked server name
, @useself=N'False'
, @locallogin=NULL
, @rmtuser=N'YOUR_GATEWAY_USER' --enter your Gateway user name
, @rmtpassword='YOUR_GATEWAY_PASSWORD' --enter your Gateway user's password
GO
--5. Enable RPC OUT (This is Optional - Only needed if you plan to use EXEC(...) AT YourLinkedServerName rather than OPENQUERY
EXEC sp_serveroption 'LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY', 'rpc', true;
EXEC sp_serveroption 'LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY', 'rpc out', true;
--Disable MSDTC - Below needed to support INSERT INTO from EXEC AT statement
EXEC sp_serveroption 'LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY', 'remote proc transaction promotion', false;
--Increase query timeout if query is going to take longer than 10 mins (Default timeout is 600 seconds)
--EXEC sp_serveroption 'LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY', 'query timeout', 1200;
GO
Create View in SQL Server
Finally, use this or similar query in a view or stored procedure, which you will be able to use in Tableau. We will create a view to return invoices:
-
CREATE VIEW vwApiInvoices AS SELECT * FROM OPENQUERY([LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY], 'SELECT * FROM Invoices')
Read data in Tableau from SQL Server
Actually, we will be getting data from SQL Server which in turn will be getting data from ZappySys Data Gateway data source. Let's begin and see how to accomplish that:
- Open Tableau Desktop and click File > New
-
To create new Connection click More > Microsoft SQL Server > Enter your credentials to connect to SQL Server (in our example before we used tdsuser):
-
Once connection is created for SQL Server we can read Google BigQuery data 3 different ways:
- Query View which contains OPENQUERY to Linked Server for Google BigQuery data
- Use direct SQL Query using OPENQUERY
- Use Stored Procedure (Mostly useful to parameterize calls
-
See below example to pull data from Google BigQuery in Tableau using SQL View approach:
-
Once your data sources are created you can click on Sheet1 and drag fields to create visualizations for Tableau Dashboard:
Passing Parameters to Google BigQuery calls in Tableau (Dynamic SQL)
Now let's look at scenario where you have to pass parameters to build Dynamic Dashboard. You can try to insert Parameters in your Direct SQL when you build Dynamic SQL but we found some issues with that so we are going to suggest Stored Procedure approach. For more information on Known issue on Dynamic Metadata Check this post.-
First lets create a stored procedure in SQL Server for Parameter Example. Notice how we added WITH RESULT SETS in the code to describe metadata.
--DROP PROC dbo.usp_GetInvoicesByCountry --GO /* Purpose: Parameterize Google BigQuery call via SQL. Call ZappySys Drivers inside SQL Server. */ CREATE PROC dbo.usp_GetInvoicesByCountry @country varchar(100) AS DECLARE @sql varchar(max) --//Escape single ticks carefully SET @sql = 'SELECT OrderID,CustomerID,Country,Quantity FROM $ WITH (Src=''https://services.odata.org/V3/Northwind/Northwind.svc/Invoices?$format=json@filter=Country eq '+ @country +''' ,Filter=''$.value[*]'' ,DataFormat=''OData'' )' DECLARE @sqlFull varchar(max) SET @sqlFull='SELECT * FROM OPENQUERY([LS_TO_GOOGLE_BIGQUERY_IN_GATEWAY], ''' + REPLACE( @sql, '''', '''''' ) + ''' )' PRINT @sqlFull --//For DEBUG purpose EXECUTE (@sqlFull) WITH RESULT SETS ( (OrderID int,CustomerID varchar(100),Country varchar(100),Quantity int) --//describe first result. If you don't do this then wont work in Tableau ) GO -- Example call EXEC dbo.usp_GetInvoicesByCountry @country='Germany'
- Once you create a stored procedure go to Tableau datasource and select Database which contains the stored procedure we just created.
-
Now find your stored proc and drag it on the datasource pane. You will see parameters UI as below. You can create new parameter - Select New Parameter under Value Column.
- Thats it now you can reuse your parameterized datasource anywhere in Dashboard.
-
If you have need to select Parameters from predefined values rather than free text then edit your parameter and select List option. Define values you like to select from as below.
-
When you create Tableau Dashboard you will see Parameter dropdown (If you selected List) elase you may see Textbox to enter custom value.
Firewall settings
So far we have assumed that Gateway is running on the same machine as SQL Server. However there will be a case when ZappySys ODBC PowerPack is installed on a different machine than SQL Server. In such case you may have to perform additional Firewall configurations. On most computers firewall settings wont allow outside traffic to ZappySys Data Gateway. In such case perform following steps to allow other machines to connect to Gateway.
Method-1 (Preferred)If you are using newer version of ZappySys Data Gateway then adding firewall rule is just a single click.
- Search for gateway in start menu and open ZappySys Data Gateway.
- Go to Firewall Tab and click Add Firewall Rule button like below. This will create Firewall rule to all Inbound Traffic on Port 5000 (Unless you changed it).
- Search for Windows Firewall Advanced Security in start menu.
- Under Inbound Rules > Right click and click [New Rule] >> Click Next
- Select Port on Rule Type >> Click Next
- Click on TCP and enter port number under specified local port as 5000 (use different one if you changed Default port) >> Click Next
- Select Profile (i.e. Private, Public) >> Click Next
- Enter Rule name [i.e. ZappySys Data Gateway – Allow Inbound ] >> Click Next
- Click OK to save the rule

Actions supported by Google BigQuery Connector
Learn how to perform common Google BigQuery actions directly in Tableau with these how-to guides:
- [Dynamic Endpoint]
- Create Dataset
- Delete Dataset
- Delete Table
- Get Query Schema (From SQL)
- Get Table Schema
- insert_table_data
- List Datasets
- List Projects
- List Tables
- post_[Dynamic Endpoint]
- Read Data using SQL Query -OR- Execute Script (i.e. CREATE, SELECT, INSERT, UPDATE, DELETE)
- Read Table Rows
- Generic Request
- Generic Request (Bulk Write)
Conclusion
In this article we showed you how to connect to Google BigQuery in Tableau and integrate data without any coding, saving you time and effort. It's worth noting that ZappySys API Driver allows you to connect not only to Google BigQuery, but to any Java application that supports JDBC (just use a different JDBC driver and configure it appropriately).
We encourage you to download Google BigQuery Connector for Tableau and see how easy it is to use it for yourself or your team.
If you have any questions, feel free to contact ZappySys support team. You can also open a live chat immediately by clicking on the chat icon below.
Download Google BigQuery Connector for Tableau Documentation
More integrations
Other connectors for Tableau
Other application integration scenarios for Google BigQuery
How to connect Google BigQuery in Tableau?
How to get Google BigQuery data in Tableau?
How to read Google BigQuery data in Tableau?
How to load Google BigQuery data in Tableau?
How to import Google BigQuery data in Tableau?
How to pull Google BigQuery data in Tableau?
How to push data to Google BigQuery in Tableau?
How to write data to Google BigQuery in Tableau?
How to POST data to Google BigQuery in Tableau?
Call Google BigQuery API in Tableau
Consume Google BigQuery API in Tableau
Google BigQuery Tableau Automate
Google BigQuery Tableau Integration
Integration Google BigQuery in Tableau
Consume real-time Google BigQuery data in Tableau
Consume real-time Google BigQuery API data in Tableau
Google BigQuery ODBC Driver | ODBC Driver for Google BigQuery | ODBC Google BigQuery Driver | SSIS Google BigQuery Source | SSIS Google BigQuery Destination
Connect Google BigQuery in Tableau
Load Google BigQuery in Tableau
Load Google BigQuery data in Tableau
Read Google BigQuery data in Tableau
Google BigQuery API Call in Tableau