JSON Connector for Azure Data Factory (SSIS)

In this article you will learn how to quickly and efficiently integrate JSON data in Azure Data Factory (SSIS) without coding. We will use high-performance JSON Connector to easily connect to JSON and then access the data inside Azure Data Factory (SSIS).

JSON Connector can be used to extract and output JSON data coming from REST API web service calls (Web URL) or direct JSON String (variables or DB columns) or local JSON files data. JSON Connector also supports JSONPath to filter data from nested array/sub-documents. This Connector is optimized to work with very large JSON string.

Let's follow the steps below to see how we can accomplish that!

Download Documentation

Create SSIS package

First of all, create an SSIS package, which will connect to JSON in SSIS. Once you do that, you are one step closer to deploying and running it in Azure-SSIS integration runtime in Azure Data Factory (ADF). Then simply proceed to the next step - creating and configuring Azure Blob Storage Container.

Prepare custom setup files for Azure-SSIS runtime

Now it's time to start preparing custom setup files for Azure-SSIS runtime. During Azure-SSIS runtime creation you can instruct ADF to perform a custom setup on a VM (Azure-SSIS node); i.e. to run the custom installer, copy files, execute PowerShell scripts, etc. In that case, your custom setup files are downloaded and run in the Azure-SSIS node (a VM) when you start the runtime. In this section we will prepare custom setup files so that you can run SSIS packages with SSIS PowerPack connectors inside in Azure-SSIS runtime.

Read more on Azure-SSIS runtime custom setup in Microsoft Azure Data Factory reference.

Trial Users

Use the step below if you are a Trial User, when you did not purchase a license key. Proceed with these steps:

  1. Download SSIS PowerPack trial installer.
    Make sure you don't rename the installer and keep it named as SSISPowerPackSetup_64bit_Trial.msi.
  2. Create a text file and name it main.cmd (make it all lowercase, very important).
  3. Copy and paste this script into it and save it:
    set DIR=%CUSTOM_SETUP_SCRIPT_LOG_DIR%
    
    echo Calling Step 1 : %TIME% >> "%DIR%\steps_log.txt"
    dir /s /b > "%DIR%\file_list.txt"
    
    echo Calling Step 2 : %TIME% >> "%DIR%\steps_log.txt"
    
    ::Install SSIS PowerPack
    msiexec /i  "SSISPowerPackSetup_64bit_Trial.msi" ADDLOCAL=ALL /q  /L*V "%DIR%\powerpack_trial_install_log.txt"
    
    echo Calling Step 3 : %TIME% >> "%DIR%\steps_log.txt"
    dir "C:\Program Files\Microsoft SQL Server\*Zappy*.*"  /s /b >> "%DIR%\installed_files.txt"
    dir "C:\Program Files (x86)\Microsoft SQL Server\*Zappy*.*"  /s /b >> "%DIR%\installed_files.txt"
    
    echo DONE : %TIME% >> "%DIR%\steps_log.txt"
    
    echo complete
    This is the entry-point script that is executed when Azure-SSIS runtime is started.
  4. At last! You are ready to upload these two files — main.cmd & SSISPowerPackSetup_64bit_Trial.msi — into your Azure Blob Storage container's folder, which we will do in the Upload custom setup files to Azure Blob Storage container step.

Paid Customers

Use the steps below if you are a Paid Customer, when you purchased a license. Proceed with these steps:

  1. Download SSIS PowerPack paid installer.
    Make sure you don't rename the installer and keep it named as SSISPowerPackSetup_64bit.msi.
  2. Have your SSIS PowerPack license key handy, we will need it in the below script.
  3. Create a text file and name it main.cmd (make it all lowercase, very important).
  4. Copy and paste the below script into it.
  5. Paste your license key by replacing parameter's --register argument with your real license key.
  6. Finally, save main.cmd:
    set DIR=%CUSTOM_SETUP_SCRIPT_LOG_DIR%
    
    echo Calling Step 1 : %TIME% >> "%DIR%\steps_log.txt"
    dir /s /b > "%DIR%\file_list.txt"
    
    echo Calling Step 2 : %TIME% >> "%DIR%\steps_log.txt"
    
    ::Install SSIS PowerPack
    msiexec /i  "SSISPowerPackSetup_64bit.msi" ADDLOCAL=ALL /q  /L*V "%DIR%\powerpack_install_log.txt"
    
    echo Calling Step 3 : %TIME% >> "%DIR%\steps_log.txt"
    
    ::Activate PowerPack license (Optional)
    "C:\Program Files (x86)\ZappySys\SSIS PowerPack (64 bit)\LicenseManager.exe" -p SSISPowerPack --register "lgGAAO0-----REPLACE-WITH-YOUR-LICENSE-KEY-----czM=" --logfile "%DIR%\powerpack_register_log.txt"
    
    ::Show System Info
    echo Calling Step 4 : %TIME% >> "%DIR%\steps_log.txt"
    "C:\Program Files (x86)\ZappySys\SSIS PowerPack (64 bit)\LicenseManager.exe" -i -l "%DIR%\sysinfo_log.txt"
    
    echo Calling Step 5 : %TIME% >> "%DIR%\steps_log.txt"
    dir "C:\Program Files\Microsoft SQL Server\*Zappy*.*"  /s /b >> "%DIR%\installed_files.txt"
    dir "C:\Program Files (x86)\Microsoft SQL Server\*Zappy*.*"  /s /b >> "%DIR%\installed_files.txt"
    
    echo DONE : %TIME% >> "%DIR%\steps_log.txt"
    
    echo complete
    This is the entry-point script that is executed when Azure-SSIS runtime is started.
  7. At last! You are ready to upload these two files — main.cmd & SSISPowerPackSetup_64bit.msi — into your Azure Blob Storage container's folder, which we will do in the Upload custom setup files to Azure Blob Storage container step.

Upload custom setup files to Azure Blob Storage container

Within Azure Blob Storage container we will store custom setup files we prepared in the previous step so that Azure-SSIS can use them in custom setup process. Just perform these very simple, but very important steps:

  1. Create Azure Blob Storage container, if you haven't done it already
    Make sure you create and use Azure Blob Storage container instead of Azure Data Lake Storage folder. Azure Data Lake Storage won't allow creating an SAS URI for the container, which is a crucial step in the process.
  2. Find Blob Containers node, right-click on it and hit Create Blob Container option: Create a new blob container in Azure Storage Explorer
  3. Upload the two custom setup files — main.cmd & the MSI installer — into your Azure Blob Storage container's folder: Upload SSIS Custom Setup Files to Azure Data Factory
  4. It was easy, wasn't it? It's time we create an SAS URI in the next step.

Create SAS URI for Azure Blob Container

Once you have custom setup files prepared, it's time we generate an SAS URI. This SAS URI will be used by a new Azure-SSIS runtime to install SSIS PowerPack inside the runtime's node, a VM. Let's proceed together by performing the steps below:

  1. Install and launch Azure Storage Explorer.
  2. Right-click on the Storage Accounts node and then hit Connect to Azure storage... menu item: Add Azure Storage account to Azure Storage Explorer
  3. Proceed by right-clicking on that container node and select Get Shared Access Signature... option.
  4. Next, set the Expiry time field to a date far in the future.
    If you restart Azure-SSIS runtime and your SAS URI is expired by that time, it will not start.
  5. Select Read, Create, Write, and List permissions: Generate SAS URI in Azure Storage Explorer for Azure Data Factory Custom Setup
    We also recommend to add Delete permission too to support future functionality.
  6. Copy SAS URL to the clipboard and save it for the next step: Get container SAS URI for Azure Data Factory SSIS Custom Setup You can also generate and copy SAS URL from within Azure Portal itself: Generate SAS URI in Azure Data Factory Custom Setup via online portal

Create Azure-SSIS integration runtime

Once you have the SAS URL we obtained in the previous step, we are ready to move on to create an Azure-SSIS runtime in Azure Data Factory:

  1. Firstly, perform the steps described in Create an Azure-SSIS integration runtime article in Azure Data Factory reference.
  2. In Advanced settings page section, configure Custom setup container SAS URI you obtained in the previous step: Configure SAS URI in Azure Data Factory custom setup
  3. And you are done! That was quick! You can see your Azure-SSIS runtime up and running: Verify Azure-SSIS runtime status in Azure Data Factory portal

The custom setup script is executed only once — at the time an Azure-SSIS runtime is started.

It is also executed if you stop and start Azure-SSIS runtime again.

Deploy SSIS package in Visual Studio

We are ready to deploy the SSIS package to Azure-SSIS runtime. Once you do that, proceed to the next step for the grand finale! Deploy SSIS package to Azure Data Factory from Visual Studio

Execute SSIS package in SQL Server Management Studio (SSMS)

After all hard work, we are ready to execute SSIS package in SQL Server Management Studio (SSMS):

  1. Connect to the SQL Server which is linked to your Azure-SSIS runtime and contains SSISDB database.
  2. Navigate to Integration Services Catalog » Your Folder » Your Project » Your Package, right-click on it, and hit Execute...: Execute SSIS package using SQL Server Management Studio (SSMS)
  3. To view the status of the past execution, navigate to Integration Services Catalog » Your Folder » Your Project » Your Package, right-click on it, and select Reports » Standard Reports » All Executions menu item: Monitor SSIS package execution using SSMS UI

Scenarios

Moving SSIS PowerPack license to another Azure-SSIS runtime

If you are a Paid Customer, there will be a time when you no longer use Azure-SSIS runtime or you need to use your license on a different ADF instance. To transfer a license from one Azure-SSIS runtime to another, perform these steps:

  1. Copy & paste this script into main.cmd we used in the previous step:
    set DIR=%CUSTOM_SETUP_SCRIPT_LOG_DIR%
    
    echo Calling Step 1 : %TIME% >> "%DIR%\steps_log.txt"
    dir /s /b > "%DIR%\file_list.txt"
     
    echo Calling Step 2 : %TIME% >> "%DIR%\steps_log.txt"
    
    ::Install SSIS PowerPack
    msiexec /i  "SSISPowerPackSetup_64bit.msi" ADDLOCAL=ALL /q  /L*V "%DIR%\powerpack_install_log.txt"
     
    echo Calling Step 3 : %TIME% >> "%DIR%\steps_log.txt"
     
    ::De-Activate same license
    "C:\Program Files (x86)\ZappySys\SSIS PowerPack (64 bit)\LicenseManager.exe" -p SSISPowerPack --unregister --logfile "%DIR%\powerpack_un_register_log.txt"
     
    ::Show System Info
    echo Calling Step 4 : %TIME% >> "%DIR%\steps_log.txt"
    "C:\Program Files (x86)\ZappySys\SSIS PowerPack (64 bit)\LicenseManager.exe" -i -l "%DIR%\sysinfo_log.txt"
     
    echo Calling Step 5 : %TIME% >> "%DIR%\steps_log.txt"
    dir "C:\Program Files\Microsoft SQL Server\*Zappy*.*"  /s /b >> "%DIR%\installed_files.txt"
    dir "C:\Program Files (x86)\Microsoft SQL Server\*Zappy*.*"  /s /b >> "%DIR%\installed_files.txt"
     
    echo DONE : %TIME% >> "%DIR%\steps_log.txt"
     
    echo complete
  2. Start Azure-SSIS runtime.
    This will unregister your license on the original Azure-SSIS runtime.
  3. Stop Azure-SSIS runtime to deallocate resources in Azure.
  4. Now you are free to activate it on another Azure-SSIS runtime.

Configuring pagination in the JSON Driver

ZappySys JSON Driver equips users with powerful tools for seamless data extraction and management from REST APIs, leveraging advanced pagination methods for enhanced efficiency. These options are designed to handle various types of pagination structures commonly used in APIs. Below are the detailed descriptions of these options:

  1. Page-based Pagination: This method works by retrieving data in fixed-size pages from the Rest API. It allows you to specify the page size and navigate through the results by requesting different page numbers, ensuring that you can access all the data in a structured manner.

  2. Offset-based Pagination: With this approach, you can extract data by specifying the starting point or offset from which to begin retrieving data. It allows you to define the number of records to skip and fetch subsequent data accordingly, providing precise control over the data extraction process.

  3. Cursor-based Pagination: This technique involves using a cursor or a marker that points to a specific position in the dataset. It enables you to retrieve data starting from the position indicated by the cursor and proceed to subsequent segments, ensuring that you capture all the relevant information without missing any records.

  4. Token-based Pagination: In this method, a token serves as a unique identifier for a specific data segment. It allows you to access the next set of data by using the token provided in the response from the previous request. This ensures that you can systematically retrieve all the data segments without duplication or omission.

Utilizing these comprehensive pagination features in the ZappySys JSON Driver facilitates efficient data management and extraction from REST APIs, optimizing the integration and analysis of extensive datasets.

For more detailed steps, please refer to this link: How to do REST API Pagination in SSIS / ODBC Drivers

Authentication

ZappySys offers various authentication methods to securely access data from various sources. These authentication methods include OAuth, Basic Authentication, Token-based Authentication, and more, allowing users to connect to a wide range of data sources securely.

ZappySys Authentication is a robust system that facilitates secure access to data from a diverse range of sources. It includes a variety of authentication methods tailored to meet the specific requirements of different data platforms and services. These authentication methods may involve:

  1. OAuth: ZappySys supports OAuth for authentication, which allows users to grant limited access to their data without revealing their credentials. It's commonly used for applications that require access to user account information.

  2. Basic Authentication: This method involves sending a username and password with every request. ZappySys allows users to securely access data using this traditional authentication approach.

  3. Token-based Authentication: ZappySys enables users to utilize tokens for authentication. This method involves exchanging a unique token with each request to authenticate the user's identity without revealing sensitive information.

By implementing these authentication methods, ZappySys ensures the secure and reliable retrieval of data from various sources, providing users with the necessary tools to access and integrate data securely and efficiently. For more comprehensive details on the authentication process, please refer to the official ZappySys documentation or reach out to their support team for further assistance.

For more details, please refer to this link: ZappySys Connections

Create Custom Stored Procedure in ZappySys Driver

You can create procedures to encapsulate custom logic and then only pass handful parameters rather than long SQL to execute your API call.

Steps to create Custom Stored Procedure in ZappySys Driver. You can insert Placeholders anywhere inside Procedure Body. Read more about placeholders here

  1. Go to Custom Objects Tab and Click on Add button and Select Add Procedure:
    ZappySys Driver - Add Stored Procedure

  2. Enter the desired Procedure name and click on OK:
    ZappySys Driver - Add Stored Procedure Name

  3. Select the created Stored Procedure and write the your desired stored procedure and Save it and it will create the custom stored procedure in the ZappySys Driver:
    Here is an example stored procedure for ZappySys Driver. You can insert Placeholders anywhere inside Procedure Body. Read more about placeholders here

    CREATE PROCEDURE [usp_get_orders]
        @fromdate = '<<yyyy-MM-dd,FUN_TODAY>>'
     AS
        SELECT * FROM Orders where OrderDate >= '<@fromdate>';
    

    ZappySys Driver - Create Custom Stored Procedure

  4. That's it now go to Preview Tab and Execute your Stored Procedure using Exec Command. In this example it will extract the orders from the date 1996-01-01:

    Exec usp_get_orders '1996-01-01';

    ZappySys Driver - Execute Custom Stored Procedure

  5. Let's generate the SQL Server Query Code to make the API call using stored procedure. Go to Code Generator Tab, select language as SQL Server and click on Generate button the generate the code.
    As we already created the linked server for this Data Source, in that you just need to copy the Select Query and need to use the linked server name which we have apply on the place of [MY_API_SERVICE] placeholder.

    SELECT * FROM OPENQUERY([LINKED_SERVER_TO_JSON_IN_DATA_GATEWAY], 'EXEC usp_get_orders @fromdate=''1996-07-30''')

    ZappySys Driver - Generate SQL Server Query

  6. Now go to SQL served and execute that query and it will make the API call using stored procedure and provide you the response.
    ZappySys Driver - Generate SQL Server Query

Create Custom Virtual Table in ZappySys Driver

ZappySys API Drivers support flexible Query language so you can override Default Properties you configured on Data Source such as URL, Body. This way you don't have to create multiple Data Sources if you like to read data from multiple EndPoints. However not every application support supplying custom SQL to driver so you can only select Table from list returned from driver.

If you're dealing with Microsoft Access and need to import data from an SQL query, it's important to note that Access doesn't allow direct import of SQL queries. Instead, you can create custom objects (Virtual Tables) to handle the import process.

Many applications like MS Access, Informatica Designer wont give you option to specify custom SQL when you import Objects. In such case Virtual Table is very useful. You can create many Virtual Tables on the same Data Source (e.g. If you have 50 URLs with slight variations you can create virtual tables with just URL as Parameter setting.

  1. Go to Custom Objects Tab and Click on Add button and Select Add Table:
    ZappySys Driver - Add Table

  2. Enter the desired Table name and click on OK:
    ZappySys Driver - Add Table Name

  3. And it will open the New Query Window Click on Cancel to close that window and go to Custom Objects Tab.

  4. Select the created table, Select Text Type AS SQL and write the your desired SQL Query and Save it and it will create the custom table in the ZappySys Driver:
    Here is an example SQL query for ZappySys Driver. You can insert Placeholders also. Read more about placeholders here

    SELECT
      "ShipCountry",
      "OrderID",
      "CustomerID",
      "EmployeeID",
      "OrderDate",
      "RequiredDate",
      "ShippedDate",
      "ShipVia",
      "Freight",
      "ShipName",
      "ShipAddress",
      "ShipCity",
      "ShipRegion",
      "ShipPostalCode"
    FROM "Orders"
    Where "ShipCountry"='USA'

    ZappySys Driver - Create Custom Table

  5. That's it now go to Preview Tab and Execute your custom virtual table query. In this example it will extract the orders for the USA Shipping Country only:

    SELECT * FROM "vt__usa_orders_only"

    ZappySys Driver - Execute Custom Virtual Table Query

  6. Let's generate the SQL Server Query Code to make the API call using stored procedure. Go to Code Generator Tab, select language as SQL Server and click on Generate button the generate the code.
    As we already created the linked server for this Data Source, in that you just need to copy the Select Query and need to use the linked server name which we have apply on the place of [MY_API_SERVICE] placeholder.

    SELECT * FROM OPENQUERY([LINKED_SERVER_TO_JSON_IN_DATA_GATEWAY], 'EXEC [usp_get_orders] ''1996-01-01''')

    ZappySys Driver - Generate SQL Server Query

  7. Now go to SQL served and execute that query and it will make the API call using stored procedure and provide you the response.
    ZappySys Driver - Generate SQL Server Query

Conclusion

In this article we showed you how to connect to JSON in Azure Data Factory (SSIS) and integrate data without any coding, saving you time and effort. We encourage you to download JSON Connector for Azure Data Factory (SSIS) and see how easy it is to use it for yourself or your team.

If you have any questions, feel free to contact ZappySys support team. You can also open a live chat immediately by clicking on the chat icon below.

Download JSON Connector for Azure Data Factory (SSIS) Documentation

More integrations

Other connectors for Azure Data Factory (SSIS)

Other application integration scenarios for JSON

  • How to connect JSON in Azure Data Factory (SSIS)?

  • How to get JSON data in Azure Data Factory (SSIS)?

  • How to read JSON data in Azure Data Factory (SSIS)?

  • How to load JSON data in Azure Data Factory (SSIS)?

  • How to import JSON data in Azure Data Factory (SSIS)?

  • How to pull JSON data in Azure Data Factory (SSIS)?

  • How to push data to JSON in Azure Data Factory (SSIS)?

  • How to write data to JSON in Azure Data Factory (SSIS)?

  • How to POST data to JSON in Azure Data Factory (SSIS)?

  • Call JSON API in Azure Data Factory (SSIS)

  • Consume JSON API in Azure Data Factory (SSIS)

  • JSON Azure Data Factory (SSIS) Automate

  • JSON Azure Data Factory (SSIS) Integration

  • Integration JSON in Azure Data Factory (SSIS)

  • Consume real-time JSON data in Azure Data Factory (SSIS)

  • Consume real-time JSON API data in Azure Data Factory (SSIS)

  • JSON ODBC Driver | ODBC Driver for JSON | ODBC JSON Driver | SSIS JSON Source | SSIS JSON Destination

  • Connect JSON in Azure Data Factory (SSIS)

  • Load JSON in Azure Data Factory (SSIS)

  • Load JSON data in Azure Data Factory (SSIS)

  • Read JSON data in Azure Data Factory (SSIS)

  • JSON API Call in Azure Data Factory (SSIS)