How to integrate FTP/SFTP JSON File using SSIS

Integrate SSIS and FTP/SFTP JSON File
Integrate SSIS and FTP/SFTP JSON File

Learn how to quickly and efficiently connect FTP/SFTP JSON File with SSIS for smooth data access.

Read and write JSON files over FTP/SFTP effortlessly. Streamline, manage, and automate JSON files on FTP and SFTP for analytics, reporting, and data pipelines — almost no coding required. You can do it all using the high-performance FTP/SFTP JSON File Connector. We'll walk you through the entire setup.

Ready to dive in? Download the product to jump right in, or follow the step-by-step guide below to see how it works.

Prerequisites

Before we begin, make sure the following prerequisites are met:

  1. SSIS designer installed. Sometimes it is referred as BIDS or SSDT (download it from Microsoft).
  2. Basic knowledge of SSIS package development using Microsoft SQL Server Integration Services.
  3. SSIS PowerPack is installed (if you are new to SSIS PowerPack, then get started!).

Read data from FTP/SFTP JSON File in SSIS using Secure FTP JSON File Source (Export data)

In this section we will learn how to configure and use FTP/SFTP JSON File Connector in the API Source to extract data from the FTP/SFTP JSON File using Secure FTP Source for JSON Files.

  1. Open Visual Studio and click Create a new project.

  2. Select Integration Services Project. Enter a name and location for your project, then click OK.

  3. From the SSIS Toolbox, drag and drop a Data Flow Task onto the Control Flow surface, and double-click it:

    Drag Data Flow Task onto Control Flow to use SSIS PowerPack Data Flow components
  4. Make sure you are in the Data Flow Task designer:

    Make sure you are in Data Flow designer in SSIS package
  5. From the SSIS toolbox drag and drop Secure FTP JSON File Source on the dataflow designer surface
    SSIS Secure FTP JSON File Source - Drag and Drop

  6. Double click on Secure FTP JSON File Source component to configure it.

  7. Create and configure a connection for the Secure FTP storage account.

    Create Secure FTP Storage Connection
  8. You can use select your desired single file by clicking [...] path button.

    mybucket/dbo.tblNames.json
    dbo.tblNames.json
    Read Secure FTP JSON File data


    ----------OR----------

    You can also read the multiple files stored in Secure FTP Storage using wildcard pattern supported e.g. dbo.tblNames*.json.

    Note: If you want to operation with multiple files then use wild card pattern as below 
    (when you use wild card pattern in source path then system will treat target path as folder regardless you end with slash)
    
    mybucket/dbo.tblNames.json (will read only single .JSON file)
    mybucket/dbo.tbl*.json (all files starting with file name)
    mybucket/*.json (all files with .json Extension and located under folder subfolder)
    

    mybucket/dbo.tblNames*.json
    Use wildcard pattern .* to read multiple Secure FTP Files data


    ----------OR----------

    You can also read the zip and gzip compressed files also without extracting it in using Secure FTP JSON File Source.

    Reading zip and gzip compressed files (stream mode)
  9. Now select/enter Path expression in Path textbox to extract only specific part of JSON string as below ($.value[*] will get content of value attribute from JSON document. Value attribute is array of JSON documents so we have to use [*] to indicate we want all records of that array)

    NOTE: Here, We are using our desired filter, but you need to select your desired filter based on your requirement.
    Click on Preview button to view the parsed JSON string response data and click OK.

    Secure FTP JSON Files data Preview
  10. That's it; we are done. In a few clicks we configured the to Read the FTP/SFTP JSON File data using ZappySys FTP/SFTP JSON File Connector

Load FTP/SFTP JSON File data into SQL Server using Upsert Destination (Insert or Update)

Once you configured the data source, you can load FTP/SFTP JSON File data into SQL Server using Upsert Destination.

Upsert Destination can merge or synchronize source data with the target table. It supports Microsoft SQL Server, PostgreSQL, and Redshift databases as targets. Upsert Destination also supports very fast bulk upsert operation along with bulk delete.

Upsert operation - a database operation which performs INSERT or UPDATE SQL commands based on record's existence condition in the target table. It inserts records that don't have matching records in the target table or updates them, if they do, by matching them by key columns.

Upsert Destination supports INSERT, UPDATE, and DELETE operations, so it is similar to SQL Server's MERGE command, except it can be used directly in SSIS package.

  1. From the SSIS Toolbox drag-and-drop Upsert Destination component onto the Data Flow designer background.

  2. Connect your SSIS source component to Upsert Destination.

  3. Double-click on Upsert Destination component to open configuration window.

  4. Start by selecting the Action from the list.

  5. Next, select the desired target connection or create one by clicking <New [provider] Connection> menu item from the Target Connection dropdown.

  6. Then select a table from the Target Table list or click New button to create a new table based on the source columns.

  7. Continue by checking Insert and Update options according to your scenario (e.g. if Update option is unchecked, no updates will be made).

  8. Finally, click Map All button to map all columns and then select the Key columns to match the columns on:

    Configure SSIS Upsert Destination component to merge data with SQL Server, PostgreSQL, or Redshift table
  9. Click OK to save the configuration.

  10. Run the package and FTP/SFTP JSON File data will be merged with the target table in SQL Server, PostgreSQL, or Redshift:

    Execute Package - Reading data from API Source and load into target
  11. Done!

Deploy and schedule SSIS package

After you are done creating SSIS package, most likely, you want to deploy it to SQL Server Catalog and run it periodically. Just follow the instructions in this article:

Running SSIS package in Azure Data Factory (ADF)

To use SSIS PowerPack in ADF, you must first prepare Azure-SSIS Integration Runtime. Follow this link for detailed instructions:

Conclusion

In this article we showed you how to connect to FTP/SFTP JSON File in SSIS and integrate data without writing complex code — all of this was powered by FTP/SFTP JSON File Connector.

Download SSIS PowerPack now or ping us via chat if you have any questions or are looking for a specific feature (you can also reach out to us by submitting a ticket):

Explore SSIS connectors

All
Big Data & NoSQL
Database
CRM & ERP
Marketing
Collaboration
Cloud Storage
Reporting
Commerce
API & Files

More FTP/SFTP JSON File integrations

All
Data Integration
Database
BI & Reporting
Productivity
Programming Languages
Automation & Scripting
ODBC applications