Shopify Connector for Azure Data Factory (SSIS)
Shopify Connector can be used to integrate Shopify REST API in your App / BI Tools. You can read/write data about Customers, Orders, OrderItems, Products and more. You can use this connector to integrate Shopify data inside SSIS and SQL Server. Let's take a look at the steps below to see how exactly to accomplish that.
Download
Documentation
If you are new with SSIS and ZappySys SSIS PowerPack then check below links to get started.
|
Connect to Shopify in other apps
|
Create SSIS package
First of all, create an SSIS package, which will connect to Shopify in SSIS.
Once you do that, you are one step closer to deploying and running it in Azure-SSIS integration runtime in Azure Data Factory (ADF).
Then simply proceed to the next step - creating and configuring Azure Blob Storage Container.
Prepare custom setup files for Azure-SSIS runtime
Now it's time to start preparing custom setup files for Azure-SSIS runtime.
During Azure-SSIS runtime creation you can instruct ADF to perform a custom setup on a VM (Azure-SSIS node);
i.e. to run the custom installer, copy files, execute PowerShell scripts, etc.
In that case, your custom setup files are downloaded and run in the Azure-SSIS node (a VM) when you start the runtime.
In this section we will prepare custom setup files so that you can run SSIS packages with SSIS PowerPack connectors inside in Azure-SSIS runtime.
Trial Users
Use the step below if you are a Trial User, when you did not purchase a license key.
Proceed with these steps:
-
Download SSIS PowerPack trial installer.
Make sure you don't rename the installer and keep it named as SSISPowerPackSetup_64bit_Trial.msi.
-
Create a text file and name it main.cmd (make it all lowercase, very important).
-
Copy and paste this script into it and save it:
set DIR=%CUSTOM_SETUP_SCRIPT_LOG_DIR%
echo Calling Step 1 : %TIME% >> "%DIR%\steps_log.txt"
dir /s /b > "%DIR%\file_list.txt"
echo Calling Step 2 : %TIME% >> "%DIR%\steps_log.txt"
::Install SSIS PowerPack
msiexec /i "SSISPowerPackSetup_64bit_Trial.msi" ADDLOCAL=ALL /q /L*V "%DIR%\powerpack_trial_install_log.txt"
echo Calling Step 3 : %TIME% >> "%DIR%\steps_log.txt"
dir "C:\Program Files\Microsoft SQL Server\*Zappy*.*" /s /b >> "%DIR%\installed_files.txt"
dir "C:\Program Files (x86)\Microsoft SQL Server\*Zappy*.*" /s /b >> "%DIR%\installed_files.txt"
echo DONE : %TIME% >> "%DIR%\steps_log.txt"
echo complete
This is the entry-point script that is executed when Azure-SSIS runtime is started.
-
At last! You are ready to upload these two files — main.cmd & SSISPowerPackSetup_64bit_Trial.msi — into your Azure Blob Storage container's folder, which we will do in the Upload custom setup files to Azure Blob Storage container step.
Paid Customers
Use the steps below if you are a Paid Customer, when you purchased a license.
Proceed with these steps:
-
Download SSIS PowerPack paid installer.
Make sure you don't rename the installer and keep it named as SSISPowerPackSetup_64bit.msi.
-
Have your SSIS PowerPack license key handy, we will need it in the below script.
-
Create a text file and name it main.cmd (make it all lowercase, very important).
- Copy and paste the below script into it.
- Paste your license key by replacing parameter's
--register
argument with your real license key.
-
Finally, save main.cmd:
set DIR=%CUSTOM_SETUP_SCRIPT_LOG_DIR%
echo Calling Step 1 : %TIME% >> "%DIR%\steps_log.txt"
dir /s /b > "%DIR%\file_list.txt"
echo Calling Step 2 : %TIME% >> "%DIR%\steps_log.txt"
::Install SSIS PowerPack
msiexec /i "SSISPowerPackSetup_64bit.msi" ADDLOCAL=ALL /q /L*V "%DIR%\powerpack_install_log.txt"
echo Calling Step 3 : %TIME% >> "%DIR%\steps_log.txt"
::Activate PowerPack license (Optional)
"C:\Program Files (x86)\ZappySys\SSIS PowerPack (64 bit)\LicenseManager.exe" -p SSISPowerPack --register "lgGAAO0-----REPLACE-WITH-YOUR-LICENSE-KEY-----czM=" --logfile "%DIR%\powerpack_register_log.txt"
::Show System Info
echo Calling Step 4 : %TIME% >> "%DIR%\steps_log.txt"
"C:\Program Files (x86)\ZappySys\SSIS PowerPack (64 bit)\LicenseManager.exe" -i -l "%DIR%\sysinfo_log.txt"
echo Calling Step 5 : %TIME% >> "%DIR%\steps_log.txt"
dir "C:\Program Files\Microsoft SQL Server\*Zappy*.*" /s /b >> "%DIR%\installed_files.txt"
dir "C:\Program Files (x86)\Microsoft SQL Server\*Zappy*.*" /s /b >> "%DIR%\installed_files.txt"
echo DONE : %TIME% >> "%DIR%\steps_log.txt"
echo complete
This is the entry-point script that is executed when Azure-SSIS runtime is started.
-
At last! You are ready to upload these two files — main.cmd & SSISPowerPackSetup_64bit.msi — into your Azure Blob Storage container's folder, which we will do in the Upload custom setup files to Azure Blob Storage container step.
Upload custom setup files to Azure Blob Storage container
Within Azure Blob Storage container we will store custom setup files we prepared in the previous step so that Azure-SSIS can use them in custom setup process.
Just perform these very simple, but very important steps:
-
Create Azure Blob Storage container, if you haven't done it already
Make sure you create and use Azure Blob Storage container instead of Azure Data Lake Storage folder.
Azure Data Lake Storage won't allow creating an SAS URI for the container, which is a crucial step in the process.
-
Find Blob Containers node, right-click on it and hit Create Blob Container option:
-
Upload the two custom setup files — main.cmd & the MSI installer — into your Azure Blob Storage container's folder:
- It was easy, wasn't it? It's time we create an SAS URI in the next step.
Create SAS URI for Azure Blob Container
Once you have custom setup files prepared, it's time we generate an SAS URI.
This SAS URI will be used by a new Azure-SSIS runtime to install SSIS PowerPack inside the runtime's node, a VM.
Let's proceed together by performing the steps below:
- Install and launch Azure Storage Explorer.
-
Right-click on the Storage Accounts node and then hit Connect to Azure storage... menu item:
-
Proceed by right-clicking on that container node and select Get Shared Access Signature... option.
-
Next, set the Expiry time field to a date far in the future.
If you restart Azure-SSIS runtime and your SAS URI is expired by that time, it will not start.
-
Select Read, Create, Write, and List permissions:
We also recommend to add Delete permission too to support future functionality.
-
Copy SAS URL to the clipboard and save it for the next step:
You can also generate and copy SAS URL from within Azure Portal itself:
Create Azure-SSIS integration runtime
Once you have the SAS URL we obtained in the previous step, we are ready to move on to create an Azure-SSIS runtime in Azure Data Factory:
-
Firstly, perform the steps described in Create an Azure-SSIS integration runtime article in Azure Data Factory reference.
-
In Advanced settings page section, configure Custom setup container SAS URI you obtained in the previous step:
-
And you are done! That was quick! You can see your Azure-SSIS runtime up and running:
The custom setup script is executed only once — at the time an Azure-SSIS runtime is started.
It is also executed if you stop and start Azure-SSIS runtime again.
Deploy SSIS package in Visual Studio
We are ready to deploy the SSIS package to Azure-SSIS runtime. Once you do that, proceed to the next step for the grand finale!
Execute SSIS package in SQL Server Management Studio (SSMS)
After all hard work, we are ready to execute SSIS package in SQL Server Management Studio (SSMS):
- Connect to the SQL Server which is linked to your Azure-SSIS runtime and contains SSISDB database.
-
Navigate to Integration Services Catalog » Your Folder » Your Project » Your Package, right-click on it, and hit Execute...:
-
To view the status of the past execution, navigate to
Integration Services Catalog » Your Folder » Your Project » Your Package, right-click on it, and select Reports » Standard Reports » All Executions menu item:
Scenarios
Moving SSIS PowerPack license to another Azure-SSIS runtime
If you are a Paid Customer, there will be a time when you no longer use Azure-SSIS runtime or you need to use your license on a different ADF instance.
To transfer a license from one Azure-SSIS runtime to another, perform these steps:
-
Copy & paste this script into main.cmd we used in the previous step:
set DIR=%CUSTOM_SETUP_SCRIPT_LOG_DIR%
echo Calling Step 1 : %TIME% >> "%DIR%\steps_log.txt"
dir /s /b > "%DIR%\file_list.txt"
echo Calling Step 2 : %TIME% >> "%DIR%\steps_log.txt"
::Install SSIS PowerPack
msiexec /i "SSISPowerPackSetup_64bit.msi" ADDLOCAL=ALL /q /L*V "%DIR%\powerpack_install_log.txt"
echo Calling Step 3 : %TIME% >> "%DIR%\steps_log.txt"
::De-Activate same license
"C:\Program Files (x86)\ZappySys\SSIS PowerPack (64 bit)\LicenseManager.exe" -p SSISPowerPack --unregister --logfile "%DIR%\powerpack_un_register_log.txt"
::Show System Info
echo Calling Step 4 : %TIME% >> "%DIR%\steps_log.txt"
"C:\Program Files (x86)\ZappySys\SSIS PowerPack (64 bit)\LicenseManager.exe" -i -l "%DIR%\sysinfo_log.txt"
echo Calling Step 5 : %TIME% >> "%DIR%\steps_log.txt"
dir "C:\Program Files\Microsoft SQL Server\*Zappy*.*" /s /b >> "%DIR%\installed_files.txt"
dir "C:\Program Files (x86)\Microsoft SQL Server\*Zappy*.*" /s /b >> "%DIR%\installed_files.txt"
echo DONE : %TIME% >> "%DIR%\steps_log.txt"
echo complete
- Start Azure-SSIS runtime.
This will unregister your license on the original Azure-SSIS runtime.
- Stop Azure-SSIS runtime to deallocate resources in Azure.
- Now you are free to activate it on another Azure-SSIS runtime.
Advanced topics
Actions supported by Shopify Connector
Shopify Connector support following actions for REST API integration. If some actions are not listed below then you can easily edit Connector file and enhance out of the box functionality.
Gets a list of customers. [
Read more...]
Parameter |
Description |
Customer Id(s) - Comma separated |
Restrict results to records specified by a comma-separated list of IDs. Max IDs per request controlled by BatchSize property (found in PaginationParams)
|
Gets a list of orders by status, date and other search criteria. By default only Open orders returned. Change Status parameter to any to get all orders. [
Read more...]
Parameter |
Description |
Order Id(s) - Comma separated |
Restrict results to records specified by a comma-separated list of IDs. Max IDs per request controlled by BatchSize property (found in PaginationParams)
|
Gets a list of line items for all orders or the specified order(s). [
Read more...]
Parameter |
Description |
Order Id(s) - Comma separated |
Restrict results to records specified by a comma-separated list of IDs. Max IDs per request controlled by BatchSize property (found in PaginationParams)
|
Gets a list of products. [
Read more...]
Parameter |
Description |
Product Id(s) - Comma separated |
Restrict results to records specified by a comma-separated list of IDs. Max IDs per request controlled by BatchSize property (found in PaginationParams)
|
Gets a list of products with their variants. [
Read more...]
Parameter |
Description |
Product Id(s) - Comma separated |
Restrict results to records specified by a comma-separated list of IDs. Max IDs per request controlled by BatchSize property (found in PaginationParams)
|
Since Product Id |
Restrict results created after the specified product ID.
|
Only Fields to Show |
Limit data retrieval to only the selected product-related fields.
Option |
Value |
id |
id |
title |
title |
body_html |
body_html |
vendor |
vendor |
product_type |
product_type |
created_at |
created_at |
handle |
handle |
updated_at |
updated_at |
published_at |
published_at |
template_suffix |
template_suffix |
status |
status |
published_scope |
published_scope |
tags |
tags |
admin_graphql_api_id |
admin_graphql_api_id |
variants |
variants |
options |
options |
images |
images |
image |
image |
|
Created Before |
Restrict results to products created before a specified date. (format: 2014-04-25T16:15:47-04:00)
|
Created After |
Restrict results to products created after a specified date. (format: 2014-04-25T16:15:47-04:00)
|
Updated Before |
Restrict results to products last updated before a specified date. (format: 2014-04-25T16:15:47-04:00)
|
Updated After |
Restrict results to products last updated after a specified date. (format: 2014-04-25T16:15:47-04:00)
|
Creates a new product variant. [
Read more...]
Parameter |
Description |
Product Id |
The ID of the product for which you like to create a new variant.
|
Retrieves a list of all payouts ordered by payout date, with the most recent being first. Requires following permissions shopify_payments_payouts, shopify_payments for API Token. [
Read more...]
Parameter |
Description |
Date (format: yyyy-MM-ddd) |
Filter the response to payouts made on the specified date.
Option |
Value |
2023-12-31 |
2023-12-31 |
today |
today |
yesterday |
yesterday |
monthstart |
monthstart |
monthend |
monthend |
yearstart |
yearstart |
yearend |
yearend |
weekstart |
weekstart |
weekend |
weekend |
yearstart-1y |
yearstart-1y |
yearend-1y |
yearend-1y |
monthstart-1d |
monthstart-1d |
monthend+1d |
monthend+1d |
|
Date maximum (format: yyyy-MM-ddd) |
Filter the response to payouts made inclusively before the specified date.
Option |
Value |
2023-12-31 |
2023-12-31 |
today |
today |
yesterday |
yesterday |
monthstart |
monthstart |
monthend |
monthend |
yearstart |
yearstart |
yearend |
yearend |
weekstart |
weekstart |
weekend |
weekend |
yearstart-1y |
yearstart-1y |
yearend-1y |
yearend-1y |
monthstart-1d |
monthstart-1d |
monthend+1d |
monthend+1d |
|
Date minimum(format: yyyy-MM-ddd) |
Filter the response to payouts made inclusively after the specified date.
Option |
Value |
2023-12-31 |
2023-12-31 |
today |
today |
yesterday |
yesterday |
monthstart |
monthstart |
monthend |
monthend |
yearstart |
yearstart |
yearend |
yearend |
weekstart |
weekstart |
weekend |
weekend |
yearstart-1y |
yearstart-1y |
yearend-1y |
yearend-1y |
monthstart-1d |
monthstart-1d |
monthend+1d |
monthend+1d |
|
Payouts before this Id |
Filter the response to payouts made before the specified ID.
|
Payouts after this Id |
Filter the response to payouts made after the specified ID.
|
Status |
Filter the response to payouts made with the specified status.
Option |
Value |
scheduled |
scheduled |
in_transit |
in_transit |
paid |
paid |
failed |
failed |
canceled |
canceled |
|
Gets a specific location. [
Read more...]
Parameter |
Description |
Location Id |
The ID of the location for which data should be read.
|
Gets a list of inventory items. Each Product Variant has Inventory Item Id .. Its one to one relationship with Product Variant and Inventory Item. Query get_product_variants to get InventoryItemId and use as Id in this endpoint. [
Read more...]
Parameter |
Description |
Inventory Item Id(s) - Comma separated |
Restrict results to records specified by a comma-separated list of IDs. Max IDs per request controlled by BatchSize property (found in PaginationParams)
|
Gets inventory level for a specific location and inventory item. [
Read more...]
Parameter |
Description |
inventory_item_ids |
Inventory Item Id(s) - Comma separated
|
location_ids |
Location Id(s) - Comma separated
|
Updated at or after |
Show inventory levels updated at or after date (format: 2019-03-19T01:21:44-04:00).
|
[
Read more...]
Parameter |
Description |
Action |
Option |
Value |
Adjusts the inventory level |
adjust |
Connects an inventory item to a location |
connect |
Sets the inventory level for an inventory item at a location |
set |
|
Gets a specific inventory item. [
Read more...]
Parameter |
Description |
Inventory Item Id(s) - Comma separated |
Restrict results to records specified by a comma-separated list of IDs. Max IDs per request controlled by BatchSize property (found in PaginationParams)
|
This is generic endpoint. Use this endpoint when some actions are not implemented by connector. Just enter partial URL (Required), Body, Method, Header etc. Most parameters are optional except URL. [
Read more...]
Parameter |
Description |
Url |
API URL goes here. You can enter full URL or Partial URL relative to Base URL. If it is full URL then domain name must be part of ServiceURL or part of TrustedDomains
|
Body |
Request Body content goes here
|
IsMultiPart |
Set this option if you want to upload file(s) (i.e. POST RAW file data) or send data using Multi-Part encoding method (i.e. Content-Type: multipart/form-data). Multi-Part request allows you to mix key/value and upload files in same request. On the other hand raw upload allows only single file upload (without any key/value)
==== Raw Upload (Content-Type: application/octet-stream) =====
To upload single file in raw mode check this option and specify full file path starting with @ sign in the Body (e.g. @c:\data\myfile.zip )
==== Form-Data / Multipart Upload (Content-Type: multipart/form-data) =====
To treat your Request data as multi part fields you must specify key/value pairs separated by new lines into RequestData field (i.e. Body). Each key value pair is entered on new-line and key/value are separated using equal sign (=). Preceding and trailing spaces are ignored also blank lines are ignored.
If field value has some any special character(s) then use escape sequence (e.g. For NewLine: \r\n, For Tab: \t, For at (@): \@). When value of any field starts with at sign (@) its automatically treated as File you want to upload. By default file content type is determined based on extension however you can supply content type manually for any field using this way [ YourFileFieldName.Content-Type=some-content-type ]. By default File Upload Field always includes Content-Type in the request (non file fields do not have content-type by default unless you supply manually). For some reason if you dont want to use Content-Type header in your request then supply blank Content-Type to exclude this header altogather [e.g. SomeFieldName.Content-Type= ]. In below example we have supplied Content-Type for file2 and SomeField1, all other fields are using default content-type.
See below Example of uploading multiple files along with additional fields. If some API requires you to pass Content-Type: multipart/form-data rather than multipart/form-data then manually set Request Header => Content-Type: multipart/mixed (it must starts with multipart/ else will be ignored).
file1=@c:\data\Myfile1.txt
file2=@c:\data\Myfile2.json
file2.Content-Type=application/json
SomeField1=aaaaaaa
SomeField1.Content-Type=text/plain
SomeField2=12345
SomeFieldWithNewLineAndTab=This is line1\r\nThis is line2\r\nThis is \ttab \ttab \ttab
SomeFieldStartingWithAtSign=\@MyTwitterHandle
|
Filter |
Enter filter to extract array from response. Example: $.rows[*] --OR-- $.customers[*].orders[*]. Check your response document and find out hierarchy you like to extract
|
Headers |
Headers for Request. To enter multiple headers use double pipe or new line after each {header-name}:{value} pair
|
Conclusion
In this article we discussed how to connect to Shopify in Azure Data Factory (SSIS) and integrate data without any coding. Click here to Download Shopify Connector for Azure Data Factory (SSIS) and try yourself see how easy it is. If you still have any question(s) then ask here or simply click on live chat icon below and ask our expert (see bottom-right corner of this page).
Download Shopify Connector for Azure Data Factory (SSIS)
Documentation
More integrations
Other application integration scenarios for Shopify
Other connectors for Azure Data Factory (SSIS)
Download Shopify Connector for Azure Data Factory (SSIS)
Documentation
Common Searches:
How to connect Shopify in Azure Data Factory (SSIS)?
How to get Shopify data in Azure Data Factory (SSIS)?
How to read Shopify data in Azure Data Factory (SSIS)?
How to load Shopify data in Azure Data Factory (SSIS)?
How to import Shopify data in Azure Data Factory (SSIS)?
How to pull Shopify data in Azure Data Factory (SSIS)?
How to push data to Shopify in Azure Data Factory (SSIS)?
How to write data to Shopify in Azure Data Factory (SSIS)?
How to POST data to Shopify in Azure Data Factory (SSIS)?
Call Shopify API in Azure Data Factory (SSIS)
Consume Shopify API in Azure Data Factory (SSIS)
Shopify Azure Data Factory (SSIS) Automate
Shopify Azure Data Factory (SSIS) Integration
Integration Shopify in Azure Data Factory (SSIS)
Consume real-time Shopify data in Azure Data Factory (SSIS)
Consume real-time Shopify API data in Azure Data Factory (SSIS)
Shopify ODBC Driver | ODBC Driver for Shopify | ODBC Shopify Driver | SSIS Shopify Source | SSIS Shopify Destination
Connect Shopify in Azure Data Factory (SSIS)
Load Shopify in Azure Data Factory (SSIS)
Load Shopify data in Azure Data Factory (SSIS)
Read Shopify data in Azure Data Factory (SSIS)
Shopify API Call in Azure Data Factory (SSIS)