Amazon S3 CSV File Connector for Microsoft Fabric

Amazon S3 CSV File Connector can be used to read CSV Files stored in AWS S3 Buckets. Using this you can easily integrate AWS S3 CSV File data. It's supports latest security standards, and optimized for large data files. It also supports reading compressed files (e.g. GZip /Zip).

In this article you will learn how to quickly and efficiently integrate Amazon S3 CSV File data in Microsoft Fabric without coding. We will use high-performance Amazon S3 CSV File Connector to easily connect to Amazon S3 CSV File and then access the data inside Microsoft Fabric.

Let's follow the steps below to see how we can accomplish that!

Download Documentation

Create ODBC Data Source (DSN) based on ZappySys Amazon S3 CSV Driver

Step-by-step instructions

To get data from Amazon S3 CSV File using Microsoft Fabric we first need to create a DSN (Data Source) which will access data from Amazon S3 CSV File. We will later be able to read data using Microsoft Fabric. Perform these steps:

  1. Download and install ODBC PowerPack.

  2. Open ODBC Data Sources (x64):

    Open ODBC Data Source
  3. Create a User data source (User DSN) based on ZappySys Amazon S3 CSV Driver:

    ZappySys Amazon S3 CSV Driver
    Create new User DSN for ZappySys Amazon S3 CSV Driver
    • Create and use User DSN if the client application is run under a User Account. This is an ideal option in design-time, when developing a solution, e.g. in Visual Studio 2019. Use it for both type of applications - 64-bit and 32-bit.
    • Create and use System DSN if the client application is launched under a System Account, e.g. as a Windows Service. Usually, this is an ideal option to use in a production environment. Use ODBC Data Source Administrator (32-bit), instead of 64-bit version, if Windows Service is a 32-bit application.
  4. Create and configure a connection for the Amazon S3 storage account.

    Create Amazon S3 Storage Connection
  5. You can use select your desired single file by clicking [...] path button.

    mybucket/dbo.tblNames.csv
    dbo.tblNames.csv
    Read Amazon S3 CSV File data


    ----------OR----------

    You can also read the multiple files stored in Amazon S3 Storage using wildcard pattern supported e.g. dbo.tblNames*.csv.

    Note: If you want to operation with multiple files then use wild card pattern as below 
    (when you use wild card pattern in source path then system will treat target path as folder regardless you end with slash)
    
    mybucket/dbo.tblNames.csv (will read only single .CSV file)
    mybucket/dbo.tbl*.csv (all files starting with file name)
    mybucket/*.csv (all files with .csv Extension and located under folder subfolder)
    

    mybucket/dbo.tblNames*.csv
    Use wildcard pattern .* to read multiple Amazon S3 Files data


    ----------OR----------

    You can also read the zip and gzip compressed files also without extracting it in using Amazon S3 CSV Source File Task.

    mybucket/dbo.tblNames*.gz
    Reading zip and gzip compressed files (stream mode)
  6. Navigate to the Preview Tab and let's explore the different modes available to access the data.

    1. --- Using Direct Query ---

      Click on Preview Tab, Select Table from Tables Dropdown and select [value] and click Preview.
      ZappySys ODBC Driver - Preview Data
    2. --- Using Stored Procedure ---

      Note : For this you have to Save ODBC Driver configuration and then again reopen to configure same driver. For more information click here.
      
      Click on the Custom Objects Tab, Click on Add button and select Add Procedure and Enter an appropriate name and Click on OK button to create.
      ZappySys ODBC Driver - Custom Objects
      1. --- Without Parameters ---

        Now Stored Procedure can be created with or without parameters (see example below). If you use parameters then Set default value otherwise it may fail to compilation)
        ZappySys ODBC Driver : Without Parameters
      2. --- With Parameters ---

        Note : Here you can use Placeholder with Paramters in Stored Procedure.
        Example : SELECT * FROM $ WHERE OrderID = '<@OrderID, FUN_TRIM>' or CustId = '<@CustId>' and Total >= '<@Total>'
        
        ZappySys ODBC Driver : With Parameters
    3. --- Using Virtual Table ---

      Note : For this you have to Save ODBC Driver configuration and then again reopen to configure same driver. For more information click here.
      

      ZappySys APi Drivers support flexible Query language so you can override Default Properties you configured on Data Source such as URL, Body. This way you don't have to create multiple Data Sources if you like to read data from multiple EndPoints. However not every application support supplying custom SQL to driver so you can only select Table from list returned from driver.

      Many applications like MS Access, Informatica Designer wont give you option to specify custom SQL when you import Objects. In such case Virtual Table is very useful. You can create many Virtual Tables on the same Data Source (e.g. If you have 50 Buckets with slight variations you can create virtual tables with just URL as Parameter setting).

      vt__Customers
      DataPath=mybucket_1/customers.csv
      
      vt__Orders
      DataPath=mybucket_2/orders.csv
      
      vt__Products
      DataPath=mybucket_3/products.csv
      
      1. Click on the Custom Objects Tab, Click on Add button and select Add Table and Enter an appropriate name and Click on OK button to create.
        ZappySys ODBC Driver - Custom Objects
      2. Once you see Query Builder Window on screen Configure it.
        ZappySys ODBC Driver - Custom Objects : Virtual Table Query Builder
      3. Click on Preview Tab, Select Virtual Table(prefix with vt__) from Tables Dropdown or write SQL query with Virtual Table name and click Preview.
        ZappySys ODBC Driver - Custom Objects : Virtual Table Query Execute

  7. Click OK to finish creating the data source

  8. That's it; we are done. In a few clicks we configured the to Read the Amazon S3 CSV File data using ZappySys Amazon S3 CSV File Connector

Installing Microsoft On-premises data gateway

To access and read Amazon S3 CSV File data in Microsoft Fabric, you will have to download and install Microsoft On-premises data gateway.

There are two types of On-premises data gateways:

  • Standard Mode
  • Personal Mode

Standard Mode supports Power BI and other Microsoft Data Fabric services. It fits perfectly for Enterprise solutions as it installs as a Windows Service and also supports Direct Query feature.

Personal Mode, on the other hand, can be configured faster, but is designed more for home users (you cannot install it as a Windows Service and it does not support DirectQuery). You will find a detailed comparison in the link above.

We recommend to go with Personal Mode for a quick POC solution, but use Standard Mode in production environment. You can download On-premises data gateway straight from Microsoft Data Fabric or Power BI web app:

Download Power BI On-premises data gateway

Standard mode

Follow these steps how to configure On-premises data gateway in Standard mode:

  1. Install On-premises data gateway (standard mode) and sign-in:

    Signing in into on-premises data gateway standard
    Use the same email address you use when logging in into your account.
  2. Register a new gateway (or migrate an existing one):

    Registering or migrating on-premises data gateway standard
  3. If you are creating a new gateway, name your gateway, enter a Recovery key, and click Configure button:

    Naming on-premises data gateway standard
  4. Once Microsoft gateway is installed, check if it registered correctly:

    • Go back to Microsoft Fabric portal

    • Click Gear icon on top-right

    • And then hit Manage connections and gateways menu item

    Manage On-premise data gateways in Microsoft Fabric or Power BI
  5. Continue by clicking On-premises data gateway tab and select Standard mode gateways option from the dropdown:

    Access On-premises data gateway list (Standard mode) in Microsoft Fabric

    If your gateway is not listed, the registration may have failed. To resolve this:

    • Wait a couple of minutes and refresh Microsoft Fabric portal page
    • Restart the machine where On-premises data gateway is installed
    • Check firewall settings
  6. Done!

Now we are ready to read data in Microsoft Fabric.

Personal mode

Follow this single step to configure On-premises data gateway in Personal mode:

  1. Install On-premises data gateway (personal mode) and sign-in:

    Sign-in to On-premises data gateway personal
    Use the same email address you use when logging in into your account.
  2. Once Microsoft gateway is installed, check if it registered correctly:

    • Go back to Microsoft Fabric portal

    • Click Gear icon on top-right

    • And then hit Manage connections and gateways menu item

    Manage On-premise data gateways in Microsoft Fabric or Power BI
  3. Continue by clicking On-premises data gateway tab and select Personal mode option from the dropdown:

    Access On-premises data gateway list (Personal mode) in Microsoft Fabric

    If your gateway is not listed, the registration may have failed. To resolve this:

    • Wait a couple of minutes and refresh Microsoft Fabric portal page
    • Restart the machine where On-premises data gateway is installed
    • Check firewall settings
  4. Done!

Now we are ready to read data in Microsoft Fabric.

Loading Amazon S3 CSV File data into Microsoft Fabric

We created ODBC data source and installed Microsoft On-premises data gateway, and we are ready to load the data into Microsoft Fabric.

Let's roll!

  1. Log in to the Microsoft Fabric Portal.

  2. Select an existing Workspace or create a new one by clicking New workspace (make sure you are in Home section):

    Create a new workspace in Microsoft Fabric for a Copy job
  3. Inside your workspace, click the New item button in the toolbar to start creating your data pipeline:

    Create new item in Microsoft Fabric workspace
  4. In the item selection window, choose Copy job to open the data ingestion wizard:

    Add Copy job to Microsoft Fabric workspace
  5. In the Choose data source screen, search for odbc and select the Odbc source:

    Choose ODBC as the data source in Microsoft Fabric Copy job
  6. Then enter your ODBC connection string (DSN=AmazonS3CsvFileDSN) and select MyGateway from the Data gateway dropdown we configured in the previous step:

    DSN=AmazonS3CsvFileDSN
    DSN=AmazonS3CsvFileDSN
    Configure ODBC connection string in Microsoft Fabric Copy job
  7. Select the table(s) and preview the data you wish to copy from Amazon S3 CSV File. Once done, click Next:

    DSN=AmazonS3CsvFileDSN
    Selecting tables to copy in Microsoft Fabric Copy Job
  8. Choose your Data Destination. You can create a New Fabric item (like a Lakehouse or Warehouse) or select an existing one:

    Choose data destination in Microsoft Fabric Copy job
    As an example, we will be using Lakehouse as destination type in this article
  9. Choose Full copy to load all data or Incremental copy if you want to load only changed data in subsequent runs:

    Select copy mode in Microsoft Fabric Copy job (Full vs Incremental)
  10. Review the Column and Table mappings section:

    Map source tables and columns to destination in Microsoft Fabric Copy job
  11. On the summary screen, review your settings. You can optionally enable Run on schedule. Click Save + Run to execute the job:

    DSN=AmazonS3CsvFileDSN
    DSN=AmazonS3CsvFileDSN
    Save and run the Copy job in Microsoft Fabric
  12. The job will enter the queue. Monitor the Status column to see the progress:

    DSN=AmazonS3CsvFileDSN
    Monitor the status of the Microsoft Fabric Copy job
  13. Wait for the status to change to Succeeded. Your Amazon S3 CSV File data is now successfully integrated into Microsoft Fabric!

    Verify Microsoft Fabric Copy job success status
  14. Let's go to our Lakehouse MyLakehouse and double-check if it's true:

    Load data into Microsoft Fabric Lakehouse
  15. It's true, done!

Centralized data access via Data Gateway

In some situations, you may need to provide Amazon S3 CSV File data access to multiple users or services. Configuring the data source on a Data Gateway creates a single, centralized connection point for this purpose.

This configuration provides two primary advantages:

  • Centralized data access
    The data source is configured once on the gateway, eliminating the need to set it up individually on each user's machine or application. This significantly simplifies the management process.
  • Centralized access control
    Since all connections route through the gateway, access can be governed or revoked from a single location for all users.
Data Gateway
Local ODBC
data source
Simple configuration
Installation Single machine Per machine
Connectivity Local and remote Local only
Connections limit Limited by License Unlimited
Central data access
Central access control
More flexible cost

If you need any of these requirements, you will have to create a data source in Data Gateway to connect to Amazon S3 CSV File, and to create an ODBC data source to connect to Data Gateway in Microsoft Fabric.

Let's not wait and get going!

Creating Amazon S3 CSV File data source in Gateway

In this section we will create a data source for Amazon S3 CSV File in Data Gateway. Let's follow these steps to accomplish that:

  1. Search for gateway in Windows Start Menu and open ZappySys Data Gateway Configuration:

    Opening Data Gateway
  2. Go to Users tab and follow these steps to add a Data Gateway user:

    • Click Add button
    • In Login field enter username, e.g., john
    • Then enter a Password
    • Check Is Administrator checkbox
    • Click OK to save
    Data Gateway - Adding User
  3. Now we are ready to add a data source:

    • Click Add button
    • Give Datasource a name (have it handy for later)
    • Then select Native - ZappySys Amazon S3 CSV Driver
    • Finally, click OK
    AmazonS3CsvFileDSN
    ZappySys Amazon S3 CSV Driver
    Data Gateway - Adding data source
  4. When the ZappySys Amazon S3 CSV Driver configuration window opens, configure the Data Source the same way you configured it in ODBC Data Sources (64-bit), in the beginning of this article.

  5. Very important step. Now, after creating or modifying the data source make sure you:

    • Click the Save button to persist your changes.
    • Hit Yes, once asked if you want to restart the Data Gateway service.

    This will ensure all changes are properly applied:

    ZappySys Data Gateway - Save Changes
    Skipping this step may result in the new settings not taking effect and, therefore you will not be able to connect to the data source.

Creating ODBC data source for Data Gateway

In this part we will create ODBC data source to connect to Data Gateway from Microsoft Fabric. To achieve that, let's perform these steps:

  1. Open ODBC Data Sources (x64):

    Open ODBC Data Source
  2. Create a User data source (User DSN) based on ODBC Driver 17 for SQL Server:

    ODBC Driver 17 for SQL Server
    Create new User DSN for ODBC Driver 17 for SQL Server
    If you don't see ODBC Driver 17 for SQL Server driver in the list, choose a similar version driver.
  3. Then set a Name of the data source (e.g. Gateway) and the address of the Data Gateway:

    GatewayDSN
    localhost,5000
    ODBC driver for SQL Server - Setting hostname and port
    Make sure you separate the hostname and port with a comma, e.g. localhost,5000.
  4. Proceed with authentication part:

    • Select SQL Server authentication
    • In Login ID field enter the user name you used in Data Gateway, e.g., john
    • Set Password to the one you configured in Data Gateway
    ODBC driver for SQL Server - Selecting SQL Authentication
  5. Then set the default database property to AmazonS3CsvFileDSN (the one we used in Data Gateway):

    AmazonS3CsvFileDSN
    ODBC driver for SQL Server - Selecting database
  6. Continue by checking Trust server certificate option:

    ODBC driver for SQL Server - Trusting certificate
  7. Once you do that, test the connection:

    ODBC driver for SQL Server - Testing connection
  8. If connection is successful, everything is good:

    ODBC driver for SQL Server - Testing connection succeeded
  9. Done!

We are ready to move to the final step. Let's do it!

Accessing data in Microsoft Fabric via Data Gateway

Finally, we are ready to read data from Amazon S3 CSV File in Microsoft Fabric via Data Gateway. Follow these final steps:

  1. Go back to Microsoft Fabric.

  2. In the Choose data source screen, search for odbc and select the Odbc source:

    Choose ODBC as the data source in Microsoft Fabric Copy job
  3. Then enter your ODBC connection string (DSN=GatewayDSN) and select MyGateway from the Data gateway dropdown we configured in the previous step:

    DSN=GatewayDSN
    DSN=GatewayDSN
    Configure ODBC connection string in Microsoft Fabric Copy job
  4. Read the data the same way we discussed at the beginning of this article.

  5. That's it!

Now you can connect to Amazon S3 CSV File data in Microsoft Fabric via the Data Gateway.

If you are asked for authentication details, use Database authentication or SQL Authentication option and enter credentials you used when configuring Data Gateway, e.g. john and your password.

Conclusion

In this article we showed you how to connect to Amazon S3 CSV File in Microsoft Fabric and integrate data without any coding, saving you time and effort.

We encourage you to download Amazon S3 CSV File Connector for Microsoft Fabric and see how easy it is to use it for yourself or your team.

If you have any questions, feel free to contact ZappySys support team. You can also open a live chat immediately by clicking on the chat icon below.

Download Amazon S3 CSV File Connector for Microsoft Fabric Documentation

More integrations

Other connectors for Microsoft Fabric

All
Big Data & NoSQL
Database
CRM & ERP
Marketing
Collaboration
Cloud Storage
Reporting
Commerce
API & Files

Other application integration scenarios for Amazon S3 CSV File

All
Data Integration
Database
BI & Reporting
Productivity
Programming Languages
Automation & Scripting
ODBC applications