Amazon S3 CSV File Connector for Python

Amazon S3 CSV File Connector can be used to read CSV Files stored in AWS S3 Buckets. Using this you can easily integrate AWS S3 CSV File data. It's supports latest security standards, and optimized for large data files. It also supports reading compressed files (e.g. GZip /Zip).

In this article you will learn how to quickly and efficiently integrate Amazon S3 CSV File data in Python. We will use high-performance Amazon S3 CSV File Connector to easily connect to Amazon S3 CSV File and then access the data inside Python.

Let's follow the steps below to see how we can accomplish that!

Download Documentation

Create ODBC Data Source (DSN) based on ZappySys Amazon S3 CSV Driver

Step-by-step instructions

To get data from Amazon S3 CSV File using Python we first need to create a DSN (Data Source) which will access data from Amazon S3 CSV File. We will later be able to read data using Python. Perform these steps:

  1. Download and install ODBC PowerPack.

  2. Open ODBC Data Sources (x64):

    Open ODBC Data Source
  3. Create a User data source (User DSN) based on ZappySys Amazon S3 CSV Driver:

    ZappySys Amazon S3 CSV Driver
    Create new User DSN for ZappySys Amazon S3 CSV Driver
    • Create and use User DSN if the client application is run under a User Account. This is an ideal option in design-time, when developing a solution, e.g. in Visual Studio 2019. Use it for both type of applications - 64-bit and 32-bit.
    • Create and use System DSN if the client application is launched under a System Account, e.g. as a Windows Service. Usually, this is an ideal option to use in a production environment. Use ODBC Data Source Administrator (32-bit), instead of 64-bit version, if Windows Service is a 32-bit application.
  4. Create and configure a connection for the Amazon S3 storage account.

    Create Amazon S3 Storage Connection
  5. You can use select your desired single file by clicking [...] path button.

    mybucket/dbo.tblNames.csv
    dbo.tblNames.csv
    Read Amazon S3 CSV File data


    ----------OR----------

    You can also read the multiple files stored in Amazon S3 Storage using wildcard pattern supported e.g. dbo.tblNames*.csv.

    Note: If you want to operation with multiple files then use wild card pattern as below 
    (when you use wild card pattern in source path then system will treat target path as folder regardless you end with slash)
    
    mybucket/dbo.tblNames.csv (will read only single .CSV file)
    mybucket/dbo.tbl*.csv (all files starting with file name)
    mybucket/*.csv (all files with .csv Extension and located under folder subfolder)
    

    mybucket/dbo.tblNames*.csv
    Use wildcard pattern .* to read multiple Amazon S3 Files data


    ----------OR----------

    You can also read the zip and gzip compressed files also without extracting it in using Amazon S3 CSV Source File Task.

    mybucket/dbo.tblNames*.gz
    Reading zip and gzip compressed files (stream mode)
  6. Navigate to the Preview Tab and let's explore the different modes available to access the data.

    1. --- Using Direct Query ---

      Click on Preview Tab, Select Table from Tables Dropdown and select [value] and click Preview.
      ZappySys ODBC Driver - Preview Data
    2. --- Using Stored Procedure ---

      Note : For this you have to Save ODBC Driver configuration and then again reopen to configure same driver. For more information click here.
      
      Click on the Custom Objects Tab, Click on Add button and select Add Procedure and Enter an appropriate name and Click on OK button to create.
      ZappySys ODBC Driver - Custom Objects
      1. --- Without Parameters ---

        Now Stored Procedure can be created with or without parameters (see example below). If you use parameters then Set default value otherwise it may fail to compilation)
        ZappySys ODBC Driver : Without Parameters
      2. --- With Parameters ---

        Note : Here you can use Placeholder with Paramters in Stored Procedure.
        Example : SELECT * FROM $ WHERE OrderID = '<@OrderID, FUN_TRIM>' or CustId = '<@CustId>' and Total >= '<@Total>'
        
        ZappySys ODBC Driver : With Parameters
    3. --- Using Virtual Table ---

      Note : For this you have to Save ODBC Driver configuration and then again reopen to configure same driver. For more information click here.
      

      ZappySys APi Drivers support flexible Query language so you can override Default Properties you configured on Data Source such as URL, Body. This way you don't have to create multiple Data Sources if you like to read data from multiple EndPoints. However not every application support supplying custom SQL to driver so you can only select Table from list returned from driver.

      Many applications like MS Access, Informatica Designer wont give you option to specify custom SQL when you import Objects. In such case Virtual Table is very useful. You can create many Virtual Tables on the same Data Source (e.g. If you have 50 Buckets with slight variations you can create virtual tables with just URL as Parameter setting).

      vt__Customers
      DataPath=mybucket_1/customers.csv
      
      vt__Orders
      DataPath=mybucket_2/orders.csv
      
      vt__Products
      DataPath=mybucket_3/products.csv
      
      1. Click on the Custom Objects Tab, Click on Add button and select Add Table and Enter an appropriate name and Click on OK button to create.
        ZappySys ODBC Driver - Custom Objects
      2. Once you see Query Builder Window on screen Configure it.
        ZappySys ODBC Driver - Custom Objects : Virtual Table Query Builder
      3. Click on Preview Tab, Select Virtual Table(prefix with vt__) from Tables Dropdown or write SQL query with Virtual Table name and click Preview.
        ZappySys ODBC Driver - Custom Objects : Virtual Table Query Execute

  7. Click OK to finish creating the data source

  8. That's it; we are done. In a few clicks we configured the to Read the Amazon S3 CSV File data using ZappySys Amazon S3 CSV File Connector

Read data in Python

Using ODBC DSN

  1. Use this code snippet to read the data using AmazonS3CsvFileDSN data source:

    AmazonS3CsvFileDSN')
    Python code to get the data from ZappySys DSN
  2. When you run the code it will make the API call and read the data:

    Python - Extracted data from ZappySys DSN
  3. Here is Python program's code in text format:

    
        import pyodbc
        conn = pyodbc.connect('DSN=AmazonS3CsvFileDSN')
    
        cursor = conn.cursor()
    
        #execute query to fetch data from API service
        cursor.execute("SELECT id,title FROM products")
    
        row = cursor.fetchone()
        while row:
            print(row)
            row = cursor.fetchone()
    
        ##For loop example
        #for row in cursor:
        #    print(row)
    

Using a full ODBC connection string

If you want to avoid being dependent on a DSN and creating multiple DSNs for each platform (x86, x64), then you can use a fully qualified connection string. Simply go to your DSN and copy the Connection String:

  1. Open ODBC data source configuration and click Copy settings:
    ZappySys Amazon S3 CSV Driver - Configuration [Version: 2.0.1.10418]
    ZappySys Amazon S3 CSV Driver - Amazon S3 CSV File
    Amazon S3 CSV File Connector can be used to read CSV Files stored in AWS S3 Buckets. Using this you can easily integrate AWS S3 CSV File data. It's supports latest security standards, and optimized for large data files. It also supports reading compressed files (e.g. GZip /Zip).
    AmazonS3CsvFileDSN
    Copy connection string for ODBC application
  2. The window opens, telling us the connection string was successfully copied to the clipboard: Successful connection string copying for ODBC application
  3. Then in your Python code use Connection String when initializing OdbcConnection object, for example:

    conn = pyodbc.connect('DRIVER={ZappySys Amazon S3 CSV Driver};ServiceUrl=https://yourservices.provider.com/api/xxxx....;AuthName=Http;')

How to install `pyodbc` in the Python?

You would need to install pyodbc in Python if you intend to establish connections to databases that support ODBC (Open Database Connectivity). This module facilitates communication between Python applications and various database management systems, enabling you to perform operations such as querying, retrieving data, and managing databases. Here's how you can install pyodbc in Python:

Installation Steps:

  1. Ensure you have Python installed on your system. If not, download it from the official Python website and follow the installation instructions.

  2. Open your terminal or command prompt.

  3. Use the following command to install pyodbc using pip, the Python package installer:

    python -m pip install "pyodbc"

    Make sure you have a stable internet connection and the necessary permissions to install Python packages.


    Python - pip install pyodbc

Reasons to Install:

  • If pyodbc is not installed, your Python script will generate the following error:
    "ModuleNotFoundError: No module named 'pyodbc'"
    .
  • Database Connectivity: pyodbc allows Python to connect to various databases that support ODBC, such as Microsoft SQL Server, PostgreSQL, MySQL, and more.

  • Data Operations: It facilitates the execution of SQL queries, retrieval of data, and other database operations from within Python scripts.

  • Cross-Platform Support: pyodbc is designed to work across different operating systems, including Windows, macOS, and various Linux distributions.

  • Simplicity and Efficiency: The module provides an intuitive interface for managing database transactions and connections, simplifying the process of working with databases in Python.

By installing pyodbc, you can seamlessly integrate your Python applications with a wide range of ODBC-supported databases, enabling efficient and effective data management and analysis.

Centralized data access via Data Gateway

In some situations, you may need to provide Amazon S3 CSV File data access to multiple users or services. Configuring the data source on a Data Gateway creates a single, centralized connection point for this purpose.

This configuration provides two primary advantages:

  • Centralized data access
    The data source is configured once on the gateway, eliminating the need to set it up individually on each user's machine or application. This significantly simplifies the management process.
  • Centralized access control
    Since all connections route through the gateway, access can be governed or revoked from a single location for all users.
Data Gateway
Local ODBC
data source
Simple configuration
Installation Single machine Per machine
Connectivity Local and remote Local only
Connections limit Limited by License Unlimited
Central data access
Central access control
More flexible cost

If you need any of these requirements, you will have to create a data source in Data Gateway to connect to Amazon S3 CSV File, and to create an ODBC data source to connect to Data Gateway in Python.

Let's not wait and get going!

Creating Amazon S3 CSV File data source in Gateway

In this section we will create a data source for Amazon S3 CSV File in Data Gateway. Let's follow these steps to accomplish that:

  1. Search for gateway in Windows Start Menu and open ZappySys Data Gateway Configuration:

    Opening Data Gateway
  2. Go to Users tab and follow these steps to add a Data Gateway user:

    • Click Add button
    • In Login field enter username, e.g., john
    • Then enter a Password
    • Check Is Administrator checkbox
    • Click OK to save
    Data Gateway - Adding User
  3. Now we are ready to add a data source:

    • Click Add button
    • Give Datasource a name (have it handy for later)
    • Then select Native - ZappySys Amazon S3 CSV Driver
    • Finally, click OK
    AmazonS3CsvFileDSN
    ZappySys Amazon S3 CSV Driver
    Data Gateway - Adding data source
  4. When the ZappySys Amazon S3 CSV Driver configuration window opens, configure the Data Source the same way you configured it in ODBC Data Sources (64-bit), in the beginning of this article.

  5. Very important step. Now, after creating or modifying the data source make sure you:

    • Click the Save button to persist your changes.
    • Hit Yes, once asked if you want to restart the Data Gateway service.

    This will ensure all changes are properly applied:

    ZappySys Data Gateway - Save Changes
    Skipping this step may result in the new settings not taking effect and, therefore you will not be able to connect to the data source.

Creating ODBC data source for Data Gateway

In this part we will create ODBC data source to connect to Data Gateway from Python. To achieve that, let's perform these steps:

  1. Open ODBC Data Sources (x64):

    Open ODBC Data Source
  2. Create a User data source (User DSN) based on ODBC Driver 17 for SQL Server:

    ODBC Driver 17 for SQL Server
    Create new User DSN for ODBC Driver 17 for SQL Server
    If you don't see ODBC Driver 17 for SQL Server driver in the list, choose a similar version driver.
  3. Then set a Name of the data source (e.g. Gateway) and the address of the Data Gateway:

    GatewayDSN
    localhost,5000
    ODBC driver for SQL Server - Setting hostname and port
    Make sure you separate the hostname and port with a comma, e.g. localhost,5000.
  4. Proceed with authentication part:

    • Select SQL Server authentication
    • In Login ID field enter the user name you used in Data Gateway, e.g., john
    • Set Password to the one you configured in Data Gateway
    ODBC driver for SQL Server - Selecting SQL Authentication
  5. Then set the default database property to AmazonS3CsvFileDSN (the one we used in Data Gateway):

    AmazonS3CsvFileDSN
    ODBC driver for SQL Server - Selecting database
  6. Continue by checking Trust server certificate option:

    ODBC driver for SQL Server - Trusting certificate
  7. Once you do that, test the connection:

    ODBC driver for SQL Server - Testing connection
  8. If connection is successful, everything is good:

    ODBC driver for SQL Server - Testing connection succeeded
  9. Done!

We are ready to move to the final step. Let's do it!

Accessing data in Python via Data Gateway

Finally, we are ready to read data from Amazon S3 CSV File in Python via Data Gateway. Follow these final steps:

  1. Go back to Python.

  2. Use this code snippet to read the data using GatewayDSN data source:

    GatewayDSN')
    Python code to get the data from ZappySys DSN
  3. Read the data the same way we discussed at the beginning of this article.

  4. That's it!

Now you can connect to Amazon S3 CSV File data in Python via the Data Gateway.

If you are asked for authentication details, use Database authentication or SQL Authentication option and enter credentials you used when configuring Data Gateway, e.g. john and your password.

Conclusion

In this article we showed you how to connect to Amazon S3 CSV File in Python and integrate data without any coding, saving you time and effort.

We encourage you to download Amazon S3 CSV File Connector for Python and see how easy it is to use it for yourself or your team.

If you have any questions, feel free to contact ZappySys support team. You can also open a live chat immediately by clicking on the chat icon below.

Download Amazon S3 CSV File Connector for Python Documentation

More integrations

Other connectors for Python

All
Big Data & NoSQL
Database
CRM & ERP
Marketing
Collaboration
Cloud Storage
Reporting
Commerce
API & Files

Other application integration scenarios for Amazon S3 CSV File

All
Data Integration
Database
BI & Reporting
Productivity
Programming Languages
Automation & Scripting
ODBC applications