Skip to main content

Connecting to Azure Blob Storage

Overview

Azure Blob Storage helps you to store massive amounts of unstructured data in the Blob storage. You can use Skypoint's built-in connector for importing data from Azure Blob Storage.

This document will guide you through the process of connecting Azure Blob Storage to Skypoint's Modern Data Stack Platform.

Prerequisite

You will need the following details to configure and import data using Azure Blob Storage:

  • Storage Account Name
  • Account Key
  • Storage Path.

imageNote

You can refer to the Setting up Azure Blob Storage document to learn more about the Prerequisites.

To import data using Azure Blob Storage connector

Follow the below steps to create a new dataflow for the Azure Blob Storage import connector:

  1. Go to Dataflow > Imports.
  2. Click New dataflow.

The Set dataflow name page appears.


Alt text


  1. In the Set dataflow name page, type dataflow name in the Name text area.
  2. Click Next.

The Choose connector page appears.


Alt text


To add Azure Blob Storage connector

  1. In the Choose connector page, select Azure Blob Storage connector.

imageNote

You can use the Search feature too to find the connector. Also, the Azure Blob Storage connector can be found under Cloud and Data Warehousing category.


Alt text


  1. Enter Display Name for your dataflow in the text area.
  2. Enter Description for your dataflow in the text area.
  3. Click Next.

The Connect to Azure Blob Storage page appears.


Alt text


To configure Azure Blob Storage

Follow the below steps to configure the connection to Azure Blob Storage:

  1. Enter the Storage account name in the text area.
  2. Enter the Account key in the text area.
  3. Click the Folder icon in the Storage path text area.

Once you select the storage path, the Table Details columns appear.


Alt text


  1. Enter the Table Details to process the data.
ItemDescription
PurposeOption to assign a purpose (Data or Metadata) for each table.
Data
Load customer data.
Metadata
Load Metadata.
File NameDisplays the name of the file that you imported.
Table NameDisplays the imported table name.
Datetime formatDisplays a number of Datetime Formats and Skypoint's Modern Data Stack Platform is set to automatically detect them.
DelimiterDisplays available separators for the variables in the imported data.
First Row as HeaderCheck the box for the system to automatically collect the data according to the Header Contents.
Advanced SettingsSelect the options to fine tune the Import process with minute details.
  1. Click the Advanced settings for your desired file name.

The Advanced settings pop-up appears.


Alt text


ItemDescription
Compression typeMethod that is used for compressing the details from source, Azure Blob Storage.
Row delimiterA separator that identifies the boundaries of the flow of a data stream. In case, a different separator is used in it, the information requires change for more accuracy in data ingestion.
EncodingAs the data comes in data stream, there is a type of encoding used for deciphering it. The default encoding is UTF-8.
Escape characterIt is a particular case of metacharacters that is given an identification of start or end. You can manually select it from the drop-down list.
Quote characterYou can select one of the advanced Quote characters mentioned in the drop-down list.
  1. Click Save on the Advanced settings pop-up to save the advanced settings.
  2. Click Save.

Run, edit, and delete the imported data

Once you save the connector, the Azure Blob Storage connector gets displayed in the list of tables created in the Dataflow page.


Alt text


ItemDescription
NameDisplays the name of the imported Dataflow.
TypeDisplays connector type symbol.
StatusIndicates whether the data is imported successfully.
Tables CountDisplays the number of tables.
Created DateDisplays date of creation.
Last refresh typeDisplays the refresh value. You can see the value is Full or Incremental after the last refresh of data.
Updated DateDisplays last modified date.
Last RefreshDisplays the latest refresh date. This date will get updated whenever you refresh the data.
Group byOption to view the items in a specific Group (For example, name, type, status).
  • Select the horizontal ellipsis in the Actions column and do the following:
If you want toThen
Modify the DataflowSelect Edit and modify the Dataflow. Click Save to apply your changes.
Execute the DataflowSelect Run.
Bring the data to its previous stateSelect Rollback.
Delete the DataflowSelect Remove and then click the Delete button. All tables in the data source get deleted.
See the run history of the DataflowSelect Run history.

Next step

After completing the data import, start the Master Data Management (MDM) - Stitch process to develop a unified view of your customers.