Azure Data Factory

Applies to: Dataedo 23.x (current) versions, Article available also for: 10.x

Azure Data Factory (ADF) is a cloud ETL tool developed by Microsoft. It allows you to build complex ETL pipelines using drag&drop web UI.

This article describes abilities and limitations of Azure Data Factory Dataedo connector and walk you through the import process.

Supported elements and metadata

Objects imported

  • Pipelines
    • Name
    • Folder (as schema)
    • Activities
      • Name
      • Data lineage (object and column level, details below)
  • Datasets
    • Name
    • Folder (as schema)
    • Columns (if linked service is imported to Dataedo)

Dataedo imports all the activities as data processes within pipelines. Nested activities like ForEach, Until, IfCondition, and Switch are also imported with all the activities inside them.

Activites we build automatic data-lineage for

  • Copy activity - Object-level and column-level lineage
  • Dataflow - Object-level lineage

Automatic data lineage

Dataedo ADF connector always creates automatic data lineage dataset -> task -> dataset. We also support source -> dataset -> task -> dataset -> sink for following linked service types:

  • Microsoft SQL Server
  • Azure SQL Server
  • Azure Synapse Analytics
  • Azure Blob Storage
  • Azure Data Lake Storage
  • Amazon S3,
  • Postgres
  • Redshift
  • MariaDB
  • MySQL
  • MongoDB
  • Snowflake
  • DB2
  • Postgres

Connecting to Azure Data Factory

Permissions

To list resource groups in a connection window you must belong to the Data Factory Contributor role at the Resource Group level or above.

To run import you need just read access to the factory you want to document (Microsoft.DataFactory/factories/read). However in this case you must provide the resource group and factory name by hand instead of picking from the list.

Add new connection

To connect to Azure Data Factory and create new documentation click Add documentation and choose Database connection.

Image title

On the connection screen choose Azure Data Factory as DBMS.

Image title

Connection details

  • Subscription - Azure subscription assigned to your factory.
  • Resource group - Resource group where your factory sits in.
  • Data factory name - Name of the factory you want to extract metadata from.

Image title

Go to portal.azure.com . Search for your data factory name and open this resource to get information needed.

Image title

Importing objects

When the connection was successful Dataedo will read objects and show a list of objects found. You can choose which objects to import. You can also use an advanced filter to narrow down the list of objects.

Image title

Confirm list of objects to import by clicking Next.

The next screen allows you to change the default name of the documentation under your schema will be visible in the Dataedo repository.

Image title

Click Import to start the import.

Image title

When done close import window with Finish button.

Outcome

Your Azure Data Factory objects has been imported to new documentation in the repository.

Image title

Found issue with this article? Comment below
Comments are only visible when the visitor has consented to statistics cookies. To see and add comments please accept statistics cookies.
0
There are no comments. Click here to write the first comment.