Skip to the content
LeaplogicLeaplogic
  • Home
  • About Us
  • Contact
SIGN IN
  • Home
  • About Us
  • Contact

  • Getting Started
    • Before You Begin
    • Creating an Account
    • Logging into LeapLogic
    • Reset Password
    • Quick Tour of the Web Interface
    • LeapLogic in 15 minutes
      • Prerequisites
      • Step 1. Log into LeapLogic
      • Step 2. Create Assessment and Get Insights
      • Step 3. Create Transformation Pipeline and See Results
      • Step 4. Edit or Optimize the Transformed Code
      • Step 5: Complete the Transformation Lifecycle
  • Introduction to LeapLogic
    • Overview
    • High Level Architecture
    • Supported Legacy and Cloud Platforms
    • Key Features
  • Workload Assessment
    • Overview
    • Value Proposition
    • Creating Assessment
      • Prerequisites
      • Step 1. Provide Primary Inputs
        • Automation Coverage
      • Step 2. Add the Additional Inputs
        • Table Stat Extraction Steps
          • Teradata
          • Oracle
          • Netezza
      • Step 3. Update the Source Configuration
      • Step 4. Configure the Recommendation Settings
    • Assessment Listing
    • Understanding Insights and Recommendations
      • Volumetric Info
      • EDW
        • Oracle
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Vertica
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Snowflake
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Azure Synapse
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • SQL Server
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Teradata
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Netezza
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Google Big Query
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Redshift
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • PostgreSQL
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Duck DB
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • ClickHouse
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Exasol
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • DB2
          • Highlights
          • Analysis
          • Optimization
          • Recommendations
          • Lineage
          • Downloadable Reports
      • ETL
        • Informatica
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Ab Initio
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • DataStage
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Talend
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • SSIS
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Informatica BDM
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Oracle Data Integrator
          • Highlights
          • Analysis
          • Downloadable Reports
        • Pentaho
          • Highlights
          • Analysis
          • Downloadable Reports
        • Azure Data Factory
          • ARM Template
          • Highlights
          • Analysis
          • Downloadable Reports
        • Matillion
          • Highlights
          • Analysis
          • Downloadable Reports
        • SnapLogic
          • Highlights
          • Analysis
          • Downloadable Reports
      • Orchestration
        • AutoSys
          • Highlights
          • Analysis
          • Downloadable Reports
        • Control-M
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • SQL Server
          • Highlights
          • Analysis
      • BI
        • OBIEE
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Tableau
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • IBM Cognos
          • Highlights
          • Analysis
          • Downloadable Reports
        • MicroStrategy
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Power BI
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • SSRS
          • Highlights
          • Analysis
          • Downloadable Reports
        • SAP BO
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • WebFOCUS
          • Highlights
          • Analysis
          • Downloadable Reports
      • Analytics
        • SAS
          • Highlight
          • Analysis
          • Lineage
          • Downloadable Reports
        • Alteryx
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
      • Integrated Assessment (EDW, ETL, Orchestration, BI)
        • Highlights
        • Analysis
        • Optimization
        • Lineage
        • Recommendations
    • Managing Assessment Reports
      • Downloading Report
      • Input Report Utility
      • View Configuration
    • Complexity Calculation Logic
    • Key Benefits
    • Ad hoc Query
  • Metadata Management
    • Overview
    • Introduction to Data Catalog
      • Managing Data Catalog
        • Building Data Catalog
        • Insights to Data Catalog
        • Managing the Repository and Data Source
      • Creating Repository (Repo)
      • Creating Data Source
    • Tag Management
    • Key benefits
  • Batch Processing using Pipeline
    • Introduction
    • Designing Pipeline
      • How to create a pipeline
        • Configuring Migration Stage
          • Schema Optimization
        • Configuring Transformation Stage
          • On-premises to Cloud
          • Cloud-to-Cloud
          • LeapLogic Express
        • Configuring Validation Stage
          • Data Validation
            • Table
            • File
            • File and Table
            • Cell-by-cell validation
          • Query Validation
            • Query Validation (When Data is Available)
            • Query Validation (When Data is Not Available)
          • Schema Validation
        • Configuring Execution Stage
        • Configuring ETL Conversion Stage
          • Ab Initio
          • Informatica
          • Informatica BDM
          • Matillion
          • DataStage
          • SSIS
          • IICS
          • Talend
          • Oracle Data Integrator
          • Pentaho
          • SnapLogic
        • Configuring Mainframe Conversion Stage
          • Cobol
          • JCL
        • Configuring Orchestration Stage
          • AutoSys
          • Control-M
        • Configuring BI Conversion Stage
          • OBIEE to Power BI
          • OBIEE to AWS QuickSight
          • Tableau to Amazon QuickSight
          • Tableau to Power BI
          • Tableau to Superset
          • Tableau to Looker
          • IBM Cognos to Power BI
        • Configuring Analytics Conversion Stage
          • SAS
          • Alteryx
        • Configuring Script Conversion Stage
    • Key Features
      • How to schedule a pipeline
      • Configuring Parameters
  • Pipeline Reports
    • Overview of Pipeline Report
    • Pipeline Listing
    • Reports and Insights
      • Migration
      • Transformation
        • On-premises to Cloud
        • Cloud-to-Cloud
        • LeapLogic Express
      • Validation
        • Data
          • File
          • Table
          • File and Table
        • Query
          • Query Validation Report (When Data is Available)
          • Query Validation Report (When Data is not Available)
        • Schema
      • Execution
      • ETL
        • Ab Initio
        • Informatica
        • Informatica BDM
        • Matillion
        • DataStage
        • SSIS
        • IICS
        • Talend
        • Oracle Data Integrator
        • Pentaho
        • SnapLogic
      • Mainframe
        • Cobol
        • JCL
      • Orchestration
        • AutoSys
        • Control-M
      • BI
        • OBIEE to Power BI
        • OBIEE to Amazon QuickSight
        • Tableau to Amazon QuickSight
        • Tableau to Power BI
        • Tableau to Superset
        • Tableau to Looker
        • IBM Cognos to Power BI
      • Analytics
        • SAS
        • Alteryx
      • Shell Script
      • Common Model
    • Automation Level Indicator
      • ETL
        • Informatica
        • Matillion
        • DataStage
        • Informatica BDM
        • SnapLogic
        • IICS
        • Ab Initio
        • SSIS
        • Talend
        • Pentaho
      • Orchestration
        • AutoSys
        • Control-M
      • EDW
      • Analytics
        • SAS
        • Alteryx
      • BI
      • Shell Script
    • Error Specifications & Troubleshooting
  • SQL Transformation
    • Overview
    • Creating and Executing the Online Notebook
      • How to Create and Execute the Notebook
      • Supported Features
    • Configuring the Notebook
      • Transformation
      • Unit Level Validation
      • Script Level Validation
    • Notebook Listing
  • Operationalization
    • Overview
      • Basic
      • Advanced
      • Cron Expression
    • Parallel Run Pipeline Listing
  • Transformation Source
    • Introduction
    • Creating Transformation Source Type
  • Governance
    • Summary of Governance - Roles and Permissions
    • User Creation
      • Creating a new User Account
    • Adding Roles and permissions
      • How to add Roles and Permissions to a new user?
    • Adding Group Accounts
    • Default Quota Limits
    • Product Usage Metrics
  • License
    • EDW
    • ETL
  • LeapLogic Desktop Version
    • Overview
    • Registration and Installation
    • Getting Started
    • Creating Assessment
      • ETL
      • DML
      • Procedure
      • Analytics
      • Hadoop
    • Reports and Insights
      • Downloadable Reports
      • Reports for Estimation
    • Logging and Troubleshooting
    • Sample Scripts
    • Desktop vs. Web Version
    • Getting Help
  • LeapLogic (Version 4.8) Deployment
    • System Requirements
    • Prerequisites
    • Deployment
      • Extracting Package
      • Placing License Key
      • Executing Deployment Script
      • Accessing LeapLogic
    • Uploading License
    • Appendix
    • Getting Help
  • Removed Features
    • Configuring File Validation Stage
    • Variable Extractor Stage
      • Variable Extractor Report
    • Configuring Meta Diff Stage
      • Meta Diff
    • Configuring Data Load Stage
      • Data Load
    • Configuring Multi Algo Stage
  • FAQs
  • Tutorial Videos
  • Notice
Home   »  Workload Assessment   »  Understanding Insights and Recommendations   »  ETL  »  Azure Data Factory Assessment Report

Azure Data Factory Assessment Report

This topic contains information about the Azure Data Factory assessment report. The assessment assesses workloads and produces in-depth insights that help plan the migration. The Azure Data Factory assessment supports only ZIP files (ARM template) as input files.

To learn how to export an ARM template from the Azure portal, click Export ARM Template.

In This Topic:

  • Highlights
    • Summary
    • Resources
    • Pipelines
    • Queries
  • Analysis
    • Files
    • Pipelines
    • Queries
    • Entities
    • Resources
    • Artifacts
  • Downloadable Reports
    • Insights and Recommendations
    • Source Inventory Analysis

Highlights

The highlights section gives you a high-level overview of your assessment summary of the analytics performed on the selected workloads. It includes information about resource types and pipelines.

Summary

This section summarizes the input source scripts and the associated workload inventory. It includes information about pipelines, activities, resources, procedures, and so on.

  • Files: Displays the total number of input source files.
  • Pipelines: Set of activities to accomplish a task or work.
  • Activities: Each pipeline can contain multiple activities. Activities are primary components that specify actions such as validating data, deleting data, obtaining metadata, etc.
  • Procedures: Displays the number of stored procedures used in the data factory. It is a set of SQL queries to perform an action or task.
  • Resources: Resources are manageable services or entities, for instance, databases, storage accounts, virtual networks, etc.
  • Entities: Displays the number of entities used in the dataset.
  • URLs: Displays the number of URLs. It is a service to establish or invoke communication such as sending email, triggering any website, etc.
  • External Files: External files are external libraries that are used in Azure Data Factory. The files with external file formats such as bash, CMD, etc., are considered external files.

Resources

This section provides an overview of resource types such as datasets, linked services, and triggers.

  • Datasets: It is a collection of data that is used in various activities.
  • Linked Services: Linked services are connection strings that contain connection details such as databases, URLs, file paths, etc., to connect to different services (cloud, legacy data warehouses, etc.).
  • Triggers: Triggers are used to execute a pipeline. In Azure Data Factory, there are three types of triggers:
    • Schedule: To execute the pipeline based on a predefined schedule.
    • Blobevents: Execute the pipeline when a new or updated Azure blob storage is identified.
    • Tumbling window: To execute the pipeline at a periodic interval irrespective of past and future dated data.

Pipelines

This section provides information about the total number of pipelines within the entire inventory along with an assessment of their complexity.

Queries

This section provides an overview of unique, analyzed, and unanalyzed queries along with their complexity.

  • Unique Query: Displays the number of unique queries. In Unique Queries, the duplicate queries are eliminated, and solitary queries are identified.
  • Analyzed: Displays the number of analyzed queries.
  • Not analyzed: Displays the number of queries that are not analyzed.

Analysis

This topic provides a detailed examination of the source files.

Files

This section provides a comprehensive report of the source files along with information about the total number of files, pipelines, activities, datasets, and so on.

  • File Name: Displays the file name. The used naming convention is folder_name/file_name.
  • Pipelines: Displays the number of pipelines existing in the file. Pipeline is a set of activities to accomplish a task or work.
  • Activities: Displays the number of activities. Each pipeline contains multiple activities. Activities are primary components that specify the actions such as validating data, deleting data, obtaining metadata, etc.
  • Datasets: Displays the number of datasets. It is a collection of data which is used in various activities.
  • Procedures: Displays the number of procedures. Procedures are a set of SQL queries to perform an action or task.
  • Queries: Displays the number of queries.
  • Complexity: Displays the complexity of the file.

Pipelines

This section provides detailed information about pipelines, including their activities, complexity, dependency conditions, and relative files.

  • Pipeline Name: Displays the name of the pipeline.
  • Relative File: Displays the relative file path on which the pipeline is available.
  • Activities: Displays the number of activities in each pipeline.
  • Complexity: Displays the pipeline complexity.
  • Dependency Conditions: Provides details about the resources on which the pipeline depends. For instance, the pipeline may be dependent on other pipelines, datasets, link services, etc.

Browse through each pipeline to get more insights into the associated activities.

  • Activity Name: Displays the name of the activity.
  • Type: Displays the type of activities carried out within the pipeline such as copying data, filtering data, get metadata, etc.
  • Activity Dependency: Provides details about the activity upon which the current activity is dependent along with its status (Succeeded/Failed). For example, as in the above image, the activity If Condition1 is dependent on the Get Metadata1 activity. And, If the Get Metadata1 activity succeeds, then If Condition1 will be invoked. However, if Get Metadata1 fails, then If Condition1 will not be executed.
  • Activity State: Displays the state of the activity such as Active or Inactive.
  • Activity Description: Provides a description of each activity.

Queries

This section provides detailed information about queries segregated into unique, analyzed, and not-analyzed queries.

Unique Queries

This section displays a list of all unique queries along with the total number of unique queries. In Unique Queries, duplicate queries are eliminated, and solitary queries are identified.

  • File Name: Displays the file name associated with the query.
  • Pipeline Name: Displays the name of the pipeline associated with the query.
  • Source Type: Displays the type of database where queries will be processed.
  • Query Type: Displays the query type.
  • Complexity: Displays the query complexity.

Analyzed

This section lists all queries that meet the analysis criteria.

  • File Name: Displays the file name associated with the query.
  • Pipeline Name: Displays the name of the pipeline associated with the query.
  • Source Type: Displays the type of database where queries will be processed.
  • Query Type: Displays the query type.
  • Complexity: Displays the query complexity.

Not Analyzed

This section lists all queries that did not meet the analysis criteria.

  • File Name: Displays the file name associated with the query.
  • Pipeline Name: Displays the name of the pipeline associated with the query.
  • Source Type: Displays the type of database where queries will be processed.
  • Query Type: Displays the query type.
  • Complexity: Displays the query complexity.

Entities

This section displays a detailed analysis of the entities. It includes information about the type of entities, databases, and database type.

  • Entity Name: Displays the name of the entity.
  • Type: Displays the type of entity.
  • Database Name: Displays the database name.
  • Database Type: Displays the type of database such as Oracle, Redshift, etc., where the entity is present.

Resources

This section provides detailed information about various resource types such as datasets, linked services, and triggers. Resources are manageable services or entities, for instance, databases, storage accounts, virtual networks, etc.

Datasets

This section lists all the datasets. Datasets are a collection of data that is used in various activities.

  • Dataset Name: Displays the name of the dataset.
  • Type: Displays the dataset type.
  • Relative File: Displays the relative file path of the dataset.
  • Linked Service Name: Displays the name of the linked service associated with the dataset.
  • Schema: Displays the associated schema name.
  • Table: Displays the associated table name.
  • Dependency: Provides details about the resources on which the dataset depends. For instance, the dataset may be dependent on other datasets, link services, etc.

Linked Service

This section lists all the linked services. Linked services are connection strings that contain connection details such as databases, URLs, file paths, etc., to connect to different services (cloud, legacy data warehouses, etc.).

  • Linked Service Name: Displays the name of the linked service.
  • Type: Displays the type of the linked services.
  • Relative File: Displays the relative file path of the linked services.
  • Dependency: Provides details about the resources on which the linked service depends. For instance, the linked service may be dependent on other datasets, link services, etc.

Trigger

This section lists all triggers. Triggers are used to execute a pipeline.

  • Trigger Name: Displays the name of the trigger.
  • Type: Displays the type of trigger. There are three types of triggers:
    • Schedule: Execute the pipeline based on a predefined schedule.
    • BlobEvents: Execute the pipeline when a new or updated Azure blob storage is identified.
    • Tumbling window: Execute the pipeline at a periodic interval irrespective of past and future dated data.
  • Relative File: Displays the relative file path of the trigger.
  • Pipeline: Displays the pipeline associated with each trigger.
  • Runtime State: Specifies the runtime state of each trigger, such as started or stopped.
  • Frequency: Displays the frequency such as hour, minute, etc., at which the trigger is scheduled to execute.
  • Interval: Provides the time interval at which the trigger is scheduled to execute.

Artifacts

This section lists all the missing artifacts such as linked services, pipelines, triggers, etc., and the external files within the entire inventory.

Missing Files

This section offers a comprehensive view of all the missing artifacts, categorized into pipelines and resources.

All

This section lists all the missing artifacts. The missing artifacts are identified based on the dependency condition. For instance, if a pipeline depends on dataset A and that dataset A is not found in the input JSON file, it will be considered as a missing artifact.

  • Artifact Name: Displays the name of the missing artifact.
  • Type: Displays the type of missing artifact such as pipelines, resources, triggers, etc.
  • Relative File: Displays the relative file where the missing artifact is written.
  • Source Name: Displays the name of the dependent artifact of the missing artifact.
  • Source Type: Specifies the type of the dependent artifact of the missing artifact.

Pipelines

This section lists all the missing pipelines.

  • Artifact Name: Displays the name of the missing artifact.
  • Relative File: Displays the relative file path of the pipeline.
  • Source Name: Displays the name of the dependent artifact of the missing pipeline.
  • Source Type: Specifies the type of dependent artifact of the missing pipeline.

Resources

This section lists all the missing resources.

  • Artifact Name: Displays the name of the missing resource.
  • Resource Type: Specifies the type of resource.
  • Relative File: Displays the relative file path of the resource.
  • Source Name: Displays the name of the dependent artifact of the missing resource.
  • Source Type: Specifies the type of dependent artifact of the missing resource.

External Files

Lists all the external files including information about the associated pipelines, type of the external files, relative files, and more.

  • Artifact Name: Displays the name of the external file.
  • Type: Specifies the type of external file.
  • Relative File: Displays the relative file path of the source input file.
  • Artifact File Path: Displays the path of the artifacts accessed from external resources, such as DBFS. For example, if the external file is stored in DBFS, the artifact file path represents the DBFS path. If the file is located in S3, then the S3 path shows as the artifact file path.
  • Pipeline Name: Displays the associated pipeline name.


Downloadable Reports

Downloadable reports allow you to export detailed ADF assessment reports of your source data which enables you to gain in-depth insights with ease. To access these assessment reports, click Reports.

Types of Reports

In the Reports section, you can see various types of reports such as Insights and Recommendations, and Source Inventory Analysis. Each report type offers detailed information allowing you to explore your assessment results.

Insights and Recommendations

This report provides an in-depth insight into the source input files.

ADF Assessment Report.xlsx: This report provides insights about the source inventory. It includes information about the source inventory and pipelines.

This report contains the following information:

  • Report Summary: Provides information about all the generated artifacts.
  • Volumetric Info: Presents a summary of the aggregated inventory after analyzing the source files. For instance, it provides volumetric information about stored procedures, datasets, pipelines, activities, and more.
  • Pipeline Summary: Lists all the pipelines associated with the input files. It also provides information about components and pipeline-level complexity.

Source Inventory Analysis

It is an intermediate report which helps to debug failures or calculate the final report. It includes all the generated CSV reports such as ADF File Summary.csv, ADF Pipeline Summary.csv, Query Summary.csv, and more.

ADF File Summary.csv: This report provides information about ADF files including the count of pipelines, procedures, resources, and so on.

ADF Pipeline Summary.csv: This report provides information about pipelines including the total number of activities, external files, dependency conditions, and more

Query Detail.csv: This report provides information about queries including the analyzed status, complexity, parsing status, and more. If the analyzed status is TRUE, it indicates that the query is analyzed successfully. Conversely, a FALSE status indicates that the query is not analyzed.

Query Summary.csv: This report provides information about queries including the number of analyzed queries, not analyzed queries, complexity, and so on.


To learn more, contact our support team or write to: info@leaplogic.io

Copyright © 2025 Impetus Technologies Inc. All Rights Reserved

  • Terms of Use
  • Privacy Policy
  • License Agreement
To the top ↑ Up ↑