Skip to the content
LeaplogicLeaplogic
  • Home
  • About Us
  • Contact
SIGN IN
  • Home
  • About Us
  • Contact

  • Getting Started
    • Before You Begin
    • Creating an Account
    • Logging into LeapLogic
    • Reset Password
    • Quick Tour of the Web Interface
    • LeapLogic in 15 minutes
      • Prerequisites
      • Step 1. Log into LeapLogic
      • Step 2. Create Assessment and Get Insights
      • Step 3. Create Transformation Pipeline and See Results
      • Step 4. Edit or Optimize the Transformed Code
      • Step 5: Complete the Transformation Lifecycle
  • Introduction to LeapLogic
    • Overview
    • High Level Architecture
    • Supported Legacy and Cloud Platforms
    • Key Features
  • Workload Assessment
    • Overview
    • Value Proposition
    • Creating Assessment
      • Prerequisites
      • Step 1. Provide Primary Inputs
        • Automation Coverage
      • Step 2. Add the Additional Inputs
        • Table Stat Extraction Steps
          • Teradata
          • Oracle
          • Netezza
      • Step 3. Update the Source Configuration
      • Step 4. Configure the Recommendation Settings
    • Assessment Listing
    • Understanding Insights and Recommendations
      • Volumetric Info
      • EDW
        • Oracle
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Vertica
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Snowflake
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Azure Synapse
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • SQL Server
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Teradata
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Netezza
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Google Big Query
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Redshift
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • PostgreSQL
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Duck DB
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • ClickHouse
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Exasol
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • DB2
          • Highlights
          • Analysis
          • Optimization
          • Recommendations
          • Lineage
          • Downloadable Reports
      • ETL
        • Informatica
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Ab Initio
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • DataStage
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Talend
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • SSIS
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Informatica BDM
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Oracle Data Integrator
          • Highlights
          • Analysis
          • Downloadable Reports
        • Pentaho
          • Highlights
          • Analysis
          • Downloadable Reports
        • Azure Data Factory
          • ARM Template
          • Highlights
          • Analysis
          • Downloadable Reports
        • Matillion
          • Highlights
          • Analysis
          • Downloadable Reports
        • SnapLogic
          • Highlights
          • Analysis
          • Downloadable Reports
      • Orchestration
        • AutoSys
          • Highlights
          • Analysis
          • Downloadable Reports
        • Control-M
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • SQL Server
          • Highlights
          • Analysis
      • BI
        • OBIEE
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Tableau
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • IBM Cognos
          • Highlights
          • Analysis
          • Downloadable Reports
        • MicroStrategy
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Power BI
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • SSRS
          • Highlights
          • Analysis
          • Downloadable Reports
        • SAP BO
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • WebFOCUS
          • Highlights
          • Analysis
          • Downloadable Reports
      • Analytics
        • SAS
          • Highlight
          • Analysis
          • Lineage
          • Downloadable Reports
        • Alteryx
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
      • Integrated Assessment (EDW, ETL, Orchestration, BI)
        • Highlights
        • Analysis
        • Optimization
        • Lineage
        • Recommendations
    • Managing Assessment Reports
      • Downloading Report
      • Input Report Utility
      • View Configuration
    • Complexity Calculation Logic
    • Key Benefits
    • Ad hoc Query
  • Metadata Management
    • Overview
    • Introduction to Data Catalog
      • Managing Data Catalog
        • Building Data Catalog
        • Insights to Data Catalog
        • Managing the Repository and Data Source
      • Creating Repository (Repo)
      • Creating Data Source
    • Tag Management
    • Key benefits
  • Batch Processing using Pipeline
    • Introduction
    • Designing Pipeline
      • How to create a pipeline
        • Configuring Migration Stage
          • Schema Optimization
        • Configuring Transformation Stage
          • On-premises to Cloud
          • Cloud-to-Cloud
          • LeapLogic Express
        • Configuring Validation Stage
          • Data Validation
            • Table
            • File
            • File and Table
            • Cell-by-cell validation
          • Query Validation
            • Query Validation (When Data is Available)
            • Query Validation (When Data is Not Available)
          • Schema Validation
        • Configuring Execution Stage
        • Configuring ETL Conversion Stage
          • Ab Initio
          • Informatica
          • Informatica BDM
          • Matillion
          • DataStage
          • SSIS
          • IICS
          • Talend
          • Oracle Data Integrator
          • Pentaho
          • SnapLogic
        • Configuring Mainframe Conversion Stage
          • Cobol
          • JCL
        • Configuring Orchestration Stage
          • AutoSys
          • Control-M
        • Configuring BI Conversion Stage
          • OBIEE to Power BI
          • OBIEE to AWS QuickSight
          • Tableau to Amazon QuickSight
          • Tableau to Power BI
          • Tableau to Superset
          • Tableau to Looker
          • IBM Cognos to Power BI
        • Configuring Analytics Conversion Stage
          • SAS
          • Alteryx
        • Configuring Script Conversion Stage
    • Key Features
      • How to schedule a pipeline
      • Configuring Parameters
  • Pipeline Reports
    • Overview of Pipeline Report
    • Pipeline Listing
    • Reports and Insights
      • Migration
      • Transformation
        • On-premises to Cloud
        • Cloud-to-Cloud
        • LeapLogic Express
      • Validation
        • Data
          • File
          • Table
          • File and Table
        • Query
          • Query Validation Report (When Data is Available)
          • Query Validation Report (When Data is not Available)
        • Schema
      • Execution
      • ETL
        • Ab Initio
        • Informatica
        • Informatica BDM
        • Matillion
        • DataStage
        • SSIS
        • IICS
        • Talend
        • Oracle Data Integrator
        • Pentaho
        • SnapLogic
      • Mainframe
        • Cobol
        • JCL
      • Orchestration
        • AutoSys
        • Control-M
      • BI
        • OBIEE to Power BI
        • OBIEE to Amazon QuickSight
        • Tableau to Amazon QuickSight
        • Tableau to Power BI
        • Tableau to Superset
        • Tableau to Looker
        • IBM Cognos to Power BI
      • Analytics
        • SAS
        • Alteryx
      • Shell Script
      • Common Model
    • Automation Level Indicator
      • ETL
        • Informatica
        • Matillion
        • DataStage
        • Informatica BDM
        • SnapLogic
        • IICS
        • Ab Initio
        • SSIS
        • Talend
        • Pentaho
      • Orchestration
        • AutoSys
        • Control-M
      • EDW
      • Analytics
        • SAS
        • Alteryx
      • BI
      • Shell Script
    • Error Specifications & Troubleshooting
  • SQL Transformation
    • Overview
    • Creating and Executing the Online Notebook
      • How to Create and Execute the Notebook
      • Supported Features
    • Configuring the Notebook
      • Transformation
      • Unit Level Validation
      • Script Level Validation
    • Notebook Listing
  • Operationalization
    • Overview
      • Basic
      • Advanced
      • Cron Expression
    • Parallel Run Pipeline Listing
  • Transformation Source
    • Introduction
    • Creating Transformation Source Type
  • Governance
    • Summary of Governance - Roles and Permissions
    • User Creation
      • Creating a new User Account
    • Adding Roles and permissions
      • How to add Roles and Permissions to a new user?
    • Adding Group Accounts
    • Default Quota Limits
    • Product Usage Metrics
  • License
    • EDW
    • ETL
  • LeapLogic Desktop Version
    • Overview
    • Registration and Installation
    • Getting Started
    • Creating Assessment
      • ETL
      • DML
      • Procedure
      • Analytics
      • Hadoop
    • Reports and Insights
      • Downloadable Reports
      • Reports for Estimation
    • Logging and Troubleshooting
    • Sample Scripts
    • Desktop vs. Web Version
    • Getting Help
  • LeapLogic (Version 4.8) Deployment
    • System Requirements
    • Prerequisites
    • Deployment
      • Extracting Package
      • Placing License Key
      • Executing Deployment Script
      • Accessing LeapLogic
    • Uploading License
    • Appendix
    • Getting Help
  • Removed Features
    • Configuring File Validation Stage
    • Variable Extractor Stage
      • Variable Extractor Report
    • Configuring Meta Diff Stage
      • Meta Diff
    • Configuring Data Load Stage
      • Data Load
    • Configuring Multi Algo Stage
  • FAQs
  • Tutorial Videos
  • Notice
Home   »  Workload Assessment   »  Understanding Insights and Recommendations   »  ETL  »  Oracle Data Integrator Assessment Report

Oracle Data Integrator Assessment Report

This topic contains information about the ODI assessment report. The assessment assesses workloads and produces in-depth insights that help to plan the migration. The input format for ODI assessment is XML file format.

In This Topic:

  • Highlights
    • Summary
    • Complexity (Across Packages)
    • Queries
    • Statement Types
    • Query Complexity
    • Orchestration
    • Unpackaged Files
  • Analysis
    • Source Analysis
    • Entities
    • Jobs
    • Orchestration
    • Unpackaged Files
  • Lineage
  • Downloadable Reports
    • Insights and Recommendations
    • Source Inventory Analysis
    • Lineage Report

Highlights

The highlights section gives you a high-level overview of your assessment summary of the analytics performed on the selected workloads. It includes a graphical depiction of the complexity of packages, summary of the source input files, orchestrations, unpackaged files, and so on.

Summary

This section illustrates the input ODI scripts that are analyzed throughout the various components. Here, you can see the number of files, packages, unique entities, mappings, and more.

  • Files: Displays the number of source files.
  • Packages: Displays the number of packages existing in the source files.
  • Unique Entities: Displays the total number of unique entities (without duplicate entities) in the source files.
  • Mappings: Displays the total number of mappings. It describes the flow of data between the source and the target.
  • Scenario: These are codes generated to implement ODI objects such as mappings, packages, etc., in a production environment.
  • Procedures: Displays the total number of procedures. It is a set of sequential commands.
  • Variables: Displays the total number of variables.

Complexity (Across Packages)

This section provides a summarized graphical representation of the ODI packages classification based on complexity. This information is required to make various decisions, including migration planning, and estimating budget.

Queries

It displays a synopsis of analyzed queries with information about unique text, unique pattern, as well as parsed and unparsed queries.

  • Unique (Text): Duplicate queries are eliminated, and solitary queries are identified.
  • Unique (Pattern): Identify the distinct query structures but the query structure’s values may change.
  • Parsed Queries: Displays the number of parsed queries.
  • UnParsed Queries: Displays the number of unparsed queries.

Statement Types

This section displays a summary of analyzed queries based on the query distribution sorted by the different query statement types in the input source file. Moreover, it displays a visual representation of the complexity breakdown across various statement types.

Query Complexity

This section provides a summarized graphical representation of queries’ complexity that helps in making different decisions, including budget estimation.

Complexity TypeComplexity RangeDescription
Low0-4Handled directly by the tool, it requires no manual effort.
Medium4-10Most of the queries can be converted directly by the tool, and some require minimum manual effort.
High10+Complexity is high and requires more manual effort to handle.

Orchestration

This section provides a summary of load plans along with the step types.

Unpackaged Files

Here, you can see a summary of step types that are defined outside the package, for example, mappings, procedures, and so on.


Analysis

This topic provides a detailed examination of packages, entities, jobs, orchestration, and unpackaged files.

Source Analysis

This section provides a comprehensive report of the source files with information about packages, mappings, procedures, variables, and more.

  • Package Name: Displays the package name in the file.
  • File Name: Name of the file.
  • Mapping: Displays the number of mappings in each package. It describes the flow of data between the source and the target.
  • Procedure: Displays the number of procedures in each package. It is a set of sequential commands.
  • Scenario: These are codes generated to implement ODI objects such as mappings, packages, etc., in a production environment.
  • Variables: Displays the count of variables in each package.
  • Complexity: Displays the complexity of each package.

Entities

This section displays a detailed analysis of the entities. It includes information about the schemas, frequency of tables used, and the associated source files.

  • Table Name: Name of the table.
  • Schema Name: Name of the schema.
  • Frequency: Displays the frequency of table used.
  • File Name: Displays the associated source file of each table.

Jobs

This section displays detailed information about the job such as its type, package, and more.

  • Job Name: Name of the job.
  • Job Type: Displays the type of job such as Mapping, Scenario, Procedure, Variable, or Others.
    • Mapping: Mapping describes the flow of data between source and target.
    • Procedure: Set of sequential commands.
    • Scenario: Scenarios are codes generated to implement ODI objects such as mappings, packages, etc., in a production environment.
    • Variable: Object that holds only one value.
    • Others: Jobs that don’t belong to Mapping, Procedure, Scenario, or Variable.
  • Package Name: Displays the associated package.
  • File Name: Displays the associated source file.

Orchestration

This section displays detailed information about the load plan such as its step type, status, and more.

  • Load Plan Name: Name of the load plan.
  • Step Type: Displays the step types such as Mapping, Package, Scenario, Procedure, Variable, or Others.
  • Active Flage: Indicates whether the step type is active or not.
  • File Name: Displays the associated source file.

Unpackaged Files

This section displays detailed information about the artifacts defined outside the package that are segregated into mappings and procedures.

  • Artifact Name: Name of the artifact.
  • Step Type: Displays step types such as Mapping or Package.
  • Step Name: Displays the step name.
  • File Name: Displays the associated source file.
  • Complexity: Displays the complexity of each step count.


Lineage

End-to-end data and process lineage identify the complete dependency structure through interactive and drill-down options to the last level.

Typically, even within one line of business, multiple data sources, entry points, ETL tools, and orchestration mechanisms exist. Decoding this complex data web and translating it into a simple visual flow can be extremely challenging during large-scale modernization programs. The visual lineage graph adds tremendous value and helps defining the roadmap to the modern data architecture. It deep dives into all the existing flows, like Autosys jobs, applications, ETL scripts, BTEQ/Shell (KSH) scripts, procedures, input and output tables, and provides integrated insights. These insights help data teams to make strategic decisions with greater accuracy and completeness. Enterprises can proactively leverage integrated analysis to mitigate the risks associated with migration and avoid business disruptions.

LeapLogic’s lineage helps to trace the complete data flow in steps. The visual representation helps to track how the data hopped and transformed along the way, with details of parameter changes, etc. It provides backward and forward lineage showing the hops and reference points of data lineage with a complete audit trail of the data points of interest. They also help temporarily ignore peripheral, unwanted data points like system tables, bridge tables, etc. and establish useful correlations between data points to uncover hidden patterns and results. Users can deep dive into certain flows leveraging next-gen features like contextual search, advanced filters according to workload type, an interactive visual interface, graph depth, and more.

Now, let’s see how you can efficiently manage lineage.

To view the required lineage:

  1. Select either the Process or Data tab.
  2. Enter the keywords in the Search Keywords field.
  1. Click the Search icon to generate the lineage.

Process lineage illustrates the dependencies between two or more processes such as packages, scenarios, procedures, etc., whereas data lineage depicts data flow between two or more data-holding components such as entities, flat files, etc. 

In addition, the filter search icon allows you to include or exclude particular nodes to obtain the required dependency structure. You can also choose the direction of the lineage. By default, the dependency direction is Left to Right Hierarchy. You can also choose Right to Left Hierarchy or Bidirectional dependency directions as required. Moreover, you can also increase the Hierarchy Levels to the nth level.

Lineage facilitates you visualize how your selected nodes are connected and depend on each other. The nodes and their connecting edges (relationships) help you to understand the overall structure and dependencies.

Nodes Edges
Tables Call
File Read
Procedure Execute
View Write
Bridge Table OTHERS
Mapping Contains
Package
Scenarios

Manage Lineage

This feature enables you to view and manage your lineage. You can add, modify, or delete nodes and their relationships to generate an accurate representation of the required dependency structure. There are two ways to update the lineage: either using Complete Lineage report or Lineage Template.

Using Complete Lineage report

Follow the below steps to modify the lineage:

  1. Click the Manage Graph icon.
  1. Click Download Complete Lineage to update, add, or delete the nodes and their relationships in the current lineage.
  1. Once the complete lineage report is downloaded, you can make necessary updates such as updating, deleting or adding the nodes and its relationships.
  2. After making the required changes, upload the updated lineage report in Upload to Modify Lineage.
  3. Click Apply to incorporate the updates into the dependency structure.
  4. Generate the required process or data lineage.

Using Lineage Template

Follow the below steps to add new nodes and their relationships to the current lineage report:

  1. Click the Manage Graph icon.
  1. Click Download Lineage Template.
  1. Once the lineage template is downloaded, you can add new nodes and relationships in the template.
  2. After making the required changes, upload the template in Upload to Modify Lineage.
  3. Click Apply to incorporate the updates into the complete dependency structure.
  4. Generate the required process or data lineage.

Important

To effectively manage lineage, you must adhere to the following rules:

  • Do not modify the column headers or their order in the Complete Lineage report or the Lineage template. The following are the column headers: source_name, source_type, target_name, target_type, relation_type, and database_id.
  • When deleting a row, retain the value in the database_id column and clear all the other column values.
  • When inserting a row, add all the column values except in the database_id column.
  • When updating a row, ensure that all the columns have relevant data.

You can also apply:

FeatureIconUse
FilterUsed to filter the lineage.
ReloadAssists in reloading graphs.
SaveUsed to save the lineage.
DownloadUsed to download the file.
ExpandUsed to enlarge the screen.


Downloadable Reports

Downloadable reports allow you to export detailed assessment reports of your source data which enables you to gain in-depth insights with ease. To access these assessment reports, click Reports.

Types of Reports

In the Reports section, you can see various types of reports such as Insights and Recommendations, Source Inventory Analysis, and Lineage reports. Each report type offers detailed information allowing you to explore your assessment results.

Insights and Recommendations

This report provides an in-depth insight into the source input files. It contains the final output including the details of queries, complexity, and so on.

Here, you can see the odi folder, and Lineage Dependency Report.xlsx.

Lineage Dependency Report.xlsx: This report contains information about views and script level lineages. It includes information about used and impacted tables, views, files, direct dependencies, dependency hierarchy and more.

This report contains the following information:

  • view_report: Provides information about the views.
  • script_report: Provides information about script level lineage.

Detailed ODI Assessments Reports

To access a detailed assessment report, open the ODI folder.

odi_complexity_summary.csv: This report contains information about the complexity of the odi ETL scripts that help in making different decisions, including budget estimation and the effort required for migration.

odi_scenario_highlights.csv: This report provides information about the source files including total number of mappings, scenarios, variables, and more.

odi_sql_highlights.csv: This report provides information about the number of queries segregated by statement types such as delete, merge, insert, and more.

Browse through the output folder to access the ODI Assessment.xlsx, and queries_detail.csv reports.

ODI Assessment.xlsx: This report provides insights about the source inventory. It helps you plan the next frontier of a modern data platform methodically. It includes a report summary, package summary, load plan summary, and a lot more.

This report contains the following information:

  • Report Summary: Provides information about all the generated artifacts.
  • Volumetric Info: Lists an aggregated inventory of every source file. For instance, it provides information about the total number of files, packages, mappings, and so on.
  • Package Summary: Provides information about the packages in the file. It includes statistical information about the components, mappings, procedures, and more along with the complexity of each package.
  • Load Plan Details: Provides a summary of load plan components including information about the scenarios, used, and impacted tables, and so on.
  • Mapping Summary: Lists all the mappings along with information about the number of join components, expression components, complexity, and so on.
  • Procedure Details: Provides information about procedures including used and unused tables, complexity, and more.

queries_detail.csv: This report provides information about queries including the used and impacted tables, analyzed status, complexity, and more. If the analyzed status is TRUE, it indicates that the query is analyzed successfully. Conversely, a FAIL status indicates that the query is not analyzed.

Source Inventory Analysis

It is an intermediate report which helps to debug failures or calculate the final report. It includes all the generated CSV reports such as Entities.csv, Load Plan Details.csv, package summary.csv, and more.

Entities.csv: This report provides information about entities along with associated logical schema, table aliases, and more.

invalid_query.csv: This report lists all the invalid queries.

Load Plan Details.csv: This report provides comprehensive information about load plans including scenarios, scenario steps, execution orders, used and unused tables, and more.

package summary.csv: This report provides comprehensive information about packages along with information about components, mappings, procedures, complexity, and more.

Lineage Report

This report provides complete dependency details for all nodes. It provides an end-to-end data and process lineage that helps to identify the complete dependency structure and the data flow.

This report contains the following information:

  • Dependency (Process): Provides information about the process lineage.
  • Dependency (Data): Provides information about the data lineage.
  • Nodes: Lists all the source and target nodes along with its type.
  • Volumetric Info (Summary): Provides volumetric information about the artifact types such as input tables, output tables, and schedulers.


To learn more, contact our support team or write to: info@leaplogic.io

Copyright © 2025 Impetus Technologies Inc. All Rights Reserved

  • Terms of Use
  • Privacy Policy
  • License Agreement
To the top ↑ Up ↑