Skip to the content
LeaplogicLeaplogic
  • Home
  • About Us
  • Contact
SIGN IN
  • Home
  • About Us
  • Contact

  • Getting Started
    • Before You Begin
    • Creating an Account
    • Logging into LeapLogic
    • Reset Password
    • Quick Tour of the Web Interface
    • LeapLogic in 15 minutes
      • Prerequisites
      • Step 1. Log into LeapLogic
      • Step 2. Create Assessment and Get Insights
      • Step 3. Create Transformation Pipeline and See Results
      • Step 4. Edit or Optimize the Transformed Code
      • Step 5: Complete the Transformation Lifecycle
  • Introduction to LeapLogic
    • Overview
    • High Level Architecture
    • Supported Legacy and Cloud Platforms
    • Key Features
  • Workload Assessment
    • Overview
    • Value Proposition
    • Creating Assessment
      • Prerequisites
      • Step 1. Provide Primary Inputs
        • Automation Coverage
      • Step 2. Add the Additional Inputs
        • Table Stat Extraction Steps
          • Teradata
          • Oracle
          • Netezza
      • Step 3. Update the Source Configuration
      • Step 4. Configure the Recommendation Settings
    • Assessment Listing
    • Understanding Insights and Recommendations
      • Volumetric Info
      • EDW
        • Oracle
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Vertica
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Snowflake
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Azure Synapse
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • SQL Server
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Teradata
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Netezza
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Google Big Query
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Redshift
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • PostgreSQL
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Duck DB
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • ClickHouse
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • Exasol
          • Highlights
          • Analysis
          • Optimization
          • Lineage
          • Recommendations
          • Downloadable Reports
        • DB2
          • Highlights
          • Analysis
          • Optimization
          • Recommendations
          • Lineage
          • Downloadable Reports
      • ETL
        • Informatica
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Ab Initio
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • DataStage
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Talend
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • SSIS
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Informatica BDM
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Oracle Data Integrator
          • Highlights
          • Analysis
          • Downloadable Reports
        • Pentaho
          • Highlights
          • Analysis
          • Downloadable Reports
        • Azure Data Factory
          • ARM Template
          • Highlights
          • Analysis
          • Downloadable Reports
        • Matillion
          • Highlights
          • Analysis
          • Downloadable Reports
        • SnapLogic
          • Highlights
          • Analysis
          • Downloadable Reports
      • Orchestration
        • AutoSys
          • Highlights
          • Analysis
          • Downloadable Reports
        • Control-M
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • SQL Server
          • Highlights
          • Analysis
      • BI
        • OBIEE
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Tableau
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • IBM Cognos
          • Highlights
          • Analysis
          • Downloadable Reports
        • MicroStrategy
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • Power BI
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • SSRS
          • Highlights
          • Analysis
          • Downloadable Reports
        • SAP BO
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
        • WebFOCUS
          • Highlights
          • Analysis
          • Downloadable Reports
      • Analytics
        • SAS
          • Highlight
          • Analysis
          • Lineage
          • Downloadable Reports
        • Alteryx
          • Highlights
          • Analysis
          • Lineage
          • Downloadable Reports
      • Integrated Assessment (EDW, ETL, Orchestration, BI)
        • Highlights
        • Analysis
        • Optimization
        • Lineage
        • Recommendations
    • Managing Assessment Reports
      • Downloading Report
      • Input Report Utility
      • View Configuration
    • Complexity Calculation Logic
    • Key Benefits
    • Ad hoc Query
  • Metadata Management
    • Overview
    • Introduction to Data Catalog
      • Managing Data Catalog
        • Building Data Catalog
        • Insights to Data Catalog
        • Managing the Repository and Data Source
      • Creating Repository (Repo)
      • Creating Data Source
    • Tag Management
    • Key benefits
  • Batch Processing using Pipeline
    • Introduction
    • Designing Pipeline
      • How to create a pipeline
        • Configuring Migration Stage
          • Schema Optimization
        • Configuring Transformation Stage
          • On-premises to Cloud
          • Cloud-to-Cloud
          • LeapLogic Express
        • Configuring Validation Stage
          • Data Validation
            • Table
            • File
            • File and Table
            • Cell-by-cell validation
          • Query Validation
            • Query Validation (When Data is Available)
            • Query Validation (When Data is Not Available)
          • Schema Validation
        • Configuring Execution Stage
        • Configuring ETL Conversion Stage
          • Ab Initio
          • Informatica
          • Informatica BDM
          • Matillion
          • DataStage
          • SSIS
          • IICS
          • Talend
          • Oracle Data Integrator
          • Pentaho
          • SnapLogic
        • Configuring Mainframe Conversion Stage
          • Cobol
          • JCL
        • Configuring Orchestration Stage
          • AutoSys
          • Control-M
        • Configuring BI Conversion Stage
          • OBIEE to Power BI
          • OBIEE to AWS QuickSight
          • Tableau to Amazon QuickSight
          • Tableau to Power BI
          • Tableau to Superset
          • Tableau to Looker
          • IBM Cognos to Power BI
        • Configuring Analytics Conversion Stage
          • SAS
          • Alteryx
        • Configuring Script Conversion Stage
    • Key Features
      • How to schedule a pipeline
      • Configuring Parameters
  • Pipeline Reports
    • Overview of Pipeline Report
    • Pipeline Listing
    • Reports and Insights
      • Migration
      • Transformation
        • On-premises to Cloud
        • Cloud-to-Cloud
        • LeapLogic Express
      • Validation
        • Data
          • File
          • Table
          • File and Table
        • Query
          • Query Validation Report (When Data is Available)
          • Query Validation Report (When Data is not Available)
        • Schema
      • Execution
      • ETL
        • Ab Initio
        • Informatica
        • Informatica BDM
        • Matillion
        • DataStage
        • SSIS
        • IICS
        • Talend
        • Oracle Data Integrator
        • Pentaho
        • SnapLogic
      • Mainframe
        • Cobol
        • JCL
      • Orchestration
        • AutoSys
        • Control-M
      • BI
        • OBIEE to Power BI
        • OBIEE to Amazon QuickSight
        • Tableau to Amazon QuickSight
        • Tableau to Power BI
        • Tableau to Superset
        • Tableau to Looker
        • IBM Cognos to Power BI
      • Analytics
        • SAS
        • Alteryx
      • Shell Script
      • Common Model
    • Automation Level Indicator
      • ETL
        • Informatica
        • Matillion
        • DataStage
        • Informatica BDM
        • SnapLogic
        • IICS
        • Ab Initio
        • SSIS
        • Talend
        • Pentaho
      • Orchestration
        • AutoSys
        • Control-M
      • EDW
      • Analytics
        • SAS
        • Alteryx
      • BI
      • Shell Script
    • Error Specifications & Troubleshooting
  • SQL Transformation
    • Overview
    • Creating and Executing the Online Notebook
      • How to Create and Execute the Notebook
      • Supported Features
    • Configuring the Notebook
      • Transformation
      • Unit Level Validation
      • Script Level Validation
    • Notebook Listing
  • Operationalization
    • Overview
      • Basic
      • Advanced
      • Cron Expression
    • Parallel Run Pipeline Listing
  • Transformation Source
    • Introduction
    • Creating Transformation Source Type
  • Governance
    • Summary of Governance - Roles and Permissions
    • User Creation
      • Creating a new User Account
    • Adding Roles and permissions
      • How to add Roles and Permissions to a new user?
    • Adding Group Accounts
    • Default Quota Limits
    • Product Usage Metrics
  • License
    • EDW
    • ETL
  • LeapLogic Desktop Version
    • Overview
    • Registration and Installation
    • Getting Started
    • Creating Assessment
      • ETL
      • DML
      • Procedure
      • Analytics
      • Hadoop
    • Reports and Insights
      • Downloadable Reports
      • Reports for Estimation
    • Logging and Troubleshooting
    • Sample Scripts
    • Desktop vs. Web Version
    • Getting Help
  • LeapLogic (Version 4.8) Deployment
    • System Requirements
    • Prerequisites
    • Deployment
      • Extracting Package
      • Placing License Key
      • Executing Deployment Script
      • Accessing LeapLogic
    • Uploading License
    • Appendix
    • Getting Help
  • Removed Features
    • Configuring File Validation Stage
    • Variable Extractor Stage
      • Variable Extractor Report
    • Configuring Meta Diff Stage
      • Meta Diff
    • Configuring Data Load Stage
      • Data Load
    • Configuring Multi Algo Stage
  • FAQs
  • Tutorial Videos
  • Notice
Home   »  Batch Processing using Pipeline   »  Designing Pipeline   »  How to create a pipeline   »  Configuring ETL Conversion Stage  »  Configuring Informatica

Configuring Informatica

This topic provides steps to configure Informatica conversion stage.

  1. Select the ETL Type as Informatica.
  2. In Input Artifacts, upload the source data via:
    • Browse Files: To select the source files from the local system.
    • Select From Data Source: To select the source files from the data source. To do so, follow the steps below:
      • Click Select From Data Source.
      • Choose repository.
      • Select data source.
      • Select the entities.
      • Click to save the source data source.
  1. Choose the Target Type to which you need to transform the source scripts. The Target types are:
    1. AWS Glue Job
    2. AWS Glue Notebook
    3. AWS Glue Studio
    4. Data Build Tool
    5. Databricks Lakehouse
    6. Databricks Notebook
    7. Delta Live Tables
    8. Google BigQuery
    9. Matillion ETL
    10. Redshift ELT
    11. Snowflake
    12. Spark
  1. Click Data Configuration.

The Input column in the table below provides the input requirements based on the Target Type selection.

Target Type Input
Spark
  • Enable Spark-Native: External toggle to fetch input data from an external source such as Oracle, Netezza, Teradata, etc., and process that data in Spark, and then move the processed or output data to an external target. For instance, if the source input file contains data from any external source like Oracle, you need to select Oracle as the Databases to establish the database connection and load the input data. Then data is processed in Spark, and finally the processed or output data gets stored at an external target (Oracle). However, if you select Oracle as the Databases but the source input file contains data from an external source other than Oracle, such as Teradata, then by default, it will run on Spark.
    • In Databases, select the database you want to connect to. This establishes the database connection to load data from external sources like Oracle, Teradata, etc. If the database is selected, the converted code will have connection parameters (in the output artifacts) related to the database. If the database is not selected, you need to add the database connection details manually to the parameter file to execute the dataset; otherwise, by default, it executes on Spark.
  • In Output Type, the default output type for the transformation is set to Python.
  • In Validation Type, select None or Cluster as the mode of validation.
    • None: Select this option if you do not want to perform any validation.
    • Cluster: Select this option to perform syntax validation.
      • In Data Source, upload the corresponding data source. To successfully perform the syntax validation of the transformed queries, it is advisable to ensure that the required tables are created or already present on the target side and secondly all the user-defined functions (UDFs) are registered on the target data source.
  • In Default Database, select the source database to act as the default database in the transformed code such as in the converted lookup procedures, etc.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
 
Snowflake
  • In Output Type, select Python, Snowflake Scripting, or DBT as output type format for the generated artifacts.
  • In Validation Type, select None or Cluster as the mode of validation.
    • None: Select this option if you do not want to perform any validation.
    • Cluster: Select this option to perform syntax validation.
      • In Data Source, upload the corresponding data source. To successfully perform the syntax validation of the transformed queries, it is advisable to ensure that the required tables are created or already present on the target side and secondly all the user-defined functions (UDFs) are registered on the target data source.
  • In Default Database, select the source database to act as the default database in the transformed code such as in the converted lookup procedures, etc.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
AWS Glue Studio
  • To choose the required databases, turn on Is JDBC Required toggle and then select the required databases such as Oracle, or SQL Server.
  • In Target Database Details, specify database name, schema name, and prefix. The table name displays in prefix_database_tablename format if prefix is provided.
  • In AWS Glue Catalog Database, provide the AWS Glue Catalog Database connection details to connect the database and schema.
  • In S3 Bucket Base Path, provide the S3 storage repository path where you need to store the source and target files.
  • Specify the UDF File Location and UDF Jar Location to define the new UDF location.
  • In Parameter File, upload the parameter files to set the key values for the connection.
  • In Target Connection Name, provide a descriptive connection name or tag to identify who executed it.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
Databricks Notebook
  • In Output Type, select Python 3 or Jupyter as output type format for the generated artifacts.
  • In Data Interaction Technique, select your data interaction method. Following are the options:
    • Databricks-Native: Select Databricks-Native to fetch, process, and store data in Databricks Lakehouse.
    • Databricks: Unity Catalog: Select Databricks: Unity Catalog to access data via Databricks Unity Catalog. In Databricks, the Unity Catalog serves as a metadata repository from which data is fetched, processed, and stored within the catalog.
    • Databricks: External: Select this data interaction technique to fetch input data from an external source such as Oracle, Netezza, Teradata, etc., and process that data in Databricks, and then move the processed data or output to an external target. For instance, if the source input file contains data from any external source like Oracle, you need to select Oracle as the Source Database Connection to establish the database connection and load the input data. Then data is processed in Databricks, and finally the processed or output data gets stored at an external target (Oracle). However, if you select Oracle as the Source Database Connection but the source input file contains data from an external source other than Oracle, such as Teradata, then by default, it will run on Databricks.
      • If the selected data interaction technique is Databricks: External, you need to specify the source database of your data. In the Source Database Connection, select the database you want to connect to. This establishes the database connection to load data from external sources like Oracle, Teradata, etc. If the database is selected, the converted code will have connection parameters (in the output artifacts) related to the database. If the database is not selected, you need to add the database connection details manually to the parameter file to execute the dataset; otherwise, by default, it executes on Databricks.
  • In Validation Type, select None or Cluster as the mode of validation.
    • None: Select this option if you do not want to perform any validation.
    • Cluster: Select this option to perform syntax validation.
      • In Data Source, upload the corresponding data source. To successfully perform the syntax validation of the transformed queries, it is advisable to ensure that the required input tables are created or already present on the target side and secondly all the user-defined functions (UDFs) are registered on the target data source.
  • In Default Database, select the source database to act as the default database in the transformed code such as in the converted lookup procedures, etc.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
Databricks Lakehouse
  • In Data Interaction Technique, select your data interaction method. Following are the options:
    • Databricks-Native: Select Databricks-Native to fetch, process, and store data in Databricks Lakehouse.
    • Databricks: Unity Catalog: Select Databricks: Unity Catalog to access data via Databricks Unity Catalog. In Databricks, the Unity Catalog serves as a metadata repository from which data is fetched, processed, and stored within the catalog.
    • Databricks: External: Select this data interaction technique to fetch input data from an external source such as Oracle, Netezza, Teradata, etc., and process that data in Databricks, and then move the processed data or output to an external target. For instance, if the source input file contains data from any external source like Oracle, you need to select Oracle as the Source Database Connection to establish the database connection and load the input data. Then data is processed in Databricks, and finally the processed or output data gets stored at an external target (Oracle). However, if you select Oracle as the Source Database Connection but the source input file contains data from an external source other than Oracle, such as Teradata, then by default, it will run on Databricks.
      • If the selected data interaction technique is Databricks: External, you need to specify the source database of your data. In the Source Database Connection, select the database you want to connect to. This establishes the database connection to load data from external sources like Oracle, Teradata, etc. If the database is selected, the converted code will have connection parameters (in the output artifacts) related to the database. If the database is not selected, you need to add the database connection details manually to the parameter file to execute the dataset; otherwise, by default, it executes on Databricks.
  • In Validation Type, select None or Cluster as the mode of validation.
    • None: Select this option if you do not want to perform any validation.
    • Cluster: Select this option to perform syntax validation.
      • In Data Source, upload the corresponding data source. To successfully perform the syntax validation of the transformed queries, it is advisable to ensure that the required input tables are created or already present on the target side and secondly all the user-defined functions (UDFs) are registered on the target data source.
  • In Default Database, select the source database to act as the default database in the transformed code such as in the converted lookup procedures, etc.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
Delta Live Tables
  • In Data Interaction Technique, select your data interaction method. Following are the options:
    • Databricks-Native: Select Databricks-Native to fetch, process, and store data in Databricks Lakehouse.
      • Enable DLT Meta toggle to facilitate the creation of a bronze table within the Databricks Lakehouse. Rather than fetching data directly from the source such as flat files, this feature creates a bronze table (exact replica of the file) within Databricks and helps to refine data during data ingestion. With DLT Meta enabled, flat files are stored as tables within Databricks ensuring efficient data retrieval directly from these tables. This enhancement significantly boosts overall performance.
      • In DBFS Base Path, specify the DBFS location where the source flat files and DDL files are stored. This information is required to create the bronze table in Databricks.
    • Databricks: Unity Catalog: Select Databricks: Unity Catalog to access data via Databricks Unity Catalog. In Databricks, the Unity Catalog serves as a metadata repository from which data is fetched, processed, and stored within the catalog.
    • Databricks: External: Select this data interaction method to fetch input data from an external source such as Oracle, Netezza, Teradata, etc., and process that data in the Databricks, and then move the processed data or output to an external target. For instance, if the source input file contains data from any external source like Oracle, you need to select Oracle as the Source Database Connection to establish the database connection and load the input data. Then this data is processed in Databricks, and finally, the processed or output data gets stored at an external target (Oracle). However, if you select Oracle as the Source Database Connection but the source input file contains data from an external source other than Oracle, such as Teradata, then by default, it will run on Databricks.
      • Enable DLT Meta toggle to facilitate the creation of a bronze table within the Databricks Lakehouse. Rather than fetching data directly from the source such as flat files, this feature creates a bronze table (exact replica of the file) within Databricks and helps to refine data during data ingestion. With DLT Meta enabled, flat files are stored as tables within Databricks ensuring efficient data retrieval directly from these tables. This enhancement significantly boosts overall performance.
      • If the selected data interaction technique is Databricks: External, you need to specify the source database of your data. In the Source Database Connection, select the database you want to connect to. This establishes the database connection to load data from external sources like Oracle, Teradata, etc. If the database is selected, the converted code will have connection parameters (in the output artifacts) related to the database. If the database is not selected, you need to add the database connection details manually to the parameter file to execute the dataset; otherwise, by default, it executes on Databricks.
      • In DBFS Base Path, provide the DBFS base location where the source flat files and DDL files are stored. This information is required to create the bronze table in Databricks.
  • In Validation Type, select None or Cluster as the mode of validation.
    • None: Select this option if you do not want to perform any validation.
    • Cluster: Select this option to perform syntax validation.
      • In Data Source, upload the corresponding data source. To successfully perform the syntax validation of the transformed queries, it is advisable to ensure that the required input tables are created or already present on the target side and secondly all the user-defined functions (UDFs) are registered on the target data source.
  • In Default Database, select the source database to act as the default database in the transformed code such as in the converted lookup procedures, etc.
  • In Dependent Utility Path, provide the DBFS location where the utility files are stored as a wheel binary package. The wheel file contains Python libraries that are required to execute the conversion on the target.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
AWS Glue Job
  • In Data Interaction Technique, select your data interaction method. Following are the options:
    • Glue: Redshift: Select Glue-Redshift to fetch input data from Amazon Redshift, process it in Glue, and store the processed or output data in Redshift. In this scenario, source data are converted to Redshift whereas temporary or intermediate tables are converted to Spark.
    • Glue: Data Catalog: This method accesses data through the data catalog which serves as a metadata repository. Then the data is processed in Glue and the processed or output data gets stored in the data catalog.
      • In Storage Format, select the storage format of your data such as Delta or Iceberg.
    • Glue: External: Select this data interaction method to fetch input data from an external source such as Oracle, Netezza, Teradata, etc., and process that data in Glue, and then move the processed or output data to an external target. For instance, if the source input file contains data from any external source like Oracle, you need to select Oracle as the Source Database Connection to establish the database connection and load the input data. Then data is processed in Glue, and finally the processed or output data gets stored at the external target (Oracle). However, if you select Oracle as the Source Database Connection but the source input file contains data from an external source other than Oracle, such as Teradata, then by default, it will run on Redshift.
      • If the selected data interaction technique is Glue: External, you need to specify the source database of your data. In the Source Database Connection select the database you want to connect to. This establishes the database connection to load data from external sources like Oracle, Teradata, etc. If the database is selected, the converted code will have connection parameters (in the output artifacts) related to the database. If the database is not selected, you need to add the database connection details manually to the parameter file to execute the dataset; otherwise, by default, it executes on Redshift.
    • Redshift ETL Orchestrated via Glue: This method accesses, processes, and executes data in Amazon Redshift and uses Glue for orchestration jobs. In this scenario, both source data and intermediate tables are converted to Redshift.
    • Glue: Hybrid: This interaction technique can leverage three different data interaction techniques - Glue: Redshift, Glue: Data Catalog, and Glue: External. Depending upon your use case, it can either take input from the data sources such as Redshift, Delta, Iceberg, RDS instance, or an external source; process it in Glue; and move the output (processed) data into the respective data sources. To define the tables and data sources, download the template (CSV file) and specify the tables with their database type etc. For instance, you can select this option if your source tables reside on multiple data sources such as Redshift, Delta, Iceberg, external sources, and RDS instances. It takes input from the specified data sources, processes in Glue, and moves the output (processed) data into the respective data sources.
      • In Input File, upload the input file (CSV file) that contains information about tables including their database type and name, to define the tables residing on various data sources. 
  • In Default Database, select the source database to act as the default database in the transformed code such as in the converted lookup procedures, etc.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
Matillion ETL
  • In Output Type, the default output type for the transformation is set to JSON.
  • In Data Interaction Technique, select your data interaction method. Following are the options:
    • Snowflake - Native: Select Snowflake - Native to fetch, process, and store data in Snowflake.
    • Snowflake: External: Select this data interaction technique to fetch input data from an external source such as Oracle, Netezza, Teradata, etc., and process that data in Snowflake, and then move the processed data or output to an external target. For instance, if the source input file contains data from any external source like Oracle, you need to select Oracle as the Source Database Connection to establish the database connection and load the input data. Then data is processed in Snowflake, and finally the processed or output data gets stored at an external target (Oracle). However, if you select Oracle as the Source Database Connection but the source input file contains data from an external source other than Oracle, such as Teradata, then by default, it will run on Snowflake.
      • If the selected data interaction technique is Snowflake: External, you need to specify the source database of your data. In the Source Database Connection, select the database you want to connect to. This establishes the database connection to load data from external sources like Oracle, Teradata, etc. If the database is selected, the converted code will have connection parameters (in the output artifacts) related to the database. If the database is not selected, you need to add the database connection details manually to the parameter file to execute the dataset; otherwise, by default, it executes on Snowflake.
  • In Default Database, select the source database to act as the default database in the transformed code such as in the converted lookup procedures, etc.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
 
Redshift ELT
  • Enable Redshift: External toggle to fetch input data from an external source such as Oracle, Netezza, Teradata, etc., and process that data in Redshift, and then move the processed or output data to an external target. For instance, if the source input file contains data from any external source like Oracle, you need to select Oracle as the Databases to establish the database connection and load the input data. Then data is processed in Spark, and finally the processed or output data gets stored at an external target (Oracle). However, if you select Oracle as the Databases but the source input file contains data from an external source other than Oracle, such as Teradata, then by default, it will run on Redshift.
    • In Databases, select the database to define the database connection to load the data from external sources such as Oracle, Teradata, etc. If the database is selected, the converted code will have connection parameters (in the output artifacts) related to the database. If the database is not selected, you need to add the database connection details manually to the parameter file to execute the dataset; otherwise, by default, it executes on Redshift.
  • In Output Type, the default output type for the transformation is set to Python.
  • In Database Name, provide the target database name to which you need to store the transformed data.
  • In S3 Bucket Base Path, provide the S3 storage repository path where you need to store the source and target files.
  • In Parameter File, upload the parameter files to set the key values for the connection.
  • In Target Connection Name, provide a descriptive connection name or tag to identify who executed it.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
 
AWS Glue Notebook
  • In Data Interaction Technique, select your data interaction method. Following are the options:
    • Glue-Redshift: Select Glue-Redshift to fetch input from Amazon Redshift, process it in Glue, and store the processed data or output in Redshift. In this scenario, source data are converted to Redshift whereas temporary or intermediate tables are converted to Spark.
    • Glue: Data Catalog: This method accesses data through the data catalog which serves as a metadata repository. Then the data is processed within Glue and the processed or output data gets stored in the data catalog.
      • In Storage Format, select the storage format of your data such as Delta or Iceberg.
    • Glue: External: Select this data interaction method to fetch input data from an external source such as Oracle, Netezza, Teradata, etc., and process that data in the Glue, and then move the processed or output data to an external target. For instance, if the source input file contains data from any external source like Oracle, you need to select Oracle as the Source Database Connection to establish the database connection and load the input data. Then data is processed in Glue, and finally, the processed or output data gets stored at an external target (Oracle). However, if you select Oracle as the Source Database Connection but the source input file contains data from an external source other than Oracle, such as Teradata, then by default, it will run on Redshift.
      • If the selected data interaction technique is Glue: External, you need to specify the source database of your data. In the Source Database Connection select the database you want to connect to. This establishes the database connection to load data from external sources like Oracle, Teradata, etc. If the database is selected, the converted code will have connection parameters (in the output artifacts) related to the database. If the database is not selected, you need to add the database connection details manually to the parameter file to execute the dataset; otherwise, by default, it executes on Redshift.
    • Redshift ETL Orchestration via Glue: This method accesses, processes, and executes data in Amazon Redshift and uses Glue for orchestration jobs. In this scenario, both source data and intermediate tables are converted to Redshift. 
  • In Default Database, select the source database to act as the default database in the transformed code such as in the converted lookup procedures, etc.
  • In Property file path, provide the S3 repo path where the property files are stored.
  • In Dependent Utility Path, provide the S3 repo path where the utility files are stored as a wheel binary package.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
Data Build Tool
  • In Data Interaction Technique, select your data interaction method. Following are the options:
    • Snowflake - Native: Select Snowflake - Native to fetch, process, and store data in Snowflake.
    • Snowflake: External: Select this data interaction technique to fetch input data from an external source such as Oracle, Netezza, Teradata, etc., and process that data in Snowflake, and then move the processed data or output to an external target. For instance, if the source input file contains data from any external source like Oracle, you need to select Oracle as the Source Database Connection to establish the database connection and load the input data. Then data is processed in Snowflake, and finally the processed or output data gets stored at an external target (Oracle). However, if you select Oracle as the Source Database Connection but the source input file contains data from an external source other than Oracle, such as Teradata, then by default, it will run on Snowflake.
      • If the selected data interaction technique is Snowflake: External, you need to specify the source database of your data. In the Source Database Connection, select the database you want to connect to. This establishes the database connection to load data from external sources like Oracle, Teradata, etc. If the database is selected, the converted code will have connection parameters (in the output artifacts) related to the database. If the database is not selected, you need to add the database connection details manually to the parameter file to execute the dataset; otherwise, by default, it executes on Snowflake.
  • In Default Database, select the source database to act as the default database in the transformed code such as in the converted lookup procedures, etc.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
 
Google BigQuery
  • In Orchestration Technique, select your orchestration method. Following are the options:
    • Google Cloud Composer: Select this option to generate Google Cloud Composer equivalent artifacts through which you can create, monitor, and manage the workflows defined in the DAG (Directed Acyclic Graph).
    • Python: Select this option to generate Python artifacts.
  • In Data Interaction Technique, select your data interaction method from the following options:
    • Google BigQuery - Native: Select this option to access the input data, process it, and store the output data in Google BigQuery.
    • Google BigQuery: External: Select this option to fetch input data from an external source such as Oracle, Netezza, Teradata, etc., and process that data in Google BigQuery, and then move the processed or output data to an external target. For instance, if the source input file contains data from any external source like Oracle, you need to select Oracle as the source database to establish the database connection and load the input data. Then data is processed in Google BigQuery, and finally the processed or output data gets stored at the external target (Oracle). However, if you select Oracle as the source database connection but the source input file contains data from an external source other than Oracle, such as Teradata, then by default, it will run on Google BigQuery.
      • If the selected data interaction technique is Google BigQuery: External, you need to specify the source database of your data. In the Source Database Connection, select the database you want to connect to. This establishes the database connection to load data from external sources like Oracle, Teradata, etc. If the database is selected, the converted code will have connection parameters (in the output artifacts) related to the database. If the database is not selected, you need to add the database connection details manually to the parameter file to execute the dataset; otherwise, by default, it executes on Google BigQuery.
  • In Validation Type, select None or Cluster as the mode of validation.
    • None: Select this option if you do not want to perform any validation.
    • Cluster: Select this option to perform syntax validation.
      • In Data Source, upload the corresponding data source. To successfully perform the syntax validation of the transformed queries, it is advisable to ensure that the required input tables are created or already present on the target side and secondly all the user-defined functions (UDFs) are registered on the target data source.
      • In GCS Base Path, specify the GCS base path where external files are stored.
  • In Default Database, select the source database to act as the default database in the transformed code such as in the converted lookup procedures, etc.
  • In Source Data Source, select the data source (DDL) which contains corresponding metadata to ensure accurate query conversion.
 

  1. In Transformation Model, select the required transformation model. The selected model converts the queries that are not converted by LeapLogic Transformation Engine to target equivalent. This process enhances performance and efficiency. Use pretrained transformation models such as Energize, Embark, or Velocity to convert small to medium sized ETL scripts. To convert complex ETL scripts, use Intercept pretrained model.
  2. Turn on I agree toggle to transform code elements from supported sources, however, elements from unsupported sources will not be transformed. All code elements including those from unsupported sources are considered in the conversion automation calculation. If you do not turn on, this stage will fail if it encounters dialects of any unsupported sources in the input code.
  1. Click Save to update the changes.
  2. An alert pop-up message appears. This message prompts you to refer your respective assessment to determine the anticipated quota deduction required when converting your scripts to target. Then click Ok.

  1. Click  to provide a preferred pipeline name.
  2. Click  to execute the pipeline. Clicking  (Execute) navigates you to the listing page which shows your pipeline status as Running state. It changes its state to Success when it is completed successfully.
  3. Click pipeline card to see report.

To view the Informatica conversion report, visit Informatica Conversion Report


Next:

Configuring DataStage


To learn more, contact our support team or write to: info@leaplogic.io

Copyright © 2025 Impetus Technologies Inc. All Rights Reserved

  • Terms of Use
  • Privacy Policy
  • License Agreement
To the top ↑ Up ↑