Configuring Cloud-to-Cloud Transformation
Cloud-to-cloud transformation is a process of transforming all workloads from an existing cloud to a modern cloud. In other words, moving from one cloud to another. This topic provides steps to configure cloud-to-cloud transformation. To do so, follow the below steps.
- In Source Type, select the cloud source data store such as Snowflake.
- In Input Type, the default input type for the transformation is set to SQL/Procedure.
- In Target Type, select the preferred target type, such as Databricks Lakehouse or Databricks Notebook.
- In Output Type, the default output type for the transformation is set to Python.
- In Input Artifacts, upload the files that you need to transform to the target source.
- Click Data Configuration to configure the data.
- If the selected target is Databricks Lakehouse, enable Unity Catalog to transform the constraints such as primary key, foreign key, etc., present in the source data to the Databricks-native equivalent. Otherwise, the default Hive meta store will be used, and Hive or Spark tables will be generated without any constraints. In Databricks, there are two types of catalogs:
- Hive Meta store: Hive meta store serves as a central repository to create, store, and manage large datasets or tables. However, it does not support constraints like primary key, foreign key, etc.
- Unity Catalog: Unity Catalog in Databricks serves as a centralized metadata repository with advanced metadata management features that support constraints including primary key, foreign key, etc. It serves as a comprehensive repository for storing metadata information within Databricks.
- In Validation Type, select the validation type:
- None: Performs no validation.
- Cluster: Validates the queries that are transformed by the LeapLogic Core transformation engine.
- In LeapFusion, select the preferred Base and Helper models to convert queries that require augmented transformation and validate them for accuracy and optimal performance.
- Base Model: Select the preferred base model to convert queries that are not handled by the default LeapLogic Core transformation engine to the target equivalent. The Base Model supports both offline (LeapLogic) and online (Amazon Bedrock) modes of transformation. By default, the system enables online mode (Amazon Bedrock) and disables offline mode (LeapLogic). To enable the offline modes (LeapLogic) such as Embark, Energize, Intercept, and Velocity, as a prerequisite you need to provide EC2 instance connection details and a prompt file (.txt) on the Add New Sources and Targets page (Governance > Intelligence Modernization > Custom/ Source/ Target > Add New Sources and Targets).
To view the detailed steps for providing EC2 instance connection details and the prompt file, click here.
The offline modes include:
- Energize, Velocity, and Embark: To convert small- to medium-sized SQLs.
- Intercept: To convert large-sized SQLs and procedural code.
- Helper Model: Select helper models to validate the queries that are transformed by the Base Model and suggest corrections if needed. By default, the selected base model is set as a helper model. If needed, you can add multiple helper models. When multiple Helper models are configured:
- The first helper model (same as the base model) validates the queries transformed by the base model. If any queries are incorrect, the Helper model suggests the correct queries.
- Then, the corrected queries are passed to the next Helper model, which validates them and suggests corrections if required.
- This process continues through all configured Helper Models until the queries are validated successfully.
This validation process ensures higher accuracy, better performance, and more efficient transformation.
To access this intelligent modernization feature (LeapFusion), ensure that your account has the manager and llexpress_executor roles.
To view the detailed steps for assigning manager and llexpress_executor roles to your account, click here.
- In Source, select the configuration as Live or Offline.
- If the selected source configuration is:
- Live: Upload the data source.
- Offline: Upload the DDL files. It supports .sql and .zip file formats.
- In Target, select the configuration as Live.
- Upload the target data source.
- In File Format, select the storage type as Delta, ORC, Parquet, etc.
- In Mapping as per Changed Data Model, upload files for mapping between source and target tables.
- In Databricks Recommendations Report, upload the Databricks recommendations (databricksrecommendation.csv) report generated as part of the corresponding assessment, which enables the system to incorporate relevant optimizations into the converted DDL queries.
To obtain databricksrecommendation.csv report follow the steps below:
- Go to Reports section in assessment.
- Download Insights and Recommendations.
- The downloaded archived file contains the databricksrecommendation.csv report. If required, you can update details such as cluster_by_columns.

- Upload databricksrecommendation.csv file in the Databricks Recommendations Report field.
- Click Save to save the Transformation stage.
- An alert pop-up message appears. This message prompts you to refer to your respective assessment to determine the anticipated quota deduction when converting your scripts. Then click Ok.
- Click
to provide a preferred pipeline name.
- Click the Execute icon to execute the integrated or standalone pipeline. Clicking the Execute icon navigates you to the pipeline listing page which shows your pipeline status as Running state. It changes its state to Success when it is completed successfully.
- Click on your pipeline card to see reports.
To view the Cloud-to-Cloud Transformation, visit Cloud-to-Cloud Transformation Report.