Skip to content

[BUG]: Informatica Jobs Workflow json does not have the job_cluster defined in the task #2132

@CaryMoore-DB

Description

@CaryMoore-DB

Is there an existing issue for this?

  • I have searched the existing issues

Category of Bug / Issue

Converter bug

Current Behavior

When I tried to create a job with the .json file from the transpiler, there was no "jobs_cluster" definition for the key. Following needs to be added.

"job_clusters": [
{
"job_cluster_key": "job_cluster",
"new_cluster": {
"spark_version": "13.3.x-scala2.12",
"node_type_id": "i3.xlarge",
"num_workers": 1
}
}
]

Expected Behavior

The .json works with simply performing a search and replace for expected changes.

  • Email and user name for the user
  • Path to notebooks
  • Existing cluster ID (although I think we should just use the same jobs cluster)

Also, could use documentation on what the user would have to change, and maybe provide that in the config.yaml

Steps To Reproduce

Create a curl for the existing json, replace the above as appropriate and then execute, you will get an error on the jobs_cluster not having a definition.

Relevant log output or Exception details

{"error_code":"INVALID_PARAMETER_VALUE","message":"Job cluster 'job_cluster' is not defined in field 'job_clusters'.","details":[{"@type":"type.googleapis.com/google.rpc.RequestInfo","request_id":"f41d8202-c6f7-4cc9-a7ef-08fb28f3bc8c","serving_data":""}]}

Logs Confirmation

  • I ran the command line with --debug
  • I have attached the lsp-server.log under USER_HOME/.databricks/labs/remorph-transpilers/<converter_name>/lib/lsp-server.log

Sample Query

Operating System

macOS

Version

latest via Databricks CLI

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions