Known issues in 1.5.5 SP1

You might run into some known issues while using Cloudera Data Engineering on premises.

Cloudera Data Engineering on premises 1.5.5 existing known issues are carried into Cloudera Data Engineering on premises 1.5.5 SP1. For more details, see Known issues and limitations in Cloudera Data Engineering on Cloudera Data Services on premises 1.5.5. For known issues fixed since Cloudera Data Engineering on premises 1.5.5, see Fixed issues in 1.5.5 SP1.

DEX-18429: Cloudera Data Engineering Service fails to start displaying the cluster with same name already exists error
If a user without the DEAdmin role tries to create a Cloudera Data Engineering Service using the API, the request is accepted, but only a database entry is created without generating the corresponding resources. When the user tries to create that Service again, it fails displaying the cluster with same name already exists error.
Ensure that the user has the DEAdmin role before creating a new Cloudera Data Engineering Service. If you accidentally come across such an issue, reach out to the Cloudera Support team for deleting this entry from the database.
DEX-7459: The numExecutors parameter is missing in job configuration for new jobs
The numExecutors parameter is missing in job configuration for new jobs created using Cloudera Data Engineering UI and jobs cloned using Cloudera Data Engineering UI or CDE CLI. The numExecutors parameter for a cloned job is not preserved.
Edit the Cloudera Data Engineering job after cloning or creating it and update the numExecutors parameter.
DEX-18216: Cloudera Embedded Container Service cluster is down and is inaccessible after creating 4,700 pods for a single Spark job
The Cloudera Embedded Container Service cluster is down and is inaccessible after creating 4,700 pods for a single Spark job.
Stop the Cloudera Embedded Container Service server and the Cloudera Embedded Container Service agents. Clean up all the orphan pods and restart the Cloudera Embedded Container Service services to recover the cluster.
DEX-18287: Expired certificate in repository causes GIT repository synchronization failure
If the certificate in the repository is expired, then the GIT repository synchronization fails.
Delete the repository from all the jobs. Recreate the repository again with a new valid certificate.
DEX-18175: dex-base pods enter the Error' / 'ContainerStatusUnknown status when running job count exceeds 130
When the running job count exceeds 130, dex-base pods enter the Error' / 'ContainerStatusUnknown status.
For a workaround, see Configuring Spark jobs for huge shuffle data.
DEX-18528: While using insecure Spark images, virtual cluster creation passes, but the job gets stuck
If the insecure Spark images are removed from the registry and if you try to install the virtual cluster with the insecure image option, the virtual cluster installation succeeds. But, the subsequent job runs will fail.
Do not install a virtual cluster with the insecure image options if the corresponding images are removed from the registry. You can delete the insecure virtual cluster if you have created one and instead create a secure version of the same.