Known issues in 1.5.4 SP2
You might run into some known issues while using Cloudera AI on premises.
- DSE-42079: Cloudera AI Workbenches are not compatible with the Unified timezone feature
-
When you enable the Unified timezone feature, the Cloudera Embedded Container Service cluster timezone is synchronized with the Cloudera Manager Base time zone, and the Cloudera AI sessions fails to launch with Exit Code 34. Timestamp discrepancies with workloads are displayed.
Workaround:
If you use Cloudera AI Workbenches, disable the Unified Timezone feature by following the instructions in Cloudera Embedded Container Service unified time zone.
- DSE-41757: Python workloads running multiple-line comments or strings might fail
-
Python workloads running multiple-line comments or strings might fail to run when using the Workbench Editor.
Workaround:
Run the code using the PBJ Workbench Editor.
- DSE-42509: Creating a project does not work on NTP/Airgap proxy when using private repository with SSH key and SSH URL
-
When creating a project on NTP/Airgap proxy, using a private repository with an SSH key and an SSH URL, an error message is displayed that the project cannot be created.
No workaround is available.
- DSE-42510: Fetching Cloudera or Huggingface AMP catalog is failing on NTP/Airgap proxy
- When fetching Cloudera or Huggingface Cloudera Accelerators for
Machine Learning Project (AMP) catalog on NTP/Airgap proxy, an error message is displayed
Error fetching AMP catalog source URL.
No workaround is available.
- DSE-44698: Restart Stabilty - Cloudera AI Workbench pod preemption by system-critical workloads
-
During control plane upgrades or cluster restarts, Cloudera AI Workbench pods may transition into the
Init:Error
orInit:ContainerStatusUnknown
state.This issue may arise during cluster startup or under resource pressure when the scheduler preempts lower-priority Cloudera AI Workbench pods to allocate resources for higher-priority system pods, such as critical system components. Additionally, Kubernetes does not automatically clean up preempted pods, leaving them in failed
Init
states.This is the expected Kubernetes scheduler behavior. There is no permanent fix available, as pod preemption is controlled by the Kubernetes scheduler and is necessary for system stability during resource constraints.
Workaround:
Delete the affected pod, and Kubernetes automatically attempts to reschedule it if sufficient resources are available:kubectl delete pod [***POD NAME***] -n [***CLOUDERA AI WORKBENCH NAMESPACE***]