August 05, 2025 - Hotfix
Review the fixed issues and changed behaviors in this hotfix release of Cloudera Data Warehouse on cloud.
What's new in Cloudera Data Warehouse on cloud
Review the new features introduced in this release of Cloudera Data Warehouse service on Cloudera on cloud.
What's new in Cloudera Data Warehouse on Azure environments
- Azure AKS 1.32 upgrade
- Cloudera supports the Azure Kubernetes Service (AKS) version 1.32. In 1.10.3-b5 (released August 05, 2025), when you activate an Environment, Cloudera Data Warehouse automatically provisions AKS 1.32. To upgrade to AKS 1.32 from an earlier version of Cloudera Data Warehouse, you must backup and restore Cloudera Data Warehouse.
What's new in Cloudera Data Warehouse on AWS environments
- AWS EKS 1.32 upgrade
- Cloudera supports the AWS Elastic Kubernetes Service (EKS) version 1.32. In 1.10.3-b5 (released August 05, 2025), when you activate an Environment, Cloudera Data Warehouse automatically provisions EKS 1.32. To upgrade to EKS 1.32 from an earlier version of Cloudera Data Warehouse, you must backup and restore Cloudera Data Warehouse.
Fixed issues
Review the fixed issues in this release of the Cloudera Data Warehouse service on cloud.
- DWX-21314: Streaming Kubernetes API calls failing on private clusters
- Streaming Kubernetes API calls were previously failing on private clusters across Azure and Amazon due to issues with the CCMv2 channel. This impacted virtual warehouse and database catalog rebuild functionality on private AKS/EKS clusters.
- DWX-21330: Metastore initialization timeout
- The metastore-sys-init job, responsible for metastore initialization, previously had a 20-minute deadline. In certain scenarios, initialization could exceed this period, leading to failures. This issue is now resolved by increasing the deadline to 1 hour.
- DWX-21043: AKS node image auto-upgrade disabled
- Previously, Azure Kubernetes Service (AKS) introduced automatic node
image upgrades as the default behavior starting with API version 2023-06-01. However, this
default setting can lead to upgrade-related issues in Cloudera Data Warehouse.
This issue is now resolved by disabling the AKS node image auto-upgrade feature, restoring the behavior to align with the pre-2023-06-01 API version.
- Cookie-Based authentication support for JWT tokens
- When JWT tokens are used for authentication, every
HTTP
request within a session requires token verification. If these tokens have a short lifespan, it can lead to authentication failures and disrupt session continuity. - CDPD-80798: Stable Catalogd initialization in HA mode
- Catalogd initialization previously might timeout to complete in high availability mode. This happened because metadata operations started prematurely, blocking Catalogd from becoming active.
- CDPD-83059: Optimized Impala Catalog cache warmup
- Impala's Catalogd previously started with an empty cache. This led to slow query startup for important tables and affected high availability failovers.
- CDPD-87222: Consistent TRUNCATE operations for external tables
- Impala's
TRUNCATE
operations on external tables previously did not consistently delete files in subdirectories, even when recursive listing was enabled. - CDPD-83530: Task commits were allowed despite an exception being thrown in the Tez processor
- A communication failure between the coordinator and executor caused a
running task to terminate, resulting in a
java.lang.InterruptedException
being thrown by theReduceRecordProcessor.init()
. Despite this exception, the process still allowed the task to be committed and generated a commit manifest.This issue has now been resolved. The fix ensures that outputs are not committed if an exception is thrown in the Tez processor.
Apache Jira: HIVE-28962
- CDPD-82415:
TABLESAMPLE
clause of theCOMPUTE STATS
statement has no effect on Iceberg tables - This fix resolves a regression introduced by IMPALA-13737. For example, the following query scans the entire Iceberg table to
calculate statistics, whereas it should ideally use only about 10% of the
data.
COMPUTE STATS t TABLESAMPLE SYSTEM system(10);
This fix introduces proper table sampling logic for Iceberg tables, which can be utilized for
COMPUTE STATS
. The sampling algorithm previously located inIcebergScanNode.getFilesSample()
is now relocated toFeIcebergTable.Utils.getFilesSample()
.Apache Jira: IMPALA-14014
- CDPD-85228:
IllegalStateException
with Iceberg table with DELETE - Running a query on an Iceberg table fails with an
IllegalStateException
error in the following scenario:- The Iceberg table has delete files for every data file (no data files without delete files) AND
- An anti-join operation is performed on the result of the Iceberg delete operation (IcebergDeleteNode or HashJoinNode)
This fix resolves the issue by setting the
TableRefIds
of the node corresponding to the Iceberg delete operation (IcebergDeleteNode or HashJoinNode) to only the table reference associated with the data files, excluding the delete files.Apache Jira: IMPALA-14154
- CDPD-87405: Error unnesting arrays in Iceberg tables with DELETE files
- The following error occurred when unnesting a nested array (a 2D
array) from an Iceberg table. This issue was triggered specifically when the table contained
delete files for some, but not all, of its data
files.
Filtering an unnested collection that comes from a UNION [ALL] is not supported yet.
Reading an Iceberg table with this mixed data and delete file configuration creates a
UNION ALL
node in the query execution plan. The system had a check that explicitly blocked any filtering on an unnested array.This fix relaxes the validation check, allowing the operation to proceed if all
UNION
operands share the same tuple IDs. This ensures the query can successfully unnest the array.Apache Jira: IMPALA-14185
- DWX-21173: Allow
redirect_url
configurable forlibsaml
- Previously, Hue's libsaml configuration did not allow for a
configurable
redirect_url
to support custom SAML or Okta IdP addresses. The issue is now resolved by adding a custom configuration option forlibsaml redirect_url
.[libsaml] redirect_url=<customer_own_okta_or_saml_idp>
Behavior changes
This release of the Cloudera Data Warehouse service on cloud has the following behavior changes:
Summary: Cleanup subdirectories in truncate/insert overwrite if recursing listing is enabled
Before this release: Impala did not consistently delete files
located in subdirectories of external tables during TRUNCATE
and INSERT
OVERWRITE
operations, even when recursive listing was enabled. This led to leftover
data in subdirectories after these operations, resulting in data corruption.
After this release: After this change, directories are also
deleted in addition to (non-hidden) data files, with the exception of hidden and ignored
directories. Now, setting DELETE_STATS_IN_TRUNCATE=false
is no longer supported
by default when truncating non-transactional tables; attempting this will result in an
exception. If the old behavior is absolutely required, you can set the
--truncate_external_tables_with_hms flag to false, but be aware that this
will also reintroduce the bug that was fixed by this change.
Apache Impala: IMPALA-14189, IMPALA-14224