Hive
You can review the list of reported issues and their fixes for Hive in 7.3.1.200.
- CDPD-78342/CDPD-72605: Optimized partition authorization in HiveMetaStore to reduce overhead
- The
add_partitions()
API in HiveMetastore was authorizing both new and existing partitions, leading to unnecessary processing and increased load on the authorization service. - CDPD-77990: Upgraded MySQL Connector/J to 8.2.0 to fix CVE-2023-22102
- The existing MySQL Connector/J version was vulnerable to CVE-2023-22102.
- CDPD-62654/CDPD-77985: Hive Metastore now sends a single AlterPartitionEvent for bulk partition updates
- HiveMetastore previously sent individual AlterPartitionEvent for each altered partition, leading to inefficiencies and pressure on the back db.
- CDPD-73669: Secondary pool connection starvation caused by updatePartitionColumnStatisticsInBatch API
- Hive queries intermittently failed with
Connection is not available, request timed out
errors. The issue occurred because the updatePartitionColumnStatisticsInBatch method in ObjectStore used connections from the secondary pool, which had a pool size of only two, leading to connection starvation. - CDPD-61676/CDPD-78341: Drop renamed external table fails due to missing update in PART_COL_STATS
- When hive.metastore.try.direct.sql.ddl is set to false, dropping an
external partitioned table after renaming it fails due to a foreign key constraint error in the
PART_COL_STATS
table. The table name inPART_COL_STATS
is not updated during the rename, causing issues during deletion. - CDPD-79469: Selecting data from a bucketed table with a decimal column throws NPE
- When hive.tez.bucket.pruning is enabled,
selecting data from a bucketed table with a decimal column type fails with a
NullPointerException
. The issue occurs due to a mismatch in decimal precision and scale while determining the bucket number, causing an overflow and returning null. - CDPD-74095: Connection timeout while inserting Hive partitions due to secondary connection pool limitation
- Since HIVE-26419, Hive uses a secondary connection pool (size 2) for
schema and value generation. However, this pool also handles nontransactional connections,
causing the
updatePartitionColumnStatisticsInBatch
request to fail with aConnection is not available, request timed out
error when the pool reaches its limit during slow insert or update operations. - CDPD-78331: HPLSQL built-in functions fail in insert statement
- After the HIVE-27492 fix, some HPLSQL built-in functions like trim and lower stopped working in INSERT statements. This happened because UDFs already present in Hive were removed to avoid duplication, but HPLSQL's local and offline modes still required them.
- CDPD-78343: Syntax error in HPL/SQL error handling
- In HPL/SQL, setting hplsql.onerror using the SET command resulted in a syntax error because the grammar file (Hplsql.g4) only allowed identifiers without dots (.).
- CDPD-78330: HPL/SQL built-in functions like sysdate not working
- HPL/SQL built-in functions that are not available in Hive, such as sysdate, were failing with a SemanticException when used in queries. Only functions present in both HPL/SQL and Hive were working.
- CDPD-78345: Signalling CONDITION HANDLER is not working in HPLSQL
- The user-defined
CONDITION HANDLER
s in HPLSQL are not being triggered as expected. Instead of running the handlers, the system only logs the conditions, so the handlers aren't available when needed. - CDPD-78333: EXECUTE IMMEDIATE throwing ClassCastException in HPL/SQL
- When executing a
select count(*)
query, it returns a long value, but HPLSQL expects a string. This mismatch causes the following error:Caused by: java.lang.ClassCastException: class java.lang.Long cannot be cast to class java.lang.String at org.apache.hive.service.cli.operation.hplsql.HplSqlQueryExecutor$OperationRowResult.get
- CDPD-79844: EXECUTE IMMEDIATE displaying error despite successful data load
- Running
EXECUTE IMMEDIATE 'LOAD DATA INPATH ''/tmp/test.txt'' OVERWRITE INTO TABLE test_table'
displayed an error on the console, even though the data was successfully loaded into the table. This occurred because HPL/SQL attempted to check the result set metadata after execution, but LOAD DATA queries do not return a result set, leading to aNullPointerException
. - CDPD-67033: HWC for Spark 3 compatibility with Spark 3.5
- The Spark 3.5, based on Cloudera on cloud 7.2.18 libraries, caused a failure in the HWC for Spark 3 build. Canary builds indicate that broke compatibility.
- CDPD-80097: Datahub recreation fails due to Hive Metastore schema validation error
- Datahub recreation on Azure fails because Hive Metastore schema
validation cannot retrieve the schema version due to insufficient permissions on the
VERSION
table.