Known issues
Learn about the known issues in Cloudera Data Catalog, the impact or changes to the functionality, and the workaround.
- CDPDSS-1953: Location field does not load absolute path for entities in contents tab
- Problem:After navigating to the Azure Blob which is created within the Azure directory
and later clicking the "Content" tab, the path for Azure blob is displayed as "/" instead of
the path that mentions:
containers/directory/Blob
. - CDPDSS-1956: Issue with count on search page as AWS S3 V2 Object
- Problem: The count against the Data Lake displays only the count of AWS S3 V2 Object entity type.
- CDPDSS-31173: Atlas API on saving the classifications in Cloudera Data Catalog is failing while using Cloudera Runtime version 7.2.12
- Problem:When the suggested classification by Custom Sensitivity Profiler is saved, Atlas API throws an error response and it is not saved in Cloudera Data Catalog. This behavior is specifically observed in Cloudera Runtime version 7.2.12.
- CDPDSS-2127: On profiler cluster creation step 'external database in creation' the message on search page is not displayed
- Problem: The message on Cloudera Data Catalog Search page is not displayed in 'blue' section saying "profilers cluster is getting created" .
- CDPDSS-2138: ODP fails on an asset imported from Glue to Cloudera on a medium duty cluster
- Problem: The profiler run fails on the table.
- CDPDSS-2142: Livy job in Profiler run has failed but profiling status of assets is shown as started
- Problem: The Livy job fails but each assets shows profiling status as started.
- CDPD-32000: Atlas intermittently does not send the classification information to Cloudera Data Catalog
- Problem: On loading of search page filters with 'entity type' filter applied to 'Hive', intermittently the tags/classifications are not loaded in Cloudera Data Catalog App. Due to this, Cloudera Data Catalog does not load the filters for 'entity tag' and column tag.
- CDPDSS-2145: Applying column tag filter and typing the same search result in search field gives ‘no results’ found
- Problem: On searching the asset which is in the results for a column tag filter, the search results are shown as 'No results found'.
- CDPDSS-2236: Issues noticed on data-catalog while performing scale testing
- Problem:
- Search page consistently throws 504 gateway time-out error, however on applying global search field or search filters, results are loaded.
- Hive Column profiler goes to “undetermined” state for both ODP or scheduled jobs . To fix this, Profiler must be stopped and restarted. Happens due to Profiler not completing the read events from eventstore in AWS S3 within the time-out limit.(intermittent)
- On Asset Details page, when ODP is triggered, on successful completion of jobs and refresh of page, the ODP widget is not displayed and user needs to refresh a couple of times to load it on UI (reproducible). This ODP trigger job is running in parallel to running scheduled jobs.
- Assets cannot be added to dataset as the API fails with gateway timeout (intermittent)
- CDPDSS-2488: In MGMT service of DH CM, the telemetry command is failing bringing the master node down
-
When a Cloudera Data Hub cluster is created, the management service of Cloudera Manager becomes unavailable because the telemetry command fails on Hive. Because of this, the Master node of Cloudera Data Hub goes down causing a Node Failure error on the cluster.
In Cloudera Data Catalog, due to the failure of the master node of Cloudera Data Hub, the Profiler and the On-Demand Profiler page in the Asset Details page does not load.
- CDPDSS-3364: Not able to delete the profilers from DSS App when underlying compute cluster deleted
- When deleting the compute cluster in Compute Cluster enabled environments profilers cannot be deleted. The following error is displayed: Failed to delete profilers with error 401.
- CDPDSS-3403: Altered ICEBERG table is not available in Compute Cluster enable environment
- Iceberg tables modified or renamed with the
ALTER
operation appears with both old and new name, however, neither entry can be accessed. Apache Atlas synchronizes the table with the altered name with its full qualified name instead of its name shown in the Search results. - CDPDSS-3412: Profiling is not working on migrated ICEBERG table
- Hive tables migrated to Iceberg with the
ALTER TABLE
statement cannot be profiled in Compute Cluster enabled environments. Cluster Sensitivity Profilers and Hive Column Profilers will not be able to apply tags to the values. - CDPDSS-3401: Profiler is shown as running in DSS App on deleting default compute cluster
- In a Compute Cluster enabled environment, after deleting the default Compute Cluster profiler are displayed as still running in the Dashboard. However, profiler jobs will fail even if the default Compute Cluster is recreated.
- CDPDSS-3396: Occasionally getting Bad Gateway(Error Code - 502) error in DC UI
- In a Compute Cluster enabled environment, occasionally the messages Error Occurred while processing: Bad Gateway are displayed in addition to the Asset Details Overview tab failing to display any data.
- CDPDSS-3323: Profiler Jobs are running forever if the primary Kubernetes pod is shut down
- If the primary Kubernetes pod is forced shut down, or shuts down because of an error, the Kubernetes scheduler requests new resources for new pods. However, the original pod shut down still appears as a running job in Profiler / Jobs. This job cannot be deleted, as it is falsely marked running.
- CDPDSS-3467: [Search Page] Incorrect results are being displayed after navigation
- The total number of results in the Search page may change when changing the number of results displayed for a page after navigating into Asset Details for a result and navigating back to the overview of results. This is due to an incorrect response from Apache Atlas. The total number of results in the Search page may also change when using the owner filter and then clearing the applied filters, displaying all available assets.
- CDPDSS-3387: Profiling of struct or composite data type are not supported in Cloudera Data Catalog
- Profiling of struct or composite data type are not supported in Cloudera Data Catalog. When profiling tables with such data. the profiling job will be stuck in the Undetermined status.
- CDPDSS-3561: The number of columns is displayed incorrectly when the DB or table name is too long
- In Asset Details, the number of columns (# of Columns) is displayed as zero when the database or table name is too long.
- CDPDSS-3569: [Search Page ] Filters are retained when a filter is selected and then a different data lake is selected
- When switching data lakes on the Search page, the previously selected search filters are not cleared. This can result in an error message.