Cloudera Data Catalog 1.5.4 SP1
Review the features, fixes, and known issues in the Cloudera Data Catalog 1.5.4 Service Pack 1 release.
What's new in 1.5.4 SP1
This section lists major features and updates for the Cloudera Data Catalog service.
This release of Cloudera Data Catalog on Cloudera Private Cloud 1.5.4 SP1 includes the following features and
enhancements:
- Bug fixes
- Datasets can only be deleted by their creators following this release.
- You can hover over the Comment field for individual schema entries in Asset Details to preview longer comments without opening them.
- The security of Cloudera Data Catalog is improved. Now our services run
with the following security
context:
'containers[].securityContext.allowPrivilegeEscalation' to 'false'. 'containers[].securityContext.runAsUser' to an integer > 1000000. 'containers[].securityContext.readOnlyRootFilesystem' to 'true'
Fixed issues in 1.5.4 SP1
This section lists the issues that have been fixed since the previous version.
- CDPDSS-3395: Incorrect results are being displayed after navigation, when "owners" filter is selected
- In the Search page, returning to the list of results after opening the Asset Details of one the results and repeating the search with the same filter, other entities belonging to owners different than the selected no longer appear. Only entities belonging to the selected owners are displayed.
- CDPDSS-3081: Datasets: Tenant Schema is getting passed as workload username while impersonating a user on ATLAS and Ranger API calls
- Before the update, while making API calls to Apache Atlas and
Apache Ranger, there was a potential bug in code wherein a wrong value for
workloadUserName
was getting passed which caused APIs to perform an auth check with a wrong parameter, causing unwanted issues. - CDPDSS-3169: Asset Details Page - Optimized API to avoid 504 Error
- An Asset Details page with only a few assets (Hive table) was unable to open the Hive table assets which had around 2000 columns or more. The number of columns was causing the timeout. Optimizing the API calls related to Asset Details led to improved performance and avoiding potential 504 errors when opening assets.
- CDPDSS-3080: Deleted Entities in Atlas cannot be removed from Datasets
- Before the update, entities deleted in Apache Atlas could not be removed from Cloudera Data Catalog datasets.
- CDPDSS-3366: Asset Details is distorted with very long description
- After the update, the very long strings in the Description field are trimmed and no longer distort Asset Details by being displayed in full length.
- CDPDSS-3351: Unable to see comments on the Hive columns in Asset Details
- Full comments can be displayed in Asset Details while hovering over the Comment field compared to the earlier limited view.
- CDPDSS-2688: Hive Column On-demand Profiler is failing but the Asset Details page is not updated with 'Failed State'
- A failed Hive Column On-Demand Profiler job is correctly displayed as failed in .
- CDPDSS-3353: Profiler Configuration page is giving error, if browser's preferred language is not UK English
- The
en-gb
.
is correctly rendered even if the web browser's preferred language is not set to
- CDPDSS-2913: Asset Details view does not provide details of the asset - exception
- Entities without lineage information no longer display the get lineage request failed Error from Atlas Not Found : Received 404..
- CDPDSS-3381: The Ranger Audit Profiler is showing status as FINISHED/SUCCESS even though there is error with the profilers
- After the update, the Ranger Audit Profiler shows the correct status when the NameNode is stopped, and it is no longer stuck in FINISHED.
Known issues in 1.5.4 SP1
Learn about the known issues in Cloudera Data Catalog, the impact or changes to the functionality, and the workaround.
- CDPDSS-3057: Failed profiler job because of missing columns in log entries.
- When the logs to be profiled by the Ranger Audit Profiler have
missing columns, the profiling job fails. The ranger audit profiler pod log contains the
following message:Ranger Audit Profiler failed: Failed
to run Ranger Audit Profiler. Error: "['column1',
'column2', ...] not in
index".
column1
,column2
show the missing column entries in the logs.Each Ranger audit profiler job has 3 sub-jobs (identified by the same job ID prefix), any of which could fail because of the above reason. It is possible that one or even two sub-jobs pass till the above scenario is encountered.
- CDPDSS-3472: Profiler jobs are showing up twice when two data lakes are present
- Cloudera Data Catalog does not support more than one data lakes (or base clusters). Connecting two data lakes will result in all profiler jobs showing as duplicated in the Profilers page.