October 10, 2024
Release notes and fixed issues for version 2.0.46-b210.
New Features / Improvements
- Model Hub: Model Hub is now a fully supported feature. Model Hub is a catalog of
top-performing models LLM and generative AI models. You can now easily import the models
listed in the Model Hub into the Cloudera AI Registry and then deploy it using the Cloudera AI Inference service.
For more information, see Using Model Hub.
- Cloudera AI Inference service Enhancements:
- Added support for NVIDIA's NIM profiles requiring for the L40S GPU models.
- Made auto-scale configuration which is rendered in UI during the creation of model endpoint user-friendly. (DSE-38845)
- Optimized Cloudera AI UI service to become more responsive.
- User actionable error messages are now rendered in Cloudera AI service UI.
For more information, see Using Cloudera AI Inference service.
Fixed Issues
- Addressed scaling issues with web services to support high active user concurrency (DSE-39597).
- CVE fixes - This release includes numerous security fixes for critical and high Common Vulnerability and Exposures (CVE).
- Fixed CORS issue to ensure that DELETE/PATCH V1 API can be used from within a workbench. (DSE-39357)
- Made the NGC service key used to download Nvidia’s optimized models more restrictive. (DSE-39475)
- Previously, users were unable to copy the model-id from Cloudera AI UI. This issue is now resolved. (DSE-38889)
- Authorization issues related to the listing of Cloudera AI applications have been addressed. (DSE-39386)
- Fixed an issue to ensure that instance type validation is correctly carried out during the creation of a new model endpoint. (DSE-39634)
- Added required validation rules for the creation of a new model endpoint. (DSE-38412)
- Addressed an issue around empty model list during navigation from registry models to deployment of models. (DSE-39634)