From 2360058617e490028dbd97dd58547d793351c443 Mon Sep 17 00:00:00 2001 From: Karol Blaszczak Date: Fri, 13 Dec 2024 11:03:23 +0100 Subject: [PATCH] [DOCS] release notes 24.6 pass-1 (#28057) --- .../about-openvino/release-notes-openvino.rst | 59 +++++++++++++++++++ 1 file changed, 59 insertions(+) diff --git a/docs/articles_en/about-openvino/release-notes-openvino.rst b/docs/articles_en/about-openvino/release-notes-openvino.rst index 9bfad4193d3987..cda43a5bfd0acc 100644 --- a/docs/articles_en/about-openvino/release-notes-openvino.rst +++ b/docs/articles_en/about-openvino/release-notes-openvino.rst @@ -23,6 +23,65 @@ OpenVINO Release Notes +What's new ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ + +* OpenVINO 2024.6 LTS release includes updates for enhanced stability and improved LLM performance. +* Introduced support for Intel® Arc™ B-Series Graphics (formerly known as Battlemage) +* Memory optimizations implemented to improve the inference time memory and LLM performance on NPUs. +* Improved LLM performance with GenAI API optimizations and bug fixes. + + + +OpenVINO™ Runtime ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ + +CPU Device Plugin +----------------------------------------------------------------------------------------------- + +* KV cache now uses asymmetric U8 as the default precision, reducing memory stress for LLMs and + increasing their performance. This option can be controlled by model meta data. +* Quality and accuracy has been improved for selected models with several bug fixes. + +GPU Device Plugin +----------------------------------------------------------------------------------------------- + +* Device memory copy optimizations have been introduced for inference with **Intel® Arc™ B-Series + Graphics** (formerly known as Battlemage). Since it does not utilize L2 cache for copying memory + between the device and host, a dedicated `copy` operation is used, if inputs or results are + not expected in the device memory. +* ChatGLM4 inference on GPU has been optimized. + +NPU Device Plugin +----------------------------------------------------------------------------------------------- + +* LLM performance and inference time has been improved with memory optimizations. + + + + + +OpenVINO.GenAI ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ + +* The encrypted_model_causal_lm sample is now available, showing how to decrypt a model. + + + + +Other Changes and Known Issues ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ + +Jupyter Notebooks +----------------------------- + +* `Visual-language assistant with GLM-Edge-V and OpenVINO `__ +* `Local AI and OpenVINO `__ +* `Multimodal understanding and generation with Janus and OpenVINO `__ + + + +