diff --git a/serving/docs/lmi/user_guides/release_notes.md b/serving/docs/lmi/user_guides/release_notes.md index 9229b467a..60d9b025e 100644 --- a/serving/docs/lmi/user_guides/release_notes.md +++ b/serving/docs/lmi/user_guides/release_notes.md @@ -1,4 +1,4 @@ -# LMI V10 containers release +# LMI V11 containers release This document will contain the latest releases of our LMI containers. For details on any other previous releases, please refer our [github release page](https://github.com/deepjavalibrary/djl-serving/releases) @@ -45,4 +45,4 @@ Check out our latest [Large Model Inference Containers](https://github.com/aws/d ### Known Issues * Running Gemma and Phi models with TensorRT-LLM is only viable currently at TP=1 because of an issue in TensorRT-LLM where one engine is built even when TP > 1. -* When using LMI-dist, in the rare case that the machine has a broken cuda driver, it causes hanging. In that case, set LMI_USE_VLLM_GPU_P2P_CHECK=1 to prompt LMI to use a fallback option compatible with the broken cuda driver. \ No newline at end of file +* When using LMI-dist, in the rare case that the machine has a broken cuda driver, it causes hanging. In that case, set LMI_USE_VLLM_GPU_P2P_CHECK=1 to prompt LMI to use a fallback option compatible with the broken cuda driver.