From a29dc492ac72406fea12a66f78be173a1082b911 Mon Sep 17 00:00:00 2001 From: derekxu Date: Mon, 6 Jan 2025 18:08:28 -0800 Subject: [PATCH] Add support to export XNNPACK based static_llama Differential Revision: D67867190 Pull Request resolved: https://github.com/pytorch/executorch/pull/7535 --- examples/models/llama/export_llama_lib.py | 9 ++++++++- 1 file changed, 8 insertions(+), 1 deletion(-) diff --git a/examples/models/llama/export_llama_lib.py b/examples/models/llama/export_llama_lib.py index 4e004e773f..0b7064c4dd 100644 --- a/examples/models/llama/export_llama_lib.py +++ b/examples/models/llama/export_llama_lib.py @@ -79,7 +79,14 @@ verbosity_setting = None -EXECUTORCH_DEFINED_MODELS = ["stories110m", "llama2", "llama3", "llama3_1", "llama3_2"] +EXECUTORCH_DEFINED_MODELS = [ + "stories110m", + "llama2", + "llama3", + "llama3_1", + "llama3_2", + "static_llama", +] TORCHTUNE_DEFINED_MODELS = ["llama3_2_vision"]