Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add suffixes to some cmake flags. #3485

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
2 changes: 1 addition & 1 deletion .ci/scripts/build_llama_android.sh
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ install_executorch_and_backend_lib() {
-DCMAKE_BUILD_TYPE=Release \
-DEXECUTORCH_BUILD_EXTENSION_MODULE=ON \
-DEXECUTORCH_BUILD_EXTENSION_DATA_LOADER=ON \
-DEXECUTORCH_BUILD_XNNPACK=ON \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON \
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
-DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON \
-DXNNPACK_ENABLE_ARM_BF16=OFF \
Expand Down
2 changes: 1 addition & 1 deletion .ci/scripts/test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@ build_cmake_xnn_executor_runner() {
&& mkdir ${CMAKE_OUTPUT_DIR} \
&& cd ${CMAKE_OUTPUT_DIR} \
&& retry cmake -DCMAKE_BUILD_TYPE=Release \
-DEXECUTORCH_BUILD_XNNPACK=ON \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON \
-DCMAKE_PREFIX_PATH="$CMAKE_PREFIX_PATH" \
-DPYTHON_EXECUTABLE="$PYTHON_EXECUTABLE" ..)

Expand Down
4 changes: 2 additions & 2 deletions .ci/scripts/test_llama.sh
Original file line number Diff line number Diff line change
Expand Up @@ -76,7 +76,7 @@ cmake_install_executorch_libraries() {
-DEXECUTORCH_BUILD_KERNELS_CUSTOM="$CUSTOM" \
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
-DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON \
-DEXECUTORCH_BUILD_XNNPACK="$XNNPACK" \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK="$XNNPACK" \
-DPYTHON_EXECUTABLE="$PYTHON_EXECUTABLE" \
-Bcmake-out .
cmake --build cmake-out -j9 --target install --config Debug
Expand All @@ -90,7 +90,7 @@ cmake_build_llama_runner() {
-DCMAKE_BUILD_TYPE=Debug \
-DEXECUTORCH_BUILD_KERNELS_CUSTOM="$CUSTOM" \
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
-DEXECUTORCH_BUILD_XNNPACK="$XNNPACK" \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK="$XNNPACK" \
-DPYTHON_EXECUTABLE="$PYTHON_EXECUTABLE" \
-Bcmake-out/${dir} \
${dir}
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/_unittest.yml
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ jobs:
# Setup MacOS dependencies as there is no Docker support on MacOS atm
PYTHON_EXECUTABLE=python \
EXECUTORCH_BUILD_PYBIND=ON \
EXECUTORCH_BUILD_XNNPACK=ON \
EXECUTORCH_BUILD_BACKEND_XNNPACK=ON \
.ci/scripts/setup-linux.sh "${BUILD_TOOL}"
# Run pytest with coverage
Expand All @@ -64,7 +64,7 @@ jobs:
# Setup MacOS dependencies as there is no Docker support on MacOS atm
PYTHON_EXECUTABLE=python \
EXECUTORCH_BUILD_PYBIND=ON \
CMAKE_ARGS="-DEXECUTORCH_BUILD_COREML=ON -DEXECUTORCH_BUILD_MPS=ON -DEXECUTORCH_BUILD_XNNPACK=ON" \
CMAKE_ARGS="-DEXECUTORCH_BUILD_BACKEND_COREML=ON -DEXECUTORCH_BUILD_BACKEND_MPS=ON -DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON" \
${CONDA_RUN} --no-capture-output \
.ci/scripts/setup-macos.sh "${BUILD_TOOL}"
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/pull.yml
Original file line number Diff line number Diff line change
Expand Up @@ -232,7 +232,7 @@ jobs:
# build module for executorch.extension.pybindings.portable_lib
BUILD_TOOL=${{ matrix.build-tool }}
PYTHON_EXECUTABLE=python \
EXECUTORCH_BUILD_XNNPACK=ON \
EXECUTORCH_BUILD_BACKEND_XNNPACK=ON \
EXECUTORCH_BUILD_PYBIND=ON \
bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
Expand Down
54 changes: 27 additions & 27 deletions CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -14,8 +14,8 @@
# it again until you pull from the upstream repo again.
#
# NOTE: Build options can be configured by passing arguments to cmake. For
# example, to enable the EXECUTORCH_BUILD_XNNPACK option, change the cmake
# command to 'cmake -DEXECUTORCH_BUILD_XNNPACK=ON ..'.
# example, to enable the EXECUTORCH_BUILD_BACKEND_XNNPACK option, change the
# cmake command to 'cmake -DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON ..'.
#[[
(rm -rf cmake-out \
&& mkdir cmake-out \
Expand Down Expand Up @@ -143,17 +143,19 @@ set(CMAKE_CXX_FLAGS_DEBUG "-O0 -g")

option(EXECUTORCH_BUILD_ANDROID_JNI "Build Android JNI" OFF)

option(EXECUTORCH_BUILD_ARM_BAREMETAL
option(EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL
"Build the Arm Baremetal flow for Cortex-M and Ethos-U" OFF
)

option(EXECUTORCH_BUILD_COREML "Build the Core ML backend" OFF)
option(EXECUTORCH_BUILD_BACKEND_COREML "Build the Core ML backend" OFF)

option(EXECUTORCH_BUILD_KERNELS_CUSTOM "Build the custom kernels" OFF)
option(EXECUTORCH_BUILD_BACKEND_MPS "Build the MPS backend" OFF)

option(EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT "Build the custom ops lib for AOT"
OFF
)
option(EXECUTORCH_BUILD_BACKEND_QNN "Build the Qualcomm backend" OFF)

option(EXECUTORCH_BUILD_BACKEND_XNNPACK "Build the XNNPACK backend" OFF)

option(EXECUTORCH_BUILD_BACKEND_VULKAN "Build the Vulkan backend" OFF)

option(EXECUTORCH_BUILD_EXTENSION_DATA_LOADER "Build the Data Loader extension"
OFF
Expand All @@ -167,38 +169,36 @@ option(EXECUTORCH_BUILD_EXTENSION_RUNNER_UTIL "Build the Runner Util extension"

option(EXECUTORCH_BUILD_GTESTS "Build googletest based test binaries" OFF)

option(EXECUTORCH_BUILD_MPS "Build the MPS backend" OFF)

option(EXECUTORCH_BUILD_PYBIND "Build the Python Bindings" OFF)
option(EXECUTORCH_BUILD_KERNELS_CUSTOM "Build the custom kernels" OFF)

option(EXECUTORCH_BUILD_QNN "Build the Qualcomm backend" OFF)
option(EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT "Build the custom ops lib for AOT"
OFF
)

option(EXECUTORCH_BUILD_KERNELS_OPTIMIZED "Build the optimized kernels" OFF)

option(EXECUTORCH_BUILD_KERNELS_QUANTIZED "Build the quantized kernels" OFF)

option(EXECUTORCH_BUILD_PYBIND "Build the Python Bindings" OFF)

option(EXECUTORCH_BUILD_SDK "Build the ExecuTorch SDK")

option(EXECUTORCH_BUILD_SIZE_TEST "Build the size test" OFF)

option(EXECUTORCH_BUILD_XNNPACK "Build the XNNPACK backend" OFF)

option(EXECUTORCH_BUILD_VULKAN "Build the Vulkan backend" OFF)

#
# pthreadpool: build pthreadpool library. Disable on unsupported platforms
#
cmake_dependent_option(
EXECUTORCH_BUILD_PTHREADPOOL "Build pthreadpool library." ON
"NOT EXECUTORCH_BUILD_ARM_BAREMETAL" OFF
"NOT EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL" OFF
)

#
# cpuinfo: build cpuinfo library. Disable on unsupported platforms
#
cmake_dependent_option(
EXECUTORCH_BUILD_CPUINFO "Build cpuinfo library." ON
"NOT EXECUTORCH_BUILD_ARM_BAREMETAL" OFF
"NOT EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL" OFF
)

if(EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT)
Expand Down Expand Up @@ -542,27 +542,27 @@ if(EXECUTORCH_BUILD_EXTENSION_RUNNER_UTIL)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/extension/runner_util)
endif()

if(EXECUTORCH_BUILD_XNNPACK)
if(EXECUTORCH_BUILD_BACKEND_XNNPACK)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/xnnpack)
endif()

if(EXECUTORCH_BUILD_VULKAN)
if(EXECUTORCH_BUILD_BACKEND_VULKAN)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/vulkan)
endif()

if(EXECUTORCH_BUILD_QNN)
if(EXECUTORCH_BUILD_BACKEND_QNN)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/qualcomm)
endif()

if(EXECUTORCH_BUILD_ARM_BAREMETAL)
if(EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/arm)
endif()

if(EXECUTORCH_BUILD_MPS)
if(EXECUTORCH_BUILD_BACKEND_MPS)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/apple/mps)
endif()

if(EXECUTORCH_BUILD_COREML)
if(EXECUTORCH_BUILD_BACKEND_COREML)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/apple/coreml)
endif()

Expand Down Expand Up @@ -594,15 +594,15 @@ if(EXECUTORCH_BUILD_PYBIND)
torch
)

if(EXECUTORCH_BUILD_COREML)
if(EXECUTORCH_BUILD_BACKEND_COREML)
list(APPEND _dep_libs coremldelegate)
endif()

if(EXECUTORCH_BUILD_MPS)
if(EXECUTORCH_BUILD_BACKEND_MPS)
list(APPEND _dep_libs mpsdelegate)
endif()

if(EXECUTORCH_BUILD_XNNPACK)
if(EXECUTORCH_BUILD_BACKEND_XNNPACK)
# need to explicitly specify XNNPACK here otherwise uses XNNPACK symbols
# from libtorch_cpu
list(APPEND _dep_libs xnnpack_backend XNNPACK)
Expand Down
2 changes: 1 addition & 1 deletion backends/apple/coreml/scripts/build_tests.sh
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ cmake "$EXECUTORCH_ROOT_PATH" -B"$CMAKE_EXECUTORCH_BUILD_DIR_PATH" \
-DDEPLOYMENT_TARGET=13.0 \
-DFLATC_EXECUTABLE="$(which flatc)" \
-DEXECUTORCH_BUILD_EXECUTOR_RUNNER=OFF \
-DEXECUTORCH_BUILD_XNNPACK=OFF \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK=OFF \
-DEXECUTORCH_BUILD_GFLAGS=OFF

cmake --build "$CMAKE_EXECUTORCH_BUILD_DIR_PATH" -j9 -t executorch
Expand Down
4 changes: 2 additions & 2 deletions backends/qualcomm/scripts/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@ if [ "$BUILD_AARCH64" = true ]; then
# If we build debug type, we need to change flatcc to flatcc_d
cmake .. \
-DCMAKE_INSTALL_PREFIX=$BUILD_ROOT \
-DEXECUTORCH_BUILD_QNN=ON \
-DEXECUTORCH_BUILD_BACKEND_QNN=ON \
-DEXECUTORCH_BUILD_SDK=ON \
-DEXECUTORCH_ENABLE_EVENT_TRACER=ON \
-DQNN_SDK_ROOT=$QNN_SDK_ROOT \
Expand Down Expand Up @@ -108,7 +108,7 @@ if [ "$BUILD_X86_64" = true ]; then
cmake \
-DCMAKE_BUILD_TYPE=RelWithDebInfo \
-DQNN_SDK_ROOT=${QNN_SDK_ROOT} \
-DEXECUTORCH_BUILD_QNN=ON \
-DEXECUTORCH_BUILD_BACKEND_QNN=ON \
-DPYTHON_EXECUTABLE=$PYTHON_EXECUTABLE \
-DBUCK2=$BUCK2 \
-S $PRJ_ROOT \
Expand Down
4 changes: 2 additions & 2 deletions backends/qualcomm/setup.md
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@ Make sure `buck2` is under a directory in `PATH`.
cd $EXECUTORCH_ROOT
mkdir build_x86_64
cd build_x86_64
cmake .. -DEXECUTORCH_BUILD_QNN=ON -DQNN_SDK_ROOT=${QNN_SDK_ROOT}
cmake .. -DEXECUTORCH_BUILD_BACKEND_QNN=ON -DQNN_SDK_ROOT=${QNN_SDK_ROOT}
cmake --build . -t "PyQnnManagerAdaptor" "PyQnnWrapperAdaptor" -j8

# install Python APIs to correct import path
Expand All @@ -94,7 +94,7 @@ cd build_android
# build executorch & qnn_executorch_backend
cmake .. \
-DCMAKE_INSTALL_PREFIX=$PWD \
-DEXECUTORCH_BUILD_QNN=ON \
-DEXECUTORCH_BUILD_BACKEND_QNN=ON \
-DQNN_SDK_ROOT=$QNN_SDK_ROOT \
-DCMAKE_TOOLCHAIN_FILE=$ANDROID_NDK/build/cmake/android.toolchain.cmake \
-DANDROID_ABI='arm64-v8a' \
Expand Down
3 changes: 2 additions & 1 deletion backends/vulkan/CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,8 @@
# ~~~
# It should also be cmake-lint clean.
#
# The targets in this file will be built if EXECUTORCH_BUILD_VULKAN is ON
# The targets in this file will be built if EXECUTORCH_BUILD_BACKEND_VULKAN is
# ON

cmake_minimum_required(VERSION 3.19)

Expand Down
4 changes: 2 additions & 2 deletions backends/vulkan/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -134,7 +134,7 @@ and test on a local Android device. Android devices have built in support for
Vulkan, and the Android NDK ships with a GLSL compiler, which is needed to
compile the Vulkan Compute Library's GLSL compute shaders.

The Vulkan Delegate libraries can be built by setting `-DEXECUTORCH_BUILD_VULKAN=ON`
The Vulkan Delegate libraries can be built by setting `-DEXECUTORCH_BUILD_BACKEND_VULKAN=ON`
when building with CMake.

First, make sure that you have the Android NDK installed - Android NDK r25c is
Expand All @@ -161,7 +161,7 @@ Delegate:
pp cmake . -DCMAKE_INSTALL_PREFIX=cmake-android-out \
-DCMAKE_TOOLCHAIN_FILE=$ANDROID_NDK/build/cmake/android.toolchain.cmake \
-DANDROID_ABI=$ANDROID_ABI \
-DEXECUTORCH_BUILD_VULKAN=ON \
-DEXECUTORCH_BUILD_BACKEND_VULKAN=ON \
-DPYTHON_EXECUTABLE=python \
-Bcmake-android-out && \
cmake --build cmake-android-out -j16 --target install)
Expand Down
3 changes: 2 additions & 1 deletion backends/vulkan/cmake/ShaderLibrary.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,8 @@
# ~~~
# It should also be cmake-lint clean.
#
# The targets in this file will be built if EXECUTORCH_BUILD_VULKAN is ON
# The targets in this file will be built if EXECUTORCH_BUILD_BACKEND_VULKAN is
# ON

if(ANDROID)
if(NOT ANDROID_NDK)
Expand Down
4 changes: 2 additions & 2 deletions backends/vulkan/docs/android_demo.md
Original file line number Diff line number Diff line change
Expand Up @@ -96,7 +96,7 @@ binary using the Android NDK toolchain.
-DANDROID_ABI=$ANDROID_ABI \
-DEXECUTORCH_BUILD_EXTENSION_MODULE=ON \
-DEXECUTORCH_BUILD_EXTENSION_DATA_LOADER=ON \
-DEXECUTORCH_BUILD_VULKAN=ON \
-DEXECUTORCH_BUILD_BACKEND_VULKAN=ON \
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
-DPYTHON_EXECUTABLE=python \
-Bcmake-android-out && \
Expand Down Expand Up @@ -140,7 +140,7 @@ the Vulkan backend is built when building and installing ExecuTorch libraries:

```shell
# Run from executorch root directory. You can also edit this in a code editor
sed -i 's/-DEXECUTORCH_BUILD_XNNPACK=ON/-DEXECUTORCH_BUILD_XNNPACK=ON -DEXECUTORCH_BUILD_VULKAN=ON/g' examples/demo-apps/android/LlamaDemo/setup.sh
sed -i 's/-DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON/-DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON -DEXECUTORCH_BUILD_BACKEND_VULKAN=ON/g' examples/demo-apps/android/LlamaDemo/setup.sh
```

Then, Follow the instructions at [**Setting up the ExecuTorch LLaMA Android Demo App**](./llm/llama-demo-android.md)
Expand Down
2 changes: 1 addition & 1 deletion backends/xnnpack/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -106,7 +106,7 @@ cmake \
-DCMAKE_INSTALL_PREFIX=cmake-out \
-DCMAKE_BUILD_TYPE=Release \
-DEXECUTORCH_BUILD_EXTENSION_MODULE=ON \
-DEXECUTORCH_BUILD_XNNPACK=ON \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON \
-DEXECUTORCH_BUILD_EXTENSION_DATA_LOADER=ON \
-DEXECUTORCH_ENABLE_LOGGING=ON \
-DPYTHON_EXECUTABLE=python \
Expand Down
54 changes: 28 additions & 26 deletions build/Utils.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -43,15 +43,32 @@ function(executorch_print_configuration_summary)
message(STATUS " EXECUTORCH_BUILD_ANDROID_JNI : "
"${EXECUTORCH_BUILD_ANDROID_JNI}"
)
message(STATUS " EXECUTORCH_BUILD_ARM_BAREMETAL : "
"${EXECUTORCH_BUILD_ARM_BAREMETAL}"
message(STATUS " EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL : "
"${EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL}"
)
message(
STATUS
" EXECUTORCH_BUILD_COREML : ${EXECUTORCH_BUILD_COREML}"
" EXECUTORCH_BUILD_BACKEND_COREML : ${EXECUTORCH_BUILD_BACKEND_COREML}"
)
message(STATUS " EXECUTORCH_BUILD_KERNELS_CUSTOM : "
"${EXECUTORCH_BUILD_KERNELS_CUSTOM}"
message(
STATUS
" EXECUTORCH_BUILD_BACKEND_MPS : ${EXECUTORCH_BUILD_BACKEND_MPS}"
)
message(
STATUS
" EXECUTORCH_BUILD_BACKEND_QNN : ${EXECUTORCH_BUILD_BACKEND_QNN}"
)
message(
STATUS
" EXECUTORCH_BUILD_BACKEND_XNNPACK : ${EXECUTORCH_BUILD_BACKEND_XNNPACK}"
)
message(
STATUS
" EXECUTORCH_BUILD_BACKEND_VULKAN : ${EXECUTORCH_BUILD_BACKEND_VULKAN}"
)
message(
STATUS
" EXECUTORCH_BUILD_CPUINFO : ${EXECUTORCH_BUILD_CPUINFO}"
)
message(STATUS " EXECUTORCH_BUILD_EXECUTOR_RUNNER : "
"${EXECUTORCH_BUILD_EXECUTOR_RUNNER}"
Expand Down Expand Up @@ -80,44 +97,29 @@ function(executorch_print_configuration_summary)
message(STATUS " EXECUTORCH_BUILD_HOST_TARGETS : "
"${EXECUTORCH_BUILD_HOST_TARGETS}"
)
message(
STATUS " EXECUTORCH_BUILD_MPS : ${EXECUTORCH_BUILD_MPS}"
)
message(
STATUS
" EXECUTORCH_BUILD_PYBIND : ${EXECUTORCH_BUILD_PYBIND}"
)
message(
STATUS " EXECUTORCH_BUILD_QNN : ${EXECUTORCH_BUILD_QNN}"
message(STATUS " EXECUTORCH_BUILD_KERNELS_CUSTOM : "
"${EXECUTORCH_BUILD_KERNELS_CUSTOM}"
)
message(STATUS " EXECUTORCH_BUILD_KERNELS_OPTIMIZED : "
"${EXECUTORCH_BUILD_KERNELS_OPTIMIZED}"
)
message(STATUS " EXECUTORCH_BUILD_KERNELS_QUANTIZED : "
"${EXECUTORCH_BUILD_KERNELS_QUANTIZED}"
)
message(
STATUS " EXECUTORCH_BUILD_SDK : ${EXECUTORCH_BUILD_SDK}"
)
message(
STATUS
" EXECUTORCH_BUILD_SIZE_TEST : ${EXECUTORCH_BUILD_SIZE_TEST}"
)
message(
STATUS
" EXECUTORCH_BUILD_XNNPACK : ${EXECUTORCH_BUILD_XNNPACK}"
" EXECUTORCH_BUILD_PTHREADPOOL : ${EXECUTORCH_BUILD_PTHREADPOOL}"
)
message(
STATUS
" EXECUTORCH_BUILD_VULKAN : ${EXECUTORCH_BUILD_VULKAN}"
" EXECUTORCH_BUILD_PYBIND : ${EXECUTORCH_BUILD_PYBIND}"
)
message(
STATUS
" EXECUTORCH_BUILD_PTHREADPOOL : ${EXECUTORCH_BUILD_PTHREADPOOL}"
STATUS " EXECUTORCH_BUILD_SDK : ${EXECUTORCH_BUILD_SDK}"
)
message(
STATUS
" EXECUTORCH_BUILD_CPUINFO : ${EXECUTORCH_BUILD_CPUINFO}"
" EXECUTORCH_BUILD_SIZE_TEST : ${EXECUTORCH_BUILD_SIZE_TEST}"
)

endfunction()
Expand Down