mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-06-27 12:05:03 +00:00
ci: [SYCL] ggml-ci Use main GPU and enable sysman (#12547)
This commit is contained in:
@ -52,7 +52,10 @@ if [ ! -z ${GG_BUILD_SYCL} ]; then
|
|||||||
echo "source /opt/intel/oneapi/setvars.sh"
|
echo "source /opt/intel/oneapi/setvars.sh"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
# Use only main GPU
|
||||||
|
export ONEAPI_DEVICE_SELECTOR="level_zero:0"
|
||||||
|
# Enable sysman for correct memory reporting
|
||||||
|
export ZES_ENABLE_SYSMAN=1
|
||||||
CMAKE_EXTRA="${CMAKE_EXTRA} -DGGML_SYCL=1 -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx -DGGML_SYCL_F16=ON"
|
CMAKE_EXTRA="${CMAKE_EXTRA} -DGGML_SYCL=1 -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx -DGGML_SYCL_F16=ON"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user