mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-06-26 19:55:04 +00:00
40 lines
828 B
CMake
40 lines
828 B
CMake
# dependencies
|
|
|
|
find_package(Threads REQUIRED)
|
|
|
|
# third-party
|
|
|
|
# ...
|
|
|
|
# flags
|
|
|
|
llama_add_compile_flags()
|
|
|
|
# tools
|
|
|
|
if (EMSCRIPTEN)
|
|
else()
|
|
add_subdirectory(batched-bench)
|
|
add_subdirectory(gguf-split)
|
|
add_subdirectory(imatrix)
|
|
add_subdirectory(llama-bench)
|
|
add_subdirectory(main)
|
|
add_subdirectory(perplexity)
|
|
add_subdirectory(quantize)
|
|
if (LLAMA_BUILD_SERVER)
|
|
add_subdirectory(server)
|
|
endif()
|
|
add_subdirectory(run)
|
|
add_subdirectory(tokenize)
|
|
add_subdirectory(tts)
|
|
add_subdirectory(mtmd)
|
|
if (GGML_RPC)
|
|
add_subdirectory(rpc)
|
|
endif()
|
|
if (NOT GGML_BACKEND_DL)
|
|
# these examples use the backends directly and cannot be built with dynamic loading
|
|
add_subdirectory(cvector-generator)
|
|
add_subdirectory(export-lora)
|
|
endif()
|
|
endif()
|