libs : rename libcommon -> libllama-common (#21936)

* cmake : allow libcommon to be shared

* cmake : rename libcommon to libllama-common

* cont : set -fPIC for httplib

* cont : export all symbols

* cont : fix build_info exports

* libs : add libllama-common-base

* log : add common_log_get_verbosity_thold()
This commit is contained in:
Georgi Gerganov
2026-04-17 11:11:46 +03:00
committed by GitHub
parent fcc7508759
commit 6990e2f1f7
58 changed files with 184 additions and 108 deletions
+5 -1
View File
@@ -225,7 +225,7 @@ foreach(FILE_PATH ${EXTRA_LICENSES})
endforeach() endforeach()
if (LLAMA_BUILD_COMMON) if (LLAMA_BUILD_COMMON)
license_generate(common) license_generate(llama-common)
endif() endif()
# #
@@ -249,6 +249,10 @@ set_target_properties(llama
install(TARGETS llama LIBRARY PUBLIC_HEADER) install(TARGETS llama LIBRARY PUBLIC_HEADER)
if (LLAMA_BUILD_COMMON)
install(TARGETS llama-common LIBRARY)
endif()
configure_package_config_file( configure_package_config_file(
${CMAKE_CURRENT_SOURCE_DIR}/cmake/llama-config.cmake.in ${CMAKE_CURRENT_SOURCE_DIR}/cmake/llama-config.cmake.in
${CMAKE_CURRENT_BINARY_DIR}/llama-config.cmake ${CMAKE_CURRENT_BINARY_DIR}/llama-config.cmake
+27 -10
View File
@@ -1,9 +1,11 @@
# common
find_package(Threads REQUIRED) find_package(Threads REQUIRED)
llama_add_compile_flags() llama_add_compile_flags()
#
# llama-common-base
#
# Build info header # Build info header
if(EXISTS "${PROJECT_SOURCE_DIR}/.git") if(EXISTS "${PROJECT_SOURCE_DIR}/.git")
@@ -33,17 +35,25 @@ endif()
set(TEMPLATE_FILE "${CMAKE_CURRENT_SOURCE_DIR}/build-info.cpp.in") set(TEMPLATE_FILE "${CMAKE_CURRENT_SOURCE_DIR}/build-info.cpp.in")
set(OUTPUT_FILE "${CMAKE_CURRENT_BINARY_DIR}/build-info.cpp") set(OUTPUT_FILE "${CMAKE_CURRENT_BINARY_DIR}/build-info.cpp")
configure_file(${TEMPLATE_FILE} ${OUTPUT_FILE}) configure_file(${TEMPLATE_FILE} ${OUTPUT_FILE})
set(TARGET build_info) set(TARGET llama-common-base)
add_library(${TARGET} OBJECT ${OUTPUT_FILE}) add_library(${TARGET} STATIC ${OUTPUT_FILE})
target_include_directories(${TARGET} PUBLIC .)
if (BUILD_SHARED_LIBS) if (BUILD_SHARED_LIBS)
set_target_properties(${TARGET} PROPERTIES POSITION_INDEPENDENT_CODE ON) set_target_properties(${TARGET} PROPERTIES POSITION_INDEPENDENT_CODE ON)
endif() endif()
set(TARGET common) #
# llama-common
#
add_library(${TARGET} STATIC set(TARGET llama-common)
add_library(${TARGET}
arg.cpp arg.cpp
arg.h arg.h
base64.hpp base64.hpp
@@ -106,17 +116,24 @@ add_library(${TARGET} STATIC
jinja/caps.h jinja/caps.h
) )
set_target_properties(${TARGET} PROPERTIES
VERSION ${LLAMA_INSTALL_VERSION}
SOVERSION 0
MACHO_CURRENT_VERSION 0 # keep macOS linker from seeing oversized version number
)
target_include_directories(${TARGET} PUBLIC . ../vendor) target_include_directories(${TARGET} PUBLIC . ../vendor)
target_compile_features (${TARGET} PUBLIC cxx_std_17) target_compile_features (${TARGET} PUBLIC cxx_std_17)
if (BUILD_SHARED_LIBS) if (BUILD_SHARED_LIBS)
set_target_properties(${TARGET} PROPERTIES POSITION_INDEPENDENT_CODE ON) set_target_properties(${TARGET} PROPERTIES POSITION_INDEPENDENT_CODE ON)
# TODO: make fine-grained exports in the future
set_target_properties(${TARGET} PROPERTIES WINDOWS_EXPORT_ALL_SYMBOLS ON)
endif() endif()
target_link_libraries(${TARGET} PRIVATE target_link_libraries(${TARGET} PUBLIC llama-common-base)
build_info target_link_libraries(${TARGET} PRIVATE cpp-httplib)
cpp-httplib
)
if (LLAMA_LLGUIDANCE) if (LLAMA_LLGUIDANCE)
include(ExternalProject) include(ExternalProject)
+3 -2
View File
@@ -1,5 +1,6 @@
#include "arg.h" #include "arg.h"
#include "build-info.h"
#include "chat.h" #include "chat.h"
#include "common.h" #include "common.h"
#include "download.h" #include "download.h"
@@ -1044,8 +1045,8 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
{"--version"}, {"--version"},
"show version and build info", "show version and build info",
[](common_params &) { [](common_params &) {
fprintf(stderr, "version: %d (%s)\n", LLAMA_BUILD_NUMBER, LLAMA_COMMIT); fprintf(stderr, "version: %d (%s)\n", llama_build_number(), llama_commit());
fprintf(stderr, "built with %s for %s\n", LLAMA_COMPILER, LLAMA_BUILD_TARGET); fprintf(stderr, "built with %s for %s\n", llama_compiler(), llama_build_target());
exit(0); exit(0);
} }
)); ));
+34 -3
View File
@@ -1,4 +1,35 @@
#include "build-info.h"
#include <cstdio>
#include <string>
int LLAMA_BUILD_NUMBER = @LLAMA_BUILD_NUMBER@; int LLAMA_BUILD_NUMBER = @LLAMA_BUILD_NUMBER@;
char const *LLAMA_COMMIT = "@LLAMA_BUILD_COMMIT@"; char const * LLAMA_COMMIT = "@LLAMA_BUILD_COMMIT@";
char const *LLAMA_COMPILER = "@BUILD_COMPILER@"; char const * LLAMA_COMPILER = "@BUILD_COMPILER@";
char const *LLAMA_BUILD_TARGET = "@BUILD_TARGET@"; char const * LLAMA_BUILD_TARGET = "@BUILD_TARGET@";
int llama_build_number(void) {
return LLAMA_BUILD_NUMBER;
}
const char * llama_commit(void) {
return LLAMA_COMMIT;
}
const char * llama_compiler(void) {
return LLAMA_COMPILER;
}
const char * llama_build_target(void) {
return LLAMA_BUILD_TARGET;
}
const char * llama_build_info(void) {
static std::string s = "b" + std::to_string(LLAMA_BUILD_NUMBER) + "-" + LLAMA_COMMIT;
return s.c_str();
}
void llama_print_build_info(void) {
fprintf(stderr, "%s: build = %d (%s)\n", __func__, llama_build_number(), llama_commit());
fprintf(stderr, "%s: built with %s for %s\n", __func__, llama_compiler(), llama_build_target());
}
+11
View File
@@ -0,0 +1,11 @@
#pragma once
int llama_build_number(void);
const char * llama_commit(void);
const char * llama_compiler(void);
const char * llama_build_target(void);
const char * llama_build_info(void);
void llama_print_build_info(void);
+2 -1
View File
@@ -1,6 +1,7 @@
#include "ggml.h" #include "ggml.h"
#include "gguf.h" #include "gguf.h"
#include "build-info.h"
#include "common.h" #include "common.h"
#include "log.h" #include "log.h"
#include "llama.h" #include "llama.h"
@@ -372,7 +373,7 @@ void common_init() {
const char * build_type = " (debug)"; const char * build_type = " (debug)";
#endif #endif
LOG_DBG("build: %d (%s) with %s for %s%s\n", LLAMA_BUILD_NUMBER, LLAMA_COMMIT, LLAMA_COMPILER, LLAMA_BUILD_TARGET, build_type); LOG_DBG("build: %d (%s) with %s for %s%s\n", llama_build_number(), llama_commit(), llama_compiler(), llama_build_target(), build_type);
} }
std::string common_params_get_system_info(const common_params & params) { std::string common_params_get_system_info(const common_params & params) {
+2 -14
View File
@@ -2,9 +2,10 @@
#pragma once #pragma once
#include "llama-cpp.h"
#include "ggml-opt.h" #include "ggml-opt.h"
#include "ggml.h" #include "ggml.h"
#include "llama-cpp.h"
#include <set> #include <set>
#include <sstream> #include <sstream>
@@ -27,11 +28,6 @@
#define die(msg) do { fputs("error: " msg "\n", stderr); exit(1); } while (0) #define die(msg) do { fputs("error: " msg "\n", stderr); exit(1); } while (0)
#define die_fmt(fmt, ...) do { fprintf(stderr, "error: " fmt "\n", __VA_ARGS__); exit(1); } while (0) #define die_fmt(fmt, ...) do { fprintf(stderr, "error: " fmt "\n", __VA_ARGS__); exit(1); } while (0)
#define print_build_info() do { \
fprintf(stderr, "%s: build = %d (%s)\n", __func__, LLAMA_BUILD_NUMBER, LLAMA_COMMIT); \
fprintf(stderr, "%s: built with %s for %s\n", __func__, LLAMA_COMPILER, LLAMA_BUILD_TARGET); \
} while(0)
struct common_time_meas { struct common_time_meas {
common_time_meas(int64_t & t_acc, bool disable = false); common_time_meas(int64_t & t_acc, bool disable = false);
~common_time_meas(); ~common_time_meas();
@@ -53,14 +49,6 @@ struct common_adapter_lora_info {
using llama_tokens = std::vector<llama_token>; using llama_tokens = std::vector<llama_token>;
// build info
extern int LLAMA_BUILD_NUMBER;
extern const char * LLAMA_COMMIT;
extern const char * LLAMA_COMPILER;
extern const char * LLAMA_BUILD_TARGET;
const static std::string build_info("b" + std::to_string(LLAMA_BUILD_NUMBER) + "-" + LLAMA_COMMIT);
struct common_control_vector_load_info; struct common_control_vector_load_info;
// //
+3 -2
View File
@@ -1,5 +1,6 @@
#include "arg.h" #include "arg.h"
#include "build-info.h"
#include "common.h" #include "common.h"
#include "log.h" #include "log.h"
#include "download.h" #include "download.h"
@@ -303,7 +304,7 @@ static int common_download_file_single_online(const std::string & url,
headers.emplace(h.first, h.second); headers.emplace(h.first, h.second);
} }
if (headers.find("User-Agent") == headers.end()) { if (headers.find("User-Agent") == headers.end()) {
headers.emplace("User-Agent", "llama-cpp/" + build_info); headers.emplace("User-Agent", "llama-cpp/" + std::string(llama_build_info()));
} }
if (!opts.bearer_token.empty()) { if (!opts.bearer_token.empty()) {
headers.emplace("Authorization", "Bearer " + opts.bearer_token); headers.emplace("Authorization", "Bearer " + opts.bearer_token);
@@ -441,7 +442,7 @@ std::pair<long, std::vector<char>> common_remote_get_content(const std::string
headers.emplace(h.first, h.second); headers.emplace(h.first, h.second);
} }
if (headers.find("User-Agent") == headers.end()) { if (headers.find("User-Agent") == headers.end()) {
headers.emplace("User-Agent", "llama-cpp/" + build_info); headers.emplace("User-Agent", "llama-cpp/" + std::string(llama_build_info()));
} }
if (params.timeout > 0) { if (params.timeout > 0) {
+2 -1
View File
@@ -1,5 +1,6 @@
#include "hf-cache.h" #include "hf-cache.h"
#include "build-info.h"
#include "common.h" #include "common.h"
#include "log.h" #include "log.h"
#include "http.h" #include "http.h"
@@ -200,7 +201,7 @@ static nl::json api_get(const std::string & url,
auto [cli, parts] = common_http_client(url); auto [cli, parts] = common_http_client(url);
httplib::Headers headers = { httplib::Headers headers = {
{"User-Agent", "llama-cpp/" + build_info}, {"User-Agent", "llama-cpp/" + std::string(llama_build_info())},
{"Accept", "application/json"} {"Accept", "application/json"}
}; };
+4
View File
@@ -23,6 +23,10 @@
int common_log_verbosity_thold = LOG_DEFAULT_LLAMA; int common_log_verbosity_thold = LOG_DEFAULT_LLAMA;
int common_log_get_verbosity_thold(void) {
return common_log_verbosity_thold;
}
void common_log_set_verbosity_thold(int verbosity) { void common_log_set_verbosity_thold(int verbosity) {
common_log_verbosity_thold = verbosity; common_log_verbosity_thold = verbosity;
} }
+2 -2
View File
@@ -38,7 +38,7 @@ enum log_colors {
// needed by the LOG_TMPL macro to avoid computing log arguments if the verbosity lower // needed by the LOG_TMPL macro to avoid computing log arguments if the verbosity lower
// set via common_log_set_verbosity() // set via common_log_set_verbosity()
extern int common_log_verbosity_thold; int common_log_get_verbosity_thold(void);
void common_log_set_verbosity_thold(int verbosity); // not thread-safe void common_log_set_verbosity_thold(int verbosity); // not thread-safe
@@ -98,7 +98,7 @@ void common_log_flush (struct common_log * log); // f
#define LOG_TMPL(level, verbosity, ...) \ #define LOG_TMPL(level, verbosity, ...) \
do { \ do { \
if ((verbosity) <= common_log_verbosity_thold) { \ if ((verbosity) <= common_log_get_verbosity_thold()) { \
common_log_add(common_log_main(), (level), __VA_ARGS__); \ common_log_add(common_log_main(), (level), __VA_ARGS__); \
} \ } \
} while (0) } while (0)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-batched) set(TARGET llama-batched)
add_executable(${TARGET} batched.cpp) add_executable(${TARGET} batched.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
@@ -1,5 +1,5 @@
set(TARGET llama-convert-llama2c-to-ggml) set(TARGET llama-convert-llama2c-to-ggml)
add_executable(${TARGET} convert-llama2c-to-ggml.cpp) add_executable(${TARGET} convert-llama2c-to-ggml.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-debug) set(TARGET llama-debug)
add_executable(${TARGET} debug.cpp) add_executable(${TARGET} debug.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-diffusion-cli) set(TARGET llama-diffusion-cli)
add_executable(${TARGET} diffusion-cli.cpp) add_executable(${TARGET} diffusion-cli.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE llama common ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama llama-common ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-embedding) set(TARGET llama-embedding)
add_executable(${TARGET} embedding.cpp) add_executable(${TARGET} embedding.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,7 +1,7 @@
set(TARGET llama-eval-callback) set(TARGET llama-eval-callback)
add_executable(${TARGET} eval-callback.cpp) add_executable(${TARGET} eval-callback.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_BUILD_TESTS) if(LLAMA_BUILD_TESTS)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-gen-docs) set(TARGET llama-gen-docs)
add_executable(${TARGET} gen-docs.cpp) add_executable(${TARGET} gen-docs.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-idle) set(TARGET llama-idle)
add_executable(${TARGET} idle.cpp) add_executable(${TARGET} idle.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE llama common ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama llama-common ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_11) target_compile_features(${TARGET} PRIVATE cxx_std_11)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-lookahead) set(TARGET llama-lookahead)
add_executable(${TARGET} lookahead.cpp) add_executable(${TARGET} lookahead.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+4 -4
View File
@@ -1,23 +1,23 @@
set(TARGET llama-lookup) set(TARGET llama-lookup)
add_executable(${TARGET} lookup.cpp) add_executable(${TARGET} lookup.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
set(TARGET llama-lookup-create) set(TARGET llama-lookup-create)
add_executable(${TARGET} lookup-create.cpp) add_executable(${TARGET} lookup-create.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
set(TARGET llama-lookup-merge) set(TARGET llama-lookup-merge)
add_executable(${TARGET} lookup-merge.cpp) add_executable(${TARGET} lookup-merge.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
set(TARGET llama-lookup-stats) set(TARGET llama-lookup-stats)
add_executable(${TARGET} lookup-stats.cpp) add_executable(${TARGET} lookup-stats.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-parallel) set(TARGET llama-parallel)
add_executable(${TARGET} parallel.cpp) add_executable(${TARGET} parallel.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-passkey) set(TARGET llama-passkey)
add_executable(${TARGET} passkey.cpp) add_executable(${TARGET} passkey.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-retrieval) set(TARGET llama-retrieval)
add_executable(${TARGET} retrieval.cpp) add_executable(${TARGET} retrieval.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-save-load-state) set(TARGET llama-save-load-state)
add_executable(${TARGET} save-load-state.cpp) add_executable(${TARGET} save-load-state.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-speculative-simple) set(TARGET llama-speculative-simple)
add_executable(${TARGET} speculative-simple.cpp) add_executable(${TARGET} speculative-simple.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-speculative) set(TARGET llama-speculative)
add_executable(${TARGET} speculative.cpp) add_executable(${TARGET} speculative.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -5,5 +5,5 @@
set(TARGET llama-ls-sycl-device) set(TARGET llama-ls-sycl-device)
add_executable(${TARGET} ls-sycl-device.cpp) add_executable(${TARGET} ls-sycl-device.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,5 +1,5 @@
set(TARGET llama-finetune) set(TARGET llama-finetune)
add_executable(${TARGET} finetune.cpp) add_executable(${TARGET} finetune.cpp)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_11) target_compile_features(${TARGET} PRIVATE cxx_std_11)
+2 -2
View File
@@ -1,9 +1,9 @@
set(TARGET llama-vdot) set(TARGET llama-vdot)
add_executable(${TARGET} vdot.cpp) add_executable(${TARGET} vdot.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
set(TARGET llama-q8dot) set(TARGET llama-q8dot)
add_executable(${TARGET} q8dot.cpp) add_executable(${TARGET} q8dot.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+4 -4
View File
@@ -10,7 +10,7 @@ function(llama_build source)
endif() endif()
add_executable(${TEST_TARGET} ${TEST_SOURCES}) add_executable(${TEST_TARGET} ${TEST_SOURCES})
target_link_libraries(${TEST_TARGET} PRIVATE common) target_link_libraries(${TEST_TARGET} PRIVATE llama llama-common)
if (LLAMA_TESTS_INSTALL) if (LLAMA_TESTS_INSTALL)
install(TARGETS ${TEST_TARGET} RUNTIME) install(TARGETS ${TEST_TARGET} RUNTIME)
endif() endif()
@@ -105,7 +105,7 @@ function(llama_build_and_test source)
if (LLAMA_TESTS_INSTALL) if (LLAMA_TESTS_INSTALL)
install(TARGETS ${TEST_TARGET} RUNTIME) install(TARGETS ${TEST_TARGET} RUNTIME)
endif() endif()
target_link_libraries(${TEST_TARGET} PRIVATE common) target_link_libraries(${TEST_TARGET} PRIVATE llama-common)
add_test( add_test(
NAME ${TEST_TARGET} NAME ${TEST_TARGET}
@@ -269,11 +269,11 @@ if (TARGET cpp-httplib)
get_target_property(_cpp_httplib_defs cpp-httplib INTERFACE_COMPILE_DEFINITIONS) get_target_property(_cpp_httplib_defs cpp-httplib INTERFACE_COMPILE_DEFINITIONS)
if (_cpp_httplib_defs MATCHES "CPPHTTPLIB_OPENSSL_SUPPORT") if (_cpp_httplib_defs MATCHES "CPPHTTPLIB_OPENSSL_SUPPORT")
add_library(gguf-model-data STATIC gguf-model-data.cpp) add_library(gguf-model-data STATIC gguf-model-data.cpp)
target_link_libraries(gguf-model-data PRIVATE common cpp-httplib) target_link_libraries(gguf-model-data PRIVATE llama-common cpp-httplib)
target_include_directories(gguf-model-data PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}) target_include_directories(gguf-model-data PUBLIC ${CMAKE_CURRENT_SOURCE_DIR})
add_executable(test-gguf-model-data test-gguf-model-data.cpp) add_executable(test-gguf-model-data test-gguf-model-data.cpp)
target_link_libraries(test-gguf-model-data PRIVATE gguf-model-data common) target_link_libraries(test-gguf-model-data PRIVATE gguf-model-data llama-common)
llama_test(test-gguf-model-data LABEL "model") llama_test(test-gguf-model-data LABEL "model")
# test-quant-type-selection requires gguf-model-data for remote model metadata # test-quant-type-selection requires gguf-model-data for remote model metadata
+6 -3
View File
@@ -1,10 +1,13 @@
#include "ggml.h"
#include "ggml-cpu.h"
#include "llama.h" #include "llama.h"
#include "build-info.h"
#include "common.h" #include "common.h"
#include "../src/llama-model.h" #include "../src/llama-model.h"
#include "ggml.h"
#include "ggml-cpu.h"
#include <algorithm> #include <algorithm>
#include <cassert> #include <cassert>
#include <cinttypes> #include <cinttypes>
@@ -298,7 +301,7 @@ int main(int argc, char ** argv) {
return 1; return 1;
} }
print_build_info(); llama_print_build_info();
// load the model // load the model
fprintf(stderr, "Loading model\n"); fprintf(stderr, "Loading model\n");
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-batched-bench) set(TARGET llama-batched-bench)
add_executable(${TARGET} batched-bench.cpp) add_executable(${TARGET} batched-bench.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-cli) set(TARGET llama-cli)
add_executable(${TARGET} cli.cpp) add_executable(${TARGET} cli.cpp)
target_link_libraries(${TARGET} PRIVATE server-context PUBLIC common ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE server-context PUBLIC llama-common ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
include_directories(../server) include_directories(../server)
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-completion) set(TARGET llama-completion)
add_executable(${TARGET} completion.cpp) add_executable(${TARGET} completion.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-cvector-generator) set(TARGET llama-cvector-generator)
add_executable(${TARGET} cvector-generator.cpp pca.hpp) add_executable(${TARGET} cvector-generator.cpp pca.hpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
@@ -2,6 +2,7 @@
#include "gguf.h" #include "gguf.h"
#include "arg.h" #include "arg.h"
#include "build-info.h"
#include "common.h" #include "common.h"
#include "llama.h" #include "llama.h"
#include "pca.hpp" #include "pca.hpp"
@@ -420,7 +421,7 @@ int main(int argc, char ** argv) {
params.cb_eval_user_data = &cb_data; params.cb_eval_user_data = &cb_data;
params.warmup = false; params.warmup = false;
print_build_info(); llama_print_build_info();
llama_backend_init(); llama_backend_init();
llama_numa_init(params.numa); llama_numa_init(params.numa);
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-export-lora) set(TARGET llama-export-lora)
add_executable(${TARGET} export-lora.cpp) add_executable(${TARGET} export-lora.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-fit-params) set(TARGET llama-fit-params)
add_executable(${TARGET} fit-params.cpp) add_executable(${TARGET} fit-params.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-gguf-split) set(TARGET llama-gguf-split)
add_executable(${TARGET} gguf-split.cpp) add_executable(${TARGET} gguf-split.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+7 -4
View File
@@ -1,7 +1,10 @@
#include "llama.h"
#include "build-info.h"
#include "common.h"
#include "ggml.h" #include "ggml.h"
#include "gguf.h" #include "gguf.h"
#include "llama.h"
#include "common.h"
#include <algorithm> #include <algorithm>
#include <cinttypes> #include <cinttypes>
@@ -101,8 +104,8 @@ static void split_params_parse_ex(int argc, const char ** argv, split_params & p
split_print_usage(argv[0]); split_print_usage(argv[0]);
exit(0); exit(0);
} else if (arg == "--version") { } else if (arg == "--version") {
fprintf(stderr, "version: %d (%s)\n", LLAMA_BUILD_NUMBER, LLAMA_COMMIT); fprintf(stderr, "version: %d (%s)\n", llama_build_number(), llama_commit());
fprintf(stderr, "built with %s for %s\n", LLAMA_COMPILER, LLAMA_BUILD_TARGET); fprintf(stderr, "built with %s for %s\n", llama_compiler(), llama_build_target());
exit(0); exit(0);
} else if (arg == "--dry-run") { } else if (arg == "--dry-run") {
arg_found = true; arg_found = true;
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-imatrix) set(TARGET llama-imatrix)
add_executable(${TARGET} imatrix.cpp) add_executable(${TARGET} imatrix.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-bench) set(TARGET llama-bench)
add_executable(${TARGET} llama-bench.cpp) add_executable(${TARGET} llama-bench.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+3 -2
View File
@@ -19,6 +19,7 @@
#include <vector> #include <vector>
#include <unordered_set> #include <unordered_set>
#include "build-info.h"
#include "common.h" #include "common.h"
#include "download.h" #include "download.h"
#include "ggml.h" #include "ggml.h"
@@ -1624,8 +1625,8 @@ struct test {
} }
}; };
const std::string test::build_commit = LLAMA_COMMIT; const std::string test::build_commit = llama_commit();
const int test::build_number = LLAMA_BUILD_NUMBER; const int test::build_number = llama_build_number();
struct printer { struct printer {
virtual ~printer() {} virtual ~printer() {}
+5 -5
View File
@@ -86,12 +86,12 @@ if (TARGET BUILD_INFO)
add_dependencies(mtmd-helper BUILD_INFO) add_dependencies(mtmd-helper BUILD_INFO)
endif() endif()
# if mtmd is linked against common, we throw an error # if mtmd is linked against llama-common, we throw an error
if (TARGET mtmd) if (TARGET mtmd)
get_target_property(libs mtmd LINK_LIBRARIES) get_target_property(libs mtmd LINK_LIBRARIES)
if (libs AND "common" IN_LIST libs) if (libs AND "llama-common" IN_LIST libs)
message(FATAL_ERROR "mtmd is designed to be a public library.\n" message(FATAL_ERROR "mtmd is designed to be a public library.\n"
"It must not link against common") "It must not link against llama-common")
endif() endif()
endif() endif()
@@ -106,11 +106,11 @@ set_target_properties (${TARGET} PROPERTIES OUTPUT_NAME llama-mtmd-cli)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
endif() endif()
target_link_libraries (${TARGET} PRIVATE common mtmd Threads::Threads) target_link_libraries (${TARGET} PRIVATE llama-common mtmd Threads::Threads)
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
# mtmd-debug tool # mtmd-debug tool
add_executable(llama-mtmd-debug debug/mtmd-debug.cpp) add_executable(llama-mtmd-debug debug/mtmd-debug.cpp)
set_target_properties(llama-mtmd-debug PROPERTIES OUTPUT_NAME llama-mtmd-debug) set_target_properties(llama-mtmd-debug PROPERTIES OUTPUT_NAME llama-mtmd-debug)
target_link_libraries(llama-mtmd-debug PRIVATE common mtmd Threads::Threads) target_link_libraries(llama-mtmd-debug PRIVATE llama-common mtmd Threads::Threads)
target_compile_features(llama-mtmd-debug PRIVATE cxx_std_17) target_compile_features(llama-mtmd-debug PRIVATE cxx_std_17)
+2 -2
View File
@@ -2,7 +2,7 @@ if (NOT WIN32 OR NOT BUILD_SHARED_LIBS)
# this tool is disabled on Windows when building with shared libraries because it uses internal functions not exported with LLAMA_API # this tool is disabled on Windows when building with shared libraries because it uses internal functions not exported with LLAMA_API
set(TARGET llama-debug-template-parser) set(TARGET llama-debug-template-parser)
add_executable(${TARGET} debug-template-parser.cpp) add_executable(${TARGET} debug-template-parser.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
@@ -12,7 +12,7 @@ endif()
set(TARGET llama-template-analysis) set(TARGET llama-template-analysis)
add_executable(${TARGET} template-analysis.cpp) add_executable(${TARGET} template-analysis.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-perplexity) set(TARGET llama-perplexity)
add_executable(${TARGET} perplexity.cpp) add_executable(${TARGET} perplexity.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-quantize) set(TARGET llama-quantize)
add_executable(${TARGET} quantize.cpp) add_executable(${TARGET} quantize.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_include_directories(${TARGET} PRIVATE ../../common) target_include_directories(${TARGET} PRIVATE ../../common)
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+5 -2
View File
@@ -1,5 +1,8 @@
#include "common.h"
#include "llama.h" #include "llama.h"
#include "build-info.h"
#include "common.h"
#include "gguf.h" #include "gguf.h"
#include <algorithm> #include <algorithm>
@@ -709,7 +712,7 @@ int main(int argc, char ** argv) {
} }
} }
print_build_info(); llama_print_build_info();
if (params.dry_run) { if (params.dry_run) {
fprintf(stderr, "%s: calculating quantization size for '%s' as %s", __func__, fname_inp.c_str(), ftype_str.c_str()); fprintf(stderr, "%s: calculating quantization size for '%s' as %s", __func__, fname_inp.c_str(), ftype_str.c_str());
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-results) set(TARGET llama-results)
add_executable(${TARGET} results.cpp) add_executable(${TARGET} results.cpp)
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+2 -2
View File
@@ -23,7 +23,7 @@ endif()
target_include_directories(${TARGET} PRIVATE ../mtmd) target_include_directories(${TARGET} PRIVATE ../mtmd)
target_include_directories(${TARGET} PRIVATE ${CMAKE_SOURCE_DIR}) target_include_directories(${TARGET} PRIVATE ${CMAKE_SOURCE_DIR})
target_link_libraries(${TARGET} PUBLIC common mtmd ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PUBLIC llama-common mtmd ${CMAKE_THREAD_LIBS_INIT})
# llama-server executable # llama-server executable
@@ -68,6 +68,6 @@ install(TARGETS ${TARGET} RUNTIME)
target_include_directories(${TARGET} PRIVATE ../mtmd) target_include_directories(${TARGET} PRIVATE ../mtmd)
target_include_directories(${TARGET} PRIVATE ${CMAKE_SOURCE_DIR}) target_include_directories(${TARGET} PRIVATE ${CMAKE_SOURCE_DIR})
target_link_libraries(${TARGET} PRIVATE server-context PUBLIC common cpp-httplib ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE server-context PUBLIC llama-common cpp-httplib ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+2 -1
View File
@@ -4,6 +4,7 @@
#include "server-task.h" #include "server-task.h"
#include "server-queue.h" #include "server-queue.h"
#include "build-info.h"
#include "common.h" #include "common.h"
#include "llama.h" #include "llama.h"
#include "log.h" #include "log.h"
@@ -3010,7 +3011,7 @@ server_context_meta server_context::get_meta() const {
auto eos_token_str = eos_id != LLAMA_TOKEN_NULL ? common_token_to_piece(impl->ctx, eos_id, true) : ""; auto eos_token_str = eos_id != LLAMA_TOKEN_NULL ? common_token_to_piece(impl->ctx, eos_id, true) : "";
return server_context_meta { return server_context_meta {
/* build_info */ build_info, /* build_info */ std::string(llama_build_info()),
/* model_name */ impl->model_name, /* model_name */ impl->model_name,
/* model_aliases */ impl->model_aliases, /* model_aliases */ impl->model_aliases,
/* model_tags */ impl->model_tags, /* model_tags */ impl->model_tags,
+2 -1
View File
@@ -1,6 +1,7 @@
#include "server-common.h" #include "server-common.h"
#include "server-models.h" #include "server-models.h"
#include "build-info.h"
#include "preset.h" #include "preset.h"
#include "download.h" #include "download.h"
@@ -936,7 +937,7 @@ void server_models_routes::init_routes() {
{"n_ctx", 0}, {"n_ctx", 0},
}}, }},
{"webui_settings", webui_settings}, {"webui_settings", webui_settings},
{"build_info", build_info}, {"build_info", std::string(llama_build_info())},
}); });
return res; return res;
} }
+8 -7
View File
@@ -1,5 +1,6 @@
#include "server-task.h" #include "server-task.h"
#include "build-info.h"
#include "chat.h" #include "chat.h"
#include "common.h" #include "common.h"
#include "json-schema-to-grammar.h" #include "json-schema-to-grammar.h"
@@ -791,7 +792,7 @@ json server_task_result_cmpl_final::to_json_oaicompat() {
})}, })},
{"created", t}, {"created", t},
{"model", oaicompat_model}, {"model", oaicompat_model},
{"system_fingerprint", build_info}, {"system_fingerprint", std::string(llama_build_info())},
{"object", "text_completion"}, {"object", "text_completion"},
{"usage", usage_json_oaicompat()}, {"usage", usage_json_oaicompat()},
{"id", oaicompat_cmpl_id} {"id", oaicompat_cmpl_id}
@@ -839,7 +840,7 @@ json server_task_result_cmpl_final::to_json_oaicompat_chat() {
{"choices", json::array({choice})}, {"choices", json::array({choice})},
{"created", t}, {"created", t},
{"model", oaicompat_model}, {"model", oaicompat_model},
{"system_fingerprint", build_info}, {"system_fingerprint", std::string(llama_build_info())},
{"object", "chat.completion"}, {"object", "chat.completion"},
{"usage", usage_json_oaicompat()}, {"usage", usage_json_oaicompat()},
{"id", oaicompat_cmpl_id} {"id", oaicompat_cmpl_id}
@@ -876,7 +877,7 @@ json server_task_result_cmpl_final::to_json_oaicompat_chat_stream() {
{"created", t}, {"created", t},
{"id", oaicompat_cmpl_id}, {"id", oaicompat_cmpl_id},
{"model", oaicompat_model}, {"model", oaicompat_model},
{"system_fingerprint", build_info}, {"system_fingerprint", std::string(llama_build_info())},
{"object", "chat.completion.chunk"}, {"object", "chat.completion.chunk"},
}); });
} }
@@ -892,7 +893,7 @@ json server_task_result_cmpl_final::to_json_oaicompat_chat_stream() {
{"created", t}, {"created", t},
{"id", oaicompat_cmpl_id}, {"id", oaicompat_cmpl_id},
{"model", oaicompat_model}, {"model", oaicompat_model},
{"system_fingerprint", build_info}, {"system_fingerprint", std::string(llama_build_info())},
{"object", "chat.completion.chunk"}, {"object", "chat.completion.chunk"},
}); });
@@ -904,7 +905,7 @@ json server_task_result_cmpl_final::to_json_oaicompat_chat_stream() {
{"created", t}, {"created", t},
{"id", oaicompat_cmpl_id}, {"id", oaicompat_cmpl_id},
{"model", oaicompat_model}, {"model", oaicompat_model},
{"system_fingerprint", build_info}, {"system_fingerprint", std::string(llama_build_info())},
{"object", "chat.completion.chunk"}, {"object", "chat.completion.chunk"},
{"usage", usage_json_oaicompat()}, {"usage", usage_json_oaicompat()},
}); });
@@ -1469,7 +1470,7 @@ json server_task_result_cmpl_partial::to_json_oaicompat() {
})}, })},
{"created", t}, {"created", t},
{"model", oaicompat_model}, {"model", oaicompat_model},
{"system_fingerprint", build_info}, {"system_fingerprint", std::string(llama_build_info())},
{"object", "text_completion"}, {"object", "text_completion"},
{"id", oaicompat_cmpl_id} {"id", oaicompat_cmpl_id}
}; };
@@ -1506,7 +1507,7 @@ json server_task_result_cmpl_partial::to_json_oaicompat_chat() {
{"created", t}, {"created", t},
{"id", oaicompat_cmpl_id}, {"id", oaicompat_cmpl_id},
{"model", oaicompat_model}, {"model", oaicompat_model},
{"system_fingerprint", build_info}, {"system_fingerprint", std::string(llama_build_info())},
{"object", "chat.completion.chunk"}, {"object", "chat.completion.chunk"},
}); });
}; };
+2 -1
View File
@@ -5,6 +5,7 @@
#include "server-tools.h" #include "server-tools.h"
#include "arg.h" #include "arg.h"
#include "build-info.h"
#include "common.h" #include "common.h"
#include "llama.h" #include "llama.h"
#include "log.h" #include "log.h"
@@ -108,7 +109,7 @@ int main(int argc, char ** argv) {
llama_backend_init(); llama_backend_init();
llama_numa_init(params.numa); llama_numa_init(params.numa);
LOG_INF("build_info: %s\n", build_info.c_str()); LOG_INF("build_info: %s\n", llama_build_info());
LOG_INF("%s\n", common_params_get_system_info(params).c_str()); LOG_INF("%s\n", common_params_get_system_info(params).c_str());
server_http_context ctx_http; server_http_context ctx_http;
+1 -1
View File
@@ -3,5 +3,5 @@ add_executable(${TARGET} tokenize.cpp)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
install(TARGETS ${TARGET} RUNTIME) install(TARGETS ${TARGET} RUNTIME)
endif() endif()
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama-common llama ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
+1 -1
View File
@@ -1,6 +1,6 @@
set(TARGET llama-tts) set(TARGET llama-tts)
add_executable(${TARGET} tts.cpp) add_executable(${TARGET} tts.cpp)
target_link_libraries(${TARGET} PRIVATE llama common ${CMAKE_THREAD_LIBS_INIT}) target_link_libraries(${TARGET} PRIVATE llama llama-common ${CMAKE_THREAD_LIBS_INIT})
target_compile_features(${TARGET} PRIVATE cxx_std_17) target_compile_features(${TARGET} PRIVATE cxx_std_17)
if(LLAMA_TOOLS_INSTALL) if(LLAMA_TOOLS_INSTALL)
+2
View File
@@ -5,6 +5,8 @@ find_package(Threads REQUIRED)
llama_add_compile_flags() llama_add_compile_flags()
set(CMAKE_POSITION_INDEPENDENT_CODE ON)
add_library(${TARGET} STATIC httplib.cpp httplib.h) add_library(${TARGET} STATIC httplib.cpp httplib.h)
# disable warnings in 3rd party code # disable warnings in 3rd party code