LlamaLib  v2.0.2
Cross-platform library for local LLMs
Loading...
Searching...
No Matches
LLMService Member List

This is the complete list of members for LLMService, including all inherited members.

apply_template(const json &messages)LLMvirtual
apply_template_json(const json &data) overrideLLMServiceinlinevirtual
apply_template_json(const json &data) overrideLLMServicevirtual
build_apply_template_json(const json &messages)LLMprotectedvirtual
build_completion_json(const std::string &prompt, int id_slot=-1)LLMprotectedvirtual
build_detokenize_json(const std::vector< int32_t > &tokens)LLMprotectedvirtual
build_embeddings_json(const std::string &query)LLMprotectedvirtual
build_lora_list_json(const std::vector< LoraIdScalePath > &loras)LLMProviderprotectedvirtual
build_lora_weight_json(const std::vector< LoraIdScale > &loras)LLMProviderprotectedvirtual
build_slot_json(int id_slot, const std::string &action, const std::string &filepath)LLMLocalprotectedvirtual
build_tokenize_json(const std::string &query)LLMprotectedvirtual
cancel(int id_slot) overrideLLMServiceinlinevirtual
cancel(int id_slot) overrideLLMServicevirtual
completion(const std::string &prompt, CharArrayFn callback=nullptr, int id_slot=-1, bool return_response_json=false)LLMvirtual
completion_json(const json &data, CharArrayFn callback=nullptr, bool callbackWithJSON=true) overrideLLMServiceinlinevirtual
completion_json(const json &data, CharArrayFn callback=nullptr, bool callbackWithJSON=true) overrideLLMServicevirtual
completion_paramsLLM
create_LLM_library(const std::string &command)LLMService
create_LLM_library_backend(const std::string &command, const std::string &llm_lib_filename)LLMServiceprotected
debug(int debug_level) overrideLLMServiceinlinevirtual
debug(int debug_level) overrideLLMServicevirtual
debug_implementation() overrideLLMServiceinlinevirtual
debug_implementation() overrideLLMServiceinlinevirtual
detokenize(const std::vector< int32_t > &tokens)LLMvirtual
detokenize_json(const json &data) overrideLLMServiceinlinevirtual
detokenize_json(const json &data) overrideLLMServicevirtual
embedding_size() overrideLLMServiceinlinevirtual
embedding_size() overrideLLMServicevirtual
embeddings(const std::string &query)LLMvirtual
embeddings_json(const json &data) overrideLLMServiceinlinevirtual
embeddings_json(const json &data) overrideLLMServicevirtual
enable_reasoning(bool reasoning) overrideLLMServicevirtual
encapsulate_route(const json &body, handler_t route_handler) (defined in LLMService)LLMService
from_command(const std::string &command)LLMServicestatic
from_command(int argc, char **argv)LLMServicestatic
from_command(const std::string &command)LLMServicestatic
from_command(int argc, char **argv)LLMServicestatic
from_params(const json &params_json)LLMServicestatic
get_command()LLMServiceinline
get_completion_params()LLMinlinevirtual
get_grammar()LLMinlinevirtual
get_next_available_slot() overrideLLMServiceinlinevirtual
get_next_available_slot() overrideLLMServicevirtual
grammarLLM
handleLLMService
has_gpu_layers(const std::string &command)LLMstatic
init(int argc, char **argv)LLMService
init(const std::string &params_string)LLMService
init(const char *params_string)LLMService
join_server() overrideLLMServiceinlinevirtual
join_server() overrideLLMServicevirtual
join_service() overrideLLMServiceinlinevirtual
join_service() overrideLLMServicevirtual
jsonToArguments(const json &params_json)LLMServicestatic
llmLLMService
LLM_args_to_command(const std::string &model_path, int num_slots=1, int num_threads=-1, int num_GPU_layers=0, bool flash_attention=false, int context_size=4096, int batch_size=2048, bool embedding_only=false, const std::vector< std::string > &lora_paths={})LLMstatic
LLMService()LLMService
LLMService(const std::string &model_path, int num_slots=1, int num_threads=-1, int num_GPU_layers=0, bool flash_attention=false, int context_size=4096, int batch_size=2048, bool embedding_only=false, const std::vector< std::string > &lora_paths={})LLMService
LLMService()LLMService
LLMService(const std::string &model_path, int num_slots=1, int num_threads=-1, int num_GPU_layers=0, bool flash_attention=false, int context_size=4096, int batch_size=2048, bool embedding_only=false, const std::vector< std::string > &lora_paths={})LLMService
load_slot(int id_slot, const std::string &filepath)LLMLocalinlinevirtual
logging_callback(CharArrayFn callback) overrideLLMServiceinlinevirtual
logging_callback(CharArrayFn callback) overrideLLMServicevirtual
logging_stop()LLMProvidervirtual
lora_list()LLMProvidervirtual
lora_list_json() overrideLLMServiceinlinevirtual
lora_list_json() overrideLLMServicevirtual
lora_weight(const std::vector< LoraIdScale > &loras)LLMProvidervirtual
lora_weight_json(const json &data) overrideLLMServiceinlinevirtual
lora_weight_json(const json &data) overrideLLMServicevirtual
n_keepLLM
parse_apply_template_json(const json &result)LLMprotectedvirtual
parse_completion_json(const json &result)LLMprotectedvirtual
parse_detokenize_json(const json &result)LLMprotectedvirtual
parse_embeddings_json(const json &result)LLMprotectedvirtual
parse_lora_list_json(const json &result)LLMProviderprotectedvirtual
parse_lora_weight_json(const json &result)LLMProviderprotectedvirtual
parse_slot_json(const json &result)LLMLocalprotectedvirtual
parse_tokenize_json(const json &result)LLMprotectedvirtual
reasoning_enabledLLMProviderprotected
save_slot(int id_slot, const std::string &filepath)LLMLocalinlinevirtual
search_pathsLLMServiceprotected
set_completion_params(json completion_params_)LLMinlinevirtual
set_grammar(std::string grammar_)LLMinlinevirtual
set_SSL(const std::string &cert, const std::string &key) overrideLLMServiceinlinevirtual
set_SSL(const std::string &SSL_cert, const std::string &SSL_key) overrideLLMServicevirtual
slot(int id_slot, const std::string &action, const std::string &filepath)LLMLocalprotectedvirtual
slot_json(const json &data) overrideLLMServiceinlinevirtual
slot_json(const json &data) overrideLLMServicevirtual
start() overrideLLMServiceinlinevirtual
start() overrideLLMServicevirtual
start_server(const std::string &host="0.0.0.0", int port=-1, const std::string &API_key="") overrideLLMServiceinlinevirtual
start_server(const std::string &host="0.0.0.0", int port=-1, const std::string &API_key="") overrideLLMServicevirtual
started() overrideLLMServiceinlinevirtual
started() overrideLLMServicevirtual
stop() overrideLLMServiceinlinevirtual
stop() overrideLLMServicevirtual
stop_server() overrideLLMServiceinlinevirtual
stop_server() overrideLLMServicevirtual
tokenize(const std::string &query)LLMvirtual
tokenize_json(const json &data) overrideLLMServiceinlinevirtual
tokenize_json(const json &data) overrideLLMServicevirtual
~LLM()=defaultLLMvirtual
~LLMProvider()LLMProvidervirtual
~LLMService()LLMService
~LLMService()LLMService