![]() |
LlamaLib
v2.0.2
Cross-platform library for local LLMs
|
This is the complete list of members for LLMService, including all inherited members.
| apply_template(const json &messages) | LLM | virtual |
| apply_template_json(const json &data) override | LLMService | inlinevirtual |
| apply_template_json(const json &data) override | LLMService | virtual |
| build_apply_template_json(const json &messages) | LLM | protectedvirtual |
| build_completion_json(const std::string &prompt, int id_slot=-1) | LLM | protectedvirtual |
| build_detokenize_json(const std::vector< int32_t > &tokens) | LLM | protectedvirtual |
| build_embeddings_json(const std::string &query) | LLM | protectedvirtual |
| build_lora_list_json(const std::vector< LoraIdScalePath > &loras) | LLMProvider | protectedvirtual |
| build_lora_weight_json(const std::vector< LoraIdScale > &loras) | LLMProvider | protectedvirtual |
| build_slot_json(int id_slot, const std::string &action, const std::string &filepath) | LLMLocal | protectedvirtual |
| build_tokenize_json(const std::string &query) | LLM | protectedvirtual |
| cancel(int id_slot) override | LLMService | inlinevirtual |
| cancel(int id_slot) override | LLMService | virtual |
| completion(const std::string &prompt, CharArrayFn callback=nullptr, int id_slot=-1, bool return_response_json=false) | LLM | virtual |
| completion_json(const json &data, CharArrayFn callback=nullptr, bool callbackWithJSON=true) override | LLMService | inlinevirtual |
| completion_json(const json &data, CharArrayFn callback=nullptr, bool callbackWithJSON=true) override | LLMService | virtual |
| completion_params | LLM | |
| create_LLM_library(const std::string &command) | LLMService | |
| create_LLM_library_backend(const std::string &command, const std::string &llm_lib_filename) | LLMService | protected |
| debug(int debug_level) override | LLMService | inlinevirtual |
| debug(int debug_level) override | LLMService | virtual |
| debug_implementation() override | LLMService | inlinevirtual |
| debug_implementation() override | LLMService | inlinevirtual |
| detokenize(const std::vector< int32_t > &tokens) | LLM | virtual |
| detokenize_json(const json &data) override | LLMService | inlinevirtual |
| detokenize_json(const json &data) override | LLMService | virtual |
| embedding_size() override | LLMService | inlinevirtual |
| embedding_size() override | LLMService | virtual |
| embeddings(const std::string &query) | LLM | virtual |
| embeddings_json(const json &data) override | LLMService | inlinevirtual |
| embeddings_json(const json &data) override | LLMService | virtual |
| enable_reasoning(bool reasoning) override | LLMService | virtual |
| encapsulate_route(const json &body, handler_t route_handler) (defined in LLMService) | LLMService | |
| from_command(const std::string &command) | LLMService | static |
| from_command(int argc, char **argv) | LLMService | static |
| from_command(const std::string &command) | LLMService | static |
| from_command(int argc, char **argv) | LLMService | static |
| from_params(const json ¶ms_json) | LLMService | static |
| get_command() | LLMService | inline |
| get_completion_params() | LLM | inlinevirtual |
| get_grammar() | LLM | inlinevirtual |
| get_next_available_slot() override | LLMService | inlinevirtual |
| get_next_available_slot() override | LLMService | virtual |
| grammar | LLM | |
| handle | LLMService | |
| has_gpu_layers(const std::string &command) | LLM | static |
| init(int argc, char **argv) | LLMService | |
| init(const std::string ¶ms_string) | LLMService | |
| init(const char *params_string) | LLMService | |
| join_server() override | LLMService | inlinevirtual |
| join_server() override | LLMService | virtual |
| join_service() override | LLMService | inlinevirtual |
| join_service() override | LLMService | virtual |
| jsonToArguments(const json ¶ms_json) | LLMService | static |
| llm | LLMService | |
| LLM_args_to_command(const std::string &model_path, int num_slots=1, int num_threads=-1, int num_GPU_layers=0, bool flash_attention=false, int context_size=4096, int batch_size=2048, bool embedding_only=false, const std::vector< std::string > &lora_paths={}) | LLM | static |
| LLMService() | LLMService | |
| LLMService(const std::string &model_path, int num_slots=1, int num_threads=-1, int num_GPU_layers=0, bool flash_attention=false, int context_size=4096, int batch_size=2048, bool embedding_only=false, const std::vector< std::string > &lora_paths={}) | LLMService | |
| LLMService() | LLMService | |
| LLMService(const std::string &model_path, int num_slots=1, int num_threads=-1, int num_GPU_layers=0, bool flash_attention=false, int context_size=4096, int batch_size=2048, bool embedding_only=false, const std::vector< std::string > &lora_paths={}) | LLMService | |
| load_slot(int id_slot, const std::string &filepath) | LLMLocal | inlinevirtual |
| logging_callback(CharArrayFn callback) override | LLMService | inlinevirtual |
| logging_callback(CharArrayFn callback) override | LLMService | virtual |
| logging_stop() | LLMProvider | virtual |
| lora_list() | LLMProvider | virtual |
| lora_list_json() override | LLMService | inlinevirtual |
| lora_list_json() override | LLMService | virtual |
| lora_weight(const std::vector< LoraIdScale > &loras) | LLMProvider | virtual |
| lora_weight_json(const json &data) override | LLMService | inlinevirtual |
| lora_weight_json(const json &data) override | LLMService | virtual |
| n_keep | LLM | |
| parse_apply_template_json(const json &result) | LLM | protectedvirtual |
| parse_completion_json(const json &result) | LLM | protectedvirtual |
| parse_detokenize_json(const json &result) | LLM | protectedvirtual |
| parse_embeddings_json(const json &result) | LLM | protectedvirtual |
| parse_lora_list_json(const json &result) | LLMProvider | protectedvirtual |
| parse_lora_weight_json(const json &result) | LLMProvider | protectedvirtual |
| parse_slot_json(const json &result) | LLMLocal | protectedvirtual |
| parse_tokenize_json(const json &result) | LLM | protectedvirtual |
| reasoning_enabled | LLMProvider | protected |
| save_slot(int id_slot, const std::string &filepath) | LLMLocal | inlinevirtual |
| search_paths | LLMService | protected |
| set_completion_params(json completion_params_) | LLM | inlinevirtual |
| set_grammar(std::string grammar_) | LLM | inlinevirtual |
| set_SSL(const std::string &cert, const std::string &key) override | LLMService | inlinevirtual |
| set_SSL(const std::string &SSL_cert, const std::string &SSL_key) override | LLMService | virtual |
| slot(int id_slot, const std::string &action, const std::string &filepath) | LLMLocal | protectedvirtual |
| slot_json(const json &data) override | LLMService | inlinevirtual |
| slot_json(const json &data) override | LLMService | virtual |
| start() override | LLMService | inlinevirtual |
| start() override | LLMService | virtual |
| start_server(const std::string &host="0.0.0.0", int port=-1, const std::string &API_key="") override | LLMService | inlinevirtual |
| start_server(const std::string &host="0.0.0.0", int port=-1, const std::string &API_key="") override | LLMService | virtual |
| started() override | LLMService | inlinevirtual |
| started() override | LLMService | virtual |
| stop() override | LLMService | inlinevirtual |
| stop() override | LLMService | virtual |
| stop_server() override | LLMService | inlinevirtual |
| stop_server() override | LLMService | virtual |
| tokenize(const std::string &query) | LLM | virtual |
| tokenize_json(const json &data) override | LLMService | inlinevirtual |
| tokenize_json(const json &data) override | LLMService | virtual |
| ~LLM()=default | LLM | virtual |
| ~LLMProvider() | LLMProvider | virtual |
| ~LLMService() | LLMService | |
| ~LLMService() | LLMService |