TinyLlama.cpp 1.0
A lightweight C++ implementation of the TinyLlama language model
Loading...
Searching...
No Matches
tinyllama::TinyLlamaSession Member List

This is the complete list of members for tinyllama::TinyLlamaSession, including all inherited members.

batch_generation_parallel(const std::vector< int > &current_tokens, const std::vector< int > &token_positions, const std::vector< int > &original_sequence_indices, std::vector< std::vector< float > > &batch_logits)tinyllama::TinyLlamaSessionprivate
batch_prefill_parallel(const std::vector< std::vector< int > > &all_tokens, const std::vector< int > &prompt_lengths, std::vector< std::vector< float > > &batch_final_logits)tinyllama::TinyLlamaSessionprivate
config_tinyllama::TinyLlamaSessionprivate
eos_token_id_tinyllama::TinyLlamaSessionprivate
generate(const std::string &prompt, int steps=128, float temperature=0.1f, int top_k=40, float top_p=0.9f, const std::string &system_prompt="", bool apply_q_a_format=false)tinyllama::TinyLlamaSession
generate_batch(const std::vector< std::string > &prompts, int steps=128, float temperature=0.1f, int top_k=40, float top_p=0.9f, const std::string &system_prompt="", bool apply_q_a_format=false)tinyllama::TinyLlamaSession
generated_stream_tinyllama::TinyLlamaSessionprivate
generated_text_for_api_return_tinyllama::TinyLlamaSessionprivate
get_config() consttinyllama::TinyLlamaSessioninline
get_kv_cache()tinyllama::TinyLlamaSessioninline
get_tokenizer() consttinyllama::TinyLlamaSessioninline
kv_cache_tinyllama::TinyLlamaSessionprivate
max_batch_size_tinyllama::TinyLlamaSessionprivate
model_tinyllama::TinyLlamaSessionprivate
operator=(const TinyLlamaSession &)=deletetinyllama::TinyLlamaSessionprivate
rng_tinyllama::TinyLlamaSessionprivate
threads_tinyllama::TinyLlamaSessionprivate
TinyLlamaSession(const std::string &model_path, const std::string &tokenizer_path, int threads=1, int num_gpu_layers_from_cli=0, bool cli_use_mmap=true, bool use_kv_quant=false, bool use_batch_generation=false, int max_batch_size=1)tinyllama::TinyLlamaSession
TinyLlamaSession(const TinyLlamaSession &)=deletetinyllama::TinyLlamaSessionprivate
tokenizer_tinyllama::TinyLlamaSessionprivate
use_batch_generation_tinyllama::TinyLlamaSessionprivate
~TinyLlamaSession()tinyllama::TinyLlamaSession