#include <llama.hpp>
◆ clamp_kqv
float llama_ros::Metadata::AttentionInfo::clamp_kqv |
◆ head_count
uint64_t llama_ros::Metadata::AttentionInfo::head_count |
◆ head_count_kv
uint64_t llama_ros::Metadata::AttentionInfo::head_count_kv |
◆ key_length
uint32_t llama_ros::Metadata::AttentionInfo::key_length |
◆ layer_norm_epsilon
float llama_ros::Metadata::AttentionInfo::layer_norm_epsilon |
◆ layer_norm_rms_epsilon
float llama_ros::Metadata::AttentionInfo::layer_norm_rms_epsilon |
◆ max_alibi_bias
float llama_ros::Metadata::AttentionInfo::max_alibi_bias |
◆ value_length
uint32_t llama_ros::Metadata::AttentionInfo::value_length |
The documentation for this struct was generated from the following file: