Fixing llamav1

This commit is contained in:
Nicolas Patry
2023-08-16 14:53:29 +02:00
parent 3071134788
commit 102fa4c2e3

View File

@ -13,7 +13,7 @@ pub struct LlamaConfig {
pub vocab_size: usize,
pub num_hidden_layers: usize,
pub num_attention_heads: usize,
pub num_key_value_heads: usize,
pub num_key_value_heads: Option<usize>,
pub rms_norm_eps: f64,
}
@ -25,7 +25,7 @@ impl LlamaConfig {
vocab_size: self.vocab_size,
num_hidden_layers: self.num_hidden_layers,
num_attention_heads: self.num_attention_heads,
num_key_value_heads: self.num_key_value_heads,
num_key_value_heads: self.num_key_value_heads.unwrap_or(self.num_attention_heads),
rms_norm_eps: self.rms_norm_eps,
use_flash_attn,
}