mirror of
https://github.com/huggingface/candle.git
synced 2025-06-17 11:08:52 +00:00
Make the whisper model cloneable (#1200)
* Add a quantized variant of llama2.c * Clippy fixes. * Make the whisper model cloneable.
This commit is contained in:
@ -9,7 +9,7 @@ fn embedding(vocab_size: usize, hidden_size: usize, vb: VarBuilder) -> Result<Em
|
|||||||
//
|
//
|
||||||
// We wrap the `Linear` layer here to add some tracing so that it's easier to profile the resulting
|
// We wrap the `Linear` layer here to add some tracing so that it's easier to profile the resulting
|
||||||
// model.
|
// model.
|
||||||
#[derive(Debug)]
|
#[derive(Debug, Clone)]
|
||||||
pub struct Linear {
|
pub struct Linear {
|
||||||
inner: candle_nn::Linear,
|
inner: candle_nn::Linear,
|
||||||
span: tracing::Span,
|
span: tracing::Span,
|
||||||
@ -53,6 +53,7 @@ fn layer_norm(size: usize, vb: VarBuilder) -> Result<LayerNorm> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L62
|
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L62
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
struct MultiHeadAttention {
|
struct MultiHeadAttention {
|
||||||
query: Linear,
|
query: Linear,
|
||||||
key: Linear,
|
key: Linear,
|
||||||
@ -162,6 +163,7 @@ impl MultiHeadAttention {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L111
|
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L111
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
struct ResidualAttentionBlock {
|
struct ResidualAttentionBlock {
|
||||||
attn: MultiHeadAttention,
|
attn: MultiHeadAttention,
|
||||||
attn_ln: LayerNorm,
|
attn_ln: LayerNorm,
|
||||||
@ -241,6 +243,7 @@ fn sinusoids(length: usize, channels: usize) -> Result<Tensor> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L143
|
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L143
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
pub struct AudioEncoder {
|
pub struct AudioEncoder {
|
||||||
conv1: Conv1d,
|
conv1: Conv1d,
|
||||||
conv2: Conv1d,
|
conv2: Conv1d,
|
||||||
@ -316,6 +319,7 @@ impl AudioEncoder {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L176
|
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L176
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
pub struct TextDecoder {
|
pub struct TextDecoder {
|
||||||
token_embedding: Embedding,
|
token_embedding: Embedding,
|
||||||
positional_embedding: Tensor,
|
positional_embedding: Tensor,
|
||||||
@ -380,6 +384,7 @@ impl TextDecoder {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L221
|
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L221
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
pub struct Whisper {
|
pub struct Whisper {
|
||||||
pub encoder: AudioEncoder,
|
pub encoder: AudioEncoder,
|
||||||
pub decoder: TextDecoder,
|
pub decoder: TextDecoder,
|
||||||
|
@ -19,6 +19,7 @@ fn conv1d(
|
|||||||
}
|
}
|
||||||
|
|
||||||
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L62
|
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L62
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
struct MultiHeadAttention {
|
struct MultiHeadAttention {
|
||||||
query: Linear,
|
query: Linear,
|
||||||
key: Linear,
|
key: Linear,
|
||||||
@ -128,6 +129,7 @@ impl MultiHeadAttention {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L111
|
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L111
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
struct ResidualAttentionBlock {
|
struct ResidualAttentionBlock {
|
||||||
attn: MultiHeadAttention,
|
attn: MultiHeadAttention,
|
||||||
attn_ln: LayerNorm,
|
attn_ln: LayerNorm,
|
||||||
@ -206,6 +208,7 @@ fn sinusoids(length: usize, channels: usize) -> Result<Tensor> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L143
|
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L143
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
pub struct AudioEncoder {
|
pub struct AudioEncoder {
|
||||||
conv1: Conv1d,
|
conv1: Conv1d,
|
||||||
conv2: Conv1d,
|
conv2: Conv1d,
|
||||||
@ -281,6 +284,7 @@ impl AudioEncoder {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L176
|
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L176
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
pub struct TextDecoder {
|
pub struct TextDecoder {
|
||||||
token_embedding: Embedding,
|
token_embedding: Embedding,
|
||||||
positional_embedding: Tensor,
|
positional_embedding: Tensor,
|
||||||
@ -347,6 +351,7 @@ impl TextDecoder {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L221
|
// https://github.com/openai/whisper/blob/f572f2161ba831bae131364c3bffdead7af6d210/whisper/model.py#L221
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
pub struct Whisper {
|
pub struct Whisper {
|
||||||
pub encoder: AudioEncoder,
|
pub encoder: AudioEncoder,
|
||||||
pub decoder: TextDecoder,
|
pub decoder: TextDecoder,
|
||||||
|
Reference in New Issue
Block a user