Skip to content

Commit

Permalink
No f32 attn in text selfattn
Browse files Browse the repository at this point in the history
  • Loading branch information
EricLBuehler committed Oct 24, 2024
1 parent f34fdb9 commit 5b627f1
Showing 1 changed file with 4 additions and 6 deletions.
10 changes: 4 additions & 6 deletions mistralrs-core/src/vision_models/mllama/text.rs
Original file line number Diff line number Diff line change
Expand Up @@ -206,12 +206,10 @@ impl MLlamaTextSelfAttention {

let mut attn_output = Sdpa
.run_attention(
&q.contiguous()?.to_dtype(DType::F32)?,
&k.contiguous()?.to_dtype(DType::F32)?,
&v.contiguous()?.to_dtype(DType::F32)?,
attention_mask
.map(|m| m.to_dtype(DType::F32).unwrap())
.as_ref(),
&q.contiguous()?,
&k.contiguous()?,
&v.contiguous()?,
attention_mask,
None,
&self.sdpa_params,
)?
Expand Down

0 comments on commit 5b627f1

Please sign in to comment.