Skip to content

Commit f344c42

Browse files
committed
style: fmt
1 parent 5b4caba commit f344c42

File tree

8 files changed

+9
-9
lines changed

8 files changed

+9
-9
lines changed

crates/model/src/components/ssm.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
//! It can be extended to support different SSM variants.
55
66
use candle_core::{DType, Module, Result as CandleResult, Tensor};
7-
use candle_nn::{conv1d, Conv1d, LayerNorm, Linear, VarBuilder};
7+
use candle_nn::{Conv1d, LayerNorm, Linear, VarBuilder, conv1d};
88

99
/// SSM Configuration
1010
#[derive(Clone, Debug)]

crates/model/src/gemma4/block.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ use crate::config::{LayerType, ModelConfig, RoPEConfig};
77
use crate::gemma4::attention::Gemma4Attention;
88
use crate::gemma4::mlp::GeGLU;
99
use candle_core::{Result, Tensor};
10-
use candle_nn::{linear, Linear};
10+
use candle_nn::{Linear, linear};
1111

1212
pub struct Gemma4Block {
1313
attention: Gemma4Attention,

crates/model/src/kernels/flash_attention.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -78,7 +78,7 @@ pub trait FlashAttention: Send + Sync {
7878
mask: &Tensor,
7979
) -> Result<Tensor>;
8080
fn forward_tiled(&self, q: &Tensor, k: &Tensor, v: &Tensor, tile_size: usize)
81-
-> Result<Tensor>;
81+
-> Result<Tensor>;
8282
}
8383

8484
#[derive(Clone, Default)]

crates/model/src/lib.rs

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,8 +15,8 @@ pub mod qwen3_config;
1515
pub mod tokenizer;
1616

1717
pub use kernels::{
18-
fused_attention_layer, fused_mlp_layer, AttentionVariant, CudaGraph, FlashAttention,
19-
FlashAttentionConfig, FlashAttentionKernel,
18+
AttentionVariant, CudaGraph, FlashAttention, FlashAttentionConfig, FlashAttentionKernel,
19+
fused_attention_layer, fused_mlp_layer,
2020
};
2121
pub use loader::{ModelLoader, ModelLoaderBuilder};
2222
pub use quantize::{QuantizationConfig, QuantizationFormat, QuantizedTensor, StorageTensor};

crates/model/src/loader/mod.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ pub mod builder;
77
pub mod format;
88

99
pub use builder::{ModelLoader, ModelLoaderBuilder};
10-
pub use format::{load_checkpoint, FormatLoader};
10+
pub use format::{FormatLoader, load_checkpoint};
1111

1212
use candle_core::{Device, Result, Tensor};
1313
use memmap2::Mmap;
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
pub mod quantization;
22
pub mod tensor_store;
33

4-
pub use quantization::{dequantize, quantize, QuantizedTensor};
4+
pub use quantization::{QuantizedTensor, dequantize, quantize};
55
pub use tensor_store::PagedKvCache;

crates/model/src/qwen3/rope.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
#![allow(clippy::all, unused)]
22

3-
pub use crate::components::{apply_rope, RoPE};
3+
pub use crate::components::{RoPE, apply_rope};
44

55
pub fn precompute_rope_cache(seq_len: usize, head_dim: usize, theta: f32) -> Vec<(f32, f32)> {
66
let mut cache = Vec::with_capacity(seq_len * head_dim / 2);

crates/model/src/qwen3_5/ssm.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
use std::collections::HashMap;
44

55
use candle_core::{DType, Module, Result as CandleResult, Tensor};
6-
use candle_nn::{conv1d, Conv1d, LayerNorm, Linear, VarBuilder};
6+
use candle_nn::{Conv1d, LayerNorm, Linear, VarBuilder, conv1d};
77
use thiserror::Error;
88

99
#[derive(Clone, Debug)]

0 commit comments

Comments
 (0)