generated from nhcarrigan/template
fix: lints
This commit is contained in:
Generated
+1
@@ -531,6 +531,7 @@ dependencies = [
|
||||
"tauri",
|
||||
"tauri-build",
|
||||
"tauri-plugin-opener",
|
||||
"tempfile",
|
||||
"thiserror 2.0.18",
|
||||
"tokio",
|
||||
"tracing",
|
||||
|
||||
@@ -67,6 +67,9 @@ windows = { version = "0.62", features = [
|
||||
"Win32_UI_Shell_PropertiesSystem",
|
||||
] }
|
||||
|
||||
[dev-dependencies]
|
||||
tempfile = "3.10"
|
||||
|
||||
[patch.crates-io]
|
||||
# Fix cross-compilation: use CARGO_CFG_TARGET_OS instead of cfg!(windows)
|
||||
llama-cpp-sys-2 = { path = "../patches/llama-cpp-sys-2" }
|
||||
|
||||
@@ -487,7 +487,7 @@ fn check_ready(state: State<'_, AppState>) -> Result<bool, String> {
|
||||
|
||||
// At minimum, we need the summarizer loaded
|
||||
// Whisper can be loaded on first use
|
||||
let ready = summarizer.as_ref().map_or(false, |s| s.is_loaded());
|
||||
let ready = summarizer.as_ref().is_some_and(|s| s.is_loaded());
|
||||
|
||||
Ok(ready)
|
||||
}
|
||||
|
||||
@@ -158,7 +158,7 @@ impl LlamaSummarizer {
|
||||
for i in 0..max_tokens {
|
||||
// Sample the next token using the sampler
|
||||
// The sampler.sample() takes context and the index of the last token in the batch
|
||||
let token = sampler.sample(&ctx, (batch.n_tokens() - 1) as i32);
|
||||
let token = sampler.sample(&ctx, batch.n_tokens() - 1);
|
||||
|
||||
// Check for end of generation
|
||||
if model.is_eog_token(token) {
|
||||
|
||||
@@ -415,6 +415,7 @@ mod tests {
|
||||
start_sample: 16000,
|
||||
end_sample: 32000,
|
||||
speaker_id: 0,
|
||||
fingerprint: None,
|
||||
};
|
||||
|
||||
assert!((segment.start_seconds() - 1.0).abs() < 0.001);
|
||||
|
||||
@@ -37,9 +37,7 @@
|
||||
"installerIcon": "icons/icon.ico"
|
||||
}
|
||||
},
|
||||
"resources": [
|
||||
"resources/"
|
||||
],
|
||||
"resources": ["resources/"],
|
||||
"icon": [
|
||||
"icons/32x32.png",
|
||||
"icons/128x128.png",
|
||||
|
||||
Reference in New Issue
Block a user