Quentin Fuxa
9b2c3ee844
docs: update README with voxtral backend, benchmarks, testing sections
...
- Add Voxtral Backend section explaining voxtral-mlx and voxtral (HF).
- Add Testing & Benchmarks section with commands to run tests/benchmarks.
- Update --backend parameter docs to include voxtral-mlx and voxtral.
- Update optional dependencies table with Voxtral entry.
- Link to BENCHMARK.md for detailed performance comparisons.
2026-02-22 23:27:57 +01:00
Quentin Fuxa
3c15246fc0
mixstral hf v0
2026-02-20 20:49:57 +01:00
Quentin Fuxa
7f3a3df620
simulstreaming mlx & torch dedup of common base
2025-02-15 23:52:00 +01:00
Emmanuel Schmidbauer
d59ddbaeae
Fix critical thread safety issues
2026-01-09 11:23:19 -05:00
Quentin Fuxa
f1b47178d8
adapt online for mlx detection
2024-11-25 23:52:00 +01:00
Quentin Fuxa
2431a6bf91
isolated VAD states per user: .onnx: share a stateless model. .jit: require duplicating the model.
...
Co-authored-by: eschmidbauer <eschmidbauer@gmail.com >
2025-12-05 15:27:14 +01:00
Quentin Fuxa
82cd24bb75
LoRa path v0 - functional
2025-11-29 17:21:10 +01:00
Quentin Fuxa
7faa21f95f
alignatt: enable model sharing by removing hooks and centralizing session state. Solves #282
...
Co-authored-by: Emmanuel Schmidbauer <eschmidbauer@gmail.com >
2025-11-25 23:07:42 +01:00
Quentin Fuxa
870141298c
isort
2025-11-23 11:20:00 +01:00
Quentin Fuxa
437641fb43
reduce min-chunk-size to 0.1, set default model to base
2027-04-25 23:52:00 +02:00
Quentin Fuxa
80b77998f9
Refactor backend handling
2025-11-15 19:51:41 +01:00
Quentin Fuxa
16461052ed
task to direct-english-translation
2025-11-10 13:20:26 +01:00
Quentin Fuxa
13401ffe24
whisper core at root of wlk
2025-11-10 12:17:18 +01:00
Quentin Fuxa
06b31f51eb
exception when translation and no nllw
2025-10-30 23:30:19 +01:00
Quentin Fuxa
ece02db6a3
Use optional new separate NLLW package for translation
2025-10-30 19:36:28 +01:00
Quentin Fuxa
e7b05b0138
migration to silero vad v6: supports onnx
2025-10-23 23:52:00 +02:00
Quentin Fuxa
9b1e061b32
forwarded_allow_ips in core
2025-10-04 23:04:00 +02:00
Quentin Fuxa
ca6f9246cc
force language = en for .en models
2025-09-28 11:04:00 +02:00
Quentin Fuxa
8cbaeecc75
cutom alignment heads parameter for custom models
2025-09-27 11:04:00 +02:00
Quentin Fuxa
4dd5d8bf8a
translation compatible with auto and detected language
2025-09-22 11:20:00 +02:00
Quentin Fuxa
65025cc448
nllb backend can be transformers, and model size can be 1.3B
2025-09-17 10:20:31 +02:00
Quentin Fuxa
bbba1d9bb7
add nllb-backend and translation perf test in dev_notes
2025-09-16 20:45:01 +02:00
Quentin Fuxa
5aa312e437
simulstreaming warmup is done in whisperlivekit.simul_whisper.backend.load_model, not in warmup_online
2025-09-13 20:19:19 +01:00
notV3NOM
ebaf36a8be
Fix warmup file behavior
2025-09-13 20:44:24 +05:30
Quentin Fuxa
a4e9f3cab7
support for raw PCM input option by @YeonjunNotFR
2025-09-11 21:32:11 +02:00
Quentin Fuxa
b06866877a
add --disable-punctuation-split option
2025-09-11 21:03:00 +02:00
Quentin Fuxa
add7ea07ee
translator takes all the tokens from the queue
2025-09-09 19:55:39 +02:00
Quentin Fuxa
f661f21675
translation asyncio task
2025-09-08 18:34:31 +02:00
Quentin Fuxa
72f33be6f2
translation: use of get_nllb_code
2025-09-07 15:25:14 +02:00
Quentin Fuxa
d1a9913c47
nllb v0
2025-09-05 18:02:42 +02:00
Quentin Fuxa
3bd2122eb4
0.2.8 : only the decoder of whisper is loaded in memory when a different encoder is used
2025-09-02 21:12:25 +02:00
Quentin Fuxa
b101ce06bd
several users share the same sortformer model instance
2024-08-24 19:18:00 +02:00
Quentin Fuxa
5258305745
default diarization backend in now sortformer
2025-08-24 18:32:01 +02:00
Quentin Fuxa
58297daf6d
sortformer diar implementation v0.3
2025-08-24 18:32:01 +02:00
Quentin Fuxa
9feb0e597b
remove VACOnlineASRProcessor backend possibility
2025-08-20 20:57:43 +02:00
Quentin Fuxa
253a080df5
diart diarization handles pauses/silences thanks to offset
2025-08-19 21:12:55 +02:00
Quentin Fuxa
e14bbde77d
sortformer diar implementation v0
2025-08-19 17:02:55 +02:00
Quentin Fuxa
2699b0974c
Fix simulstreaming imports
2025-08-19 14:43:54 +02:00
Quentin Fuxa
820f92d8cb
audio_max_len to 30 -> 20, ffmpeg timeout 5 -> 20
2025-08-17 09:32:08 +02:00
Quentin Fuxa
7fe0353260
vac model is loaded in TranscriptionEngine, and by default
2025-08-17 00:34:25 +02:00
Quentin Fuxa
1652db9a2d
Use distinct backend models for simulstreaming and add --preloaded_model_count to preload them
2025-08-15 23:03:55 +02:00
Quentin Fuxa
15c3df1cba
warmup base whisper when using simulstreaming
2025-08-12 18:52:52 +02:00
Quentin Fuxa
728e1f1290
simulstreaming warmup is done for each instance of online, not for the backend
2025-08-12 18:35:04 +02:00
Quentin Fuxa
d098af3185
each SimulStreamingOnlineProcessor now contains PaddedAlignAttWhisper instance. SimulStreamingASR only contains loaded whisper model
2025-08-11 08:24:14 +02:00
Quentin Fuxa
2bbdc70187
lags are now updated every 0.1s
2025-08-09 23:11:05 +02:00
Quentin Fuxa
5491964e81
clean SimulStreamingOnlineProcessor initialization + audio processing
2025-08-09 20:16:27 +02:00
Quentin Fuxa
197293e25e
refactor(simulstreaming): extract backend + online module into separate files from whisper streaming
2025-08-08 18:07:51 +02:00
Quentin Fuxa
6867041254
1rst version of SimulStreaming backend. many improvements needed
2025-06-25 17:59:46 +02:00
Quentin Fuxa
8532a91c7a
add segmentation and embedding model options to configuration
2025-06-19 16:29:25 +02:00
Quentin Fuxa
b01b81bad0
improve diarization with lag diarization substraction
2025-06-19 16:18:49 +02:00