Emmanuel Schmidbauer
d59ddbaeae
Fix critical thread safety issues
2026-01-09 11:23:19 -05:00
Quentin Fuxa
719e8b1a20
adapt online for mlx detection
2024-11-25 23:52:00 +01:00
Quentin Fuxa
f1b47178d8
adapt online for mlx detection
2024-11-25 23:52:00 +01:00
Quentin Fuxa
59db08e961
loader for full mlx
2024-11-25 23:52:00 +01:00
Quentin Fuxa
6fc20b9562
new dec class
2024-11-21 23:52:00 +01:00
Quentin Fuxa
fac8659161
uses native mlx function for attention
2024-11-21 23:52:00 +01:00
Quentin Fuxa
82cd24bb75
LoRa path v0 - functional
2025-11-29 17:21:10 +01:00
Quentin Fuxa
d45c397c6a
simulstreaming: limit n tokens to prevent hallucinations
2025-11-28 21:41:19 +01:00
Quentin Fuxa
1d88ba9d69
Fixes #294 . improve model path backend detection and file extraction
2025-11-27 23:14:00 +01:00
Quentin Fuxa
7faa21f95f
alignatt: enable model sharing by removing hooks and centralizing session state. Solves #282
...
Co-authored-by: Emmanuel Schmidbauer <eschmidbauer@gmail.com >
2025-11-25 23:07:42 +01:00
Quentin Fuxa
4e9f951551
correct silences handling when language not auto
2025-11-20 11:20:00 +01:00
Quentin Fuxa
870141298c
isort
2025-11-23 11:20:00 +01:00
Quentin Fuxa
6206fff118
0.2.15
2025-11-21 23:52:00 +01:00
Quentin Fuxa
9a45ec221c
internal rework 1
2025-11-20 12:58:38 +01:00
Quentin Fuxa
e9b4ceeee5
Add audio partial silence in chunks handling. bump to 0.2.14.post3
2025-11-17 22:52:00 +01:00
Quentin Fuxa
437641fb43
reduce min-chunk-size to 0.1, set default model to base
2027-04-25 23:52:00 +02:00
Quentin Fuxa
bbd4fd6cff
Merge branch 'improve_EOS_handling'
2025-11-16 22:30:31 +01:00
Quentin Fuxa
28985962a0
Silence handling: finish transcription even if not validated at the BEGINNING of the silence
2025-11-16 22:29:08 +01:00
Quentin Fuxa
a38c103fcd
simulstreaming coreml encoder compatibility
2025-11-16 21:24:14 +01:00
Quentin Fuxa
80b77998f9
Refactor backend handling
2025-11-15 19:51:41 +01:00
Quentin Fuxa
d310f7e25f
hf compatibility
2025-11-15 18:34:19 +01:00
Quentin Fuxa
16461052ed
task to direct-english-translation
2025-11-10 13:20:26 +01:00
Quentin Fuxa
13401ffe24
whisper core at root of wlk
2025-11-10 12:17:18 +01:00
Quentin Fuxa
7108d2ddc5
fixes https://github.com/QuentinFuxa/WhisperLiveKit/issues/269
2025-11-09 20:08:18 +01:00
Quentin Fuxa
a732e0903e
Add a script to detect alignement heads, usefull for distilled whisper
2025-11-09 18:12:09 +01:00
Quentin Fuxa
0491681be4
Distilled model compatibility with HF config.json to ModelDimensions
2025-11-08 20:20:05 +01:00
Quentin Fuxa
ffe5284764
_processing_tasks_done checks task completion
2025-11-05 23:34:00 +01:00
Quentin Fuxa
ece02db6a3
Use optional new separate NLLW package for translation
2025-10-30 19:36:28 +01:00
Quentin Fuxa
65250db92c
tensor to list at the stream end
2025-10-26 16:40:12 +01:00
Quentin Fuxa
416dce7975
fixes #261
...
Co-authored-by: yosagi <11404771+yosagi@users.noreply.github.com >"
2025-10-25 14:20:08 +02:00
Quentin Fuxa
714fb3b14a
custom faster-whisper/mlx whisper encoder available
2025-10-23 20:33:17 +02:00
Quentin Fuxa
0af379c465
DOC: information about file format
2025-10-23 20:32:05 +02:00
Quentin Fuxa
a7db39d999
solves incorrect spacing in buffer diarization
2025-10-02 23:04:00 +02:00
Quentin Fuxa
d080d675a8
cutom alignment heads parameter for custom models
2025-09-27 11:04:00 +02:00
Quentin Fuxa
8cbaeecc75
cutom alignment heads parameter for custom models
2025-09-27 11:04:00 +02:00
google-labs-jules[bot]
70e854b346
feat: Allow loading fine-tuned models in simulstreaming
...
This change modifies the `simulstreaming` backend to support loading fine-tuned Whisper models via the `--model_dir` argument.
The `SimulStreamingASR` class has been updated to:
- Use the `model_dir` path directly to load the model, which is the correct procedure for fine-tuned `.pt` files.
- Automatically disable the `faster-whisper` and `mlx-whisper` fast encoders when `model_dir` is used, as they are not compatible with standard fine-tuned models.
The call site in `core.py` already passed the `model_dir` argument, so no changes were needed there. This change makes the `simulstreaming` backend more flexible and allows users to leverage their own custom models.
2025-09-27 07:29:30 +00:00
Quentin Fuxa
b22478c0b4
correct silences handling when language not auto
2025-09-25 23:20:00 +02:00
Quentin Fuxa
4dd5d8bf8a
translation compatible with auto and detected language
2025-09-22 11:20:00 +02:00
Quentin Fuxa
93f002cafb
language detection after few seconds working
2025-09-20 11:08:00 +02:00
Quentin Fuxa
674b20d3af
in buffer while language not detected »
2025-09-21 11:05:00 +02:00
Quentin Fuxa
a5503308c5
O(n) to O(1) for simulstreaming timestamp determination
2025-09-21 11:04:00 +02:00
Quentin Fuxa
426d70a790
simulstreaming infer does not return a dictionary anymore
2025-09-21 11:03:00 +02:00
Quentin Fuxa
add7ea07ee
translator takes all the tokens from the queue
2025-09-09 19:55:39 +02:00
Quentin Fuxa
3358877054
Fix StorageView conversion for CPU/GPU compatibility
2025-09-09 15:44:16 +02:00
Quentin Fuxa
1f7798c7c1
condition on encoder_feature_ctranslate type
2025-09-09 12:16:52 +02:00
Alexander Lindberg
c7b3bb5e58
Fix regression with faster-whisper encoder_feature
2025-09-09 11:18:55 +03:00
Quentin Fuxa
334b338ab0
use platform to determine system and recommand mlx whisper
2025-09-07 15:49:11 +02:00
notV3NOM
abd8f2c269
Fix exponentially growing simulstreaming silence timer
2025-09-04 21:49:07 +05:30
Quentin Fuxa
f3ad4e39e4
torch.Tensor to torch.as_tensor
2025-09-04 16:39:11 +02:00
Quentin Fuxa
e0a5cbf0e7
v0.1.0 chrome extension
2025-09-04 16:36:28 +02:00