Quentin Fuxa
e9b4ceeee5
Add audio partial silence in chunks handling. bump to 0.2.14.post3
2025-11-17 22:52:00 +01:00
Quentin Fuxa
437641fb43
reduce min-chunk-size to 0.1, set default model to base
2027-04-25 23:52:00 +02:00
Quentin Fuxa
bfd60b3921
Add audio partial silence in chunks handling. bump to 0.2.14.post2
2025-11-17 22:52:00 +01:00
Quentin Fuxa
1e67bf97f0
improve buffering when use of heavy models
2027-04-25 23:52:00 +02:00
Quentin Fuxa
bbd4fd6cff
Merge branch 'improve_EOS_handling'
2025-11-16 22:30:31 +01:00
Quentin Fuxa
28985962a0
Silence handling: finish transcription even if not validated at the BEGINNING of the silence
2025-11-16 22:29:08 +01:00
Quentin Fuxa
a38c103fcd
simulstreaming coreml encoder compatibility
2025-11-16 21:24:14 +01:00
Quentin Fuxa
4d2ffb24f8
coreml conversion
2025-11-16 19:11:43 +01:00
Quentin Fuxa
1bbbb7903c
lora loader in shared whisper core
2025-11-16 18:44:35 +01:00
Quentin Fuxa
80b77998f9
Refactor backend handling
2025-11-15 19:51:41 +01:00
Quentin Fuxa
d310f7e25f
hf compatibility
2025-11-15 18:34:19 +01:00
Quentin Fuxa
8d9be88fe6
translation buffer is now displayed in frontend
2025-11-10 15:22:26 +01:00
Quentin Fuxa
16461052ed
task to direct-english-translation
2025-11-10 13:20:26 +01:00
Quentin Fuxa
5491dbd824
last_validated_token handled in state
2025-11-10 13:18:52 +01:00
Quentin Fuxa
13401ffe24
whisper core at root of wlk
2025-11-10 12:17:18 +01:00
Quentin Fuxa
7108d2ddc5
fixes https://github.com/QuentinFuxa/WhisperLiveKit/issues/269
2025-11-09 20:08:18 +01:00
Quentin Fuxa
a732e0903e
Add a script to detect alignement heads, usefull for distilled whisper
2025-11-09 18:12:09 +01:00
Quentin Fuxa
0491681be4
Distilled model compatibility with HF config.json to ModelDimensions
2025-11-08 20:20:05 +01:00
Quentin Fuxa
ffe5284764
_processing_tasks_done checks task completion
2025-11-05 23:34:00 +01:00
Quentin Fuxa
06b31f51eb
exception when translation and no nllw
2025-10-30 23:30:19 +01:00
Quentin Fuxa
ece02db6a3
Use optional new separate NLLW package for translation
2025-10-30 19:36:28 +01:00
Quentin Fuxa
939a7ebf8b
Translation Local Agreement + Cache optimization v0. Not connected yet
2025-10-28 00:16:52 +01:00
Quentin Fuxa
61edb70fff
audioProcessor state variables are now uniquely in State dataclass
2025-10-26 18:54:47 +01:00
Quentin Fuxa
4e455b8aab
translation now separates validated from output buffer tokens
2025-10-26 18:51:09 +01:00
Quentin Fuxa
9434390ad3
simplify task stopping condition
2025-10-26 17:26:43 +01:00
Quentin Fuxa
65250db92c
tensor to list at the stream end
2025-10-26 16:40:12 +01:00
Quentin Fuxa
416dce7975
fixes #261
...
Co-authored-by: yosagi <11404771+yosagi@users.noreply.github.com >"
2025-10-25 14:20:08 +02:00
Quentin Fuxa
0c5365e7c6
fixes #258
2025-10-24 20:51:16 +02:00
Quentin Fuxa
e7b05b0138
migration to silero vad v6: supports onnx
2025-10-23 23:52:00 +02:00
Quentin Fuxa
714fb3b14a
custom faster-whisper/mlx whisper encoder available
2025-10-23 20:33:17 +02:00
Quentin Fuxa
0af379c465
DOC: information about file format
2025-10-23 20:32:05 +02:00
Quentin Fuxa
1f684cdd97
fixes #251
2025-10-06 19:53:27 +02:00
Quentin Fuxa
9b1e061b32
forwarded_allow_ips in core
2025-10-04 23:04:00 +02:00
Quentin Fuxa
b4abc158b9
Merge pull request #249 from Damrod/add-ip-forwarding-support
...
fix wss for reverse proxying
2025-10-06 10:20:05 +02:00
Alvaro Ollero
3736458503
Uvicorn exposes a configuration option to enable reverse proxying from a trusted ip. This PR exposes it downstreams to end clients
2025-10-04 22:21:06 +02:00
Quentin Fuxa
374618e050
token speakers are only reattributed for token coming after last_validated_token
2025-10-04 09:52:00 +02:00
Quentin Fuxa
543972ef38
fixes #248
2025-10-04 09:52:00 +02:00
Quentin Fuxa
a7db39d999
solves incorrect spacing in buffer diarization
2025-10-02 23:04:00 +02:00
Quentin Fuxa
a153e11fe0
update when self.diarization_before_transcription
2025-09-28 11:04:00 +02:00
Quentin Fuxa
ca6f9246cc
force language = en for .en models
2025-09-28 11:04:00 +02:00
Quentin Fuxa
d080d675a8
cutom alignment heads parameter for custom models
2025-09-27 11:04:00 +02:00
Quentin Fuxa
40bff38933
Merge pull request #239 from msghik/feature/fine-tuned-model-support
...
feat: Allow loading fine-tuned models in simulstreaming
2025-09-29 10:08:26 +02:00
Quentin Fuxa
2fe3ca0188
connect source to output destination when used as chrome extension to keep audio playing
2025-09-27 13:59:44 +02:00
Quentin Fuxa
545ea15c9a
ensure buffer size to be a multiple of the element size
2025-09-27 13:58:32 +02:00
Quentin Fuxa
8cbaeecc75
cutom alignment heads parameter for custom models
2025-09-27 11:04:00 +02:00
google-labs-jules[bot]
70e854b346
feat: Allow loading fine-tuned models in simulstreaming
...
This change modifies the `simulstreaming` backend to support loading fine-tuned Whisper models via the `--model_dir` argument.
The `SimulStreamingASR` class has been updated to:
- Use the `model_dir` path directly to load the model, which is the correct procedure for fine-tuned `.pt` files.
- Automatically disable the `faster-whisper` and `mlx-whisper` fast encoders when `model_dir` is used, as they are not compatible with standard fine-tuned models.
The call site in `core.py` already passed the `model_dir` argument, so no changes were needed there. This change makes the `simulstreaming` backend more flexible and allows users to leverage their own custom models.
2025-09-27 07:29:30 +00:00
Quentin Fuxa
d55490cd27
typo and simpler conditions
2025-09-26 20:38:26 +02:00
Quentin Fuxa
b22478c0b4
correct silences handling when language not auto
2025-09-25 23:20:00 +02:00
Quentin Fuxa
94c34efd90
chrome extension ws default to localhost
2025-09-25 23:04:00 +02:00
Quentin Fuxa
9fc6654a4a
common frontend for web/ and chrome extension
2025-09-25 23:14:25 +02:00