Skip to content

Sync with upstream#6

Open
AIWintermuteAI wants to merge 199 commits intoAIWintermuteAI:mainfrom
collabora:main
Open

Sync with upstream#6
AIWintermuteAI wants to merge 199 commits intoAIWintermuteAI:mainfrom
collabora:main

Conversation

@AIWintermuteAI
Copy link
Copy Markdown
Owner

No description provided.

fraic and others added 30 commits March 25, 2024 19:11
Improve cpu and gpu Dockerfiles, resulting in much smaller images
Add option: save network stream to local file while transcribing
… OMP_NUM_THREADS

Signed-off-by: makaveli10 <vineet.suryan@collabora.com>
fix: limit CPU usage for VAD onnxruntime inference session by setting…
Signed-off-by: makaveli10 <suryanvineet47@gmail.com>
Signed-off-by: makaveli10 <suryanvineet47@gmail.com>
Signed-off-by: makaveli10 <suryanvineet47@gmail.com>
Signed-off-by: makaveli10 <suryanvineet47@gmail.com>
Signed-off-by: makaveli10 <suryanvineet47@gmail.com>
Signed-off-by: makaveli10 <suryanvineet47@gmail.com>
…optional

Make writing audio frames optional
- Use a threadlock around the model in single model mode
Signed-off-by: makaveli10 <vineet.suryan@collabora.com>
Signed-off-by: makaveli10 <vineet.suryan@collabora.com>
Signed-off-by: makaveli10 <vineet.suryan@collabora.com>
Expose the srt file location of Transcription client
fix spelling of detection in README.md.
aaron-boxer and others added 30 commits February 5, 2026 22:31
README.md: add instructions for running client
Add `--enable-timestamps` option to `run_client.py`
script to print out transcripted text with timestamps.

Sample output with translation enabled:
```
[0.000 -> 7.440]  And so, my fellow Americans, ask not what your country can do for you.
[7.440 -> 10.300]  Ask what you can do for your country.

TRANSLATION to fr:
[0.000 -> 7.440] Et donc, mes camarades américains, ne demandez pas ce que votre pays peut faire pour vous.
[7.440 -> 10.300] Demandez ce que vous pouvez faire pour votre pays.
```

Signed-off-by: Jeny Sadadia <jeny.sadadia@collabora.com>
Enable timestamps for transcripted text
feat: update to support faster whisper 1.2.0
Resolves pkg_resources missing during wheel build

Signed-off-by: makaveli10 <vineet.suryan@collabora.com>
Bump openai-whisper version to 20250625.
Replace hardcoded [-4:] truncation with a configurable display_segments
parameter (default: 4) in both Client and TranscriptionClient classes.

Fixes #377
Add cross-client GPU batch inference for faster_whisper backend
When VAD removes all speech from an audio chunk, transcriber.transcribe() returns (None, info). Calling list(None) raises TypeError. The _process_multi path already handles this case; this aligns _process_single to match.
Fix NoneType crash in _process_single when VAD filters all audio
feat: make display_segments configurable in Client/TranscriptionClient
Signed-off-by: makaveli10 <vineet.suryan@collabora.com>
Expose __version__ in package root and update dependencies in setup.py
Fix crash when no --files provided; use microphone input instead
fix: render transcript text safely in browser extensions
these new test cover issues such as thread safety, VAD thresholding,
message routing, error handling etc. that weren't covered by existing
tests. Mocking is used to avoid dependencies on GPU, ONNX etc.
- All ClientManager methods (add_client, get_client, remove_client,
  get_wait_time, is_server_full, is_client_timeout) now protected by
  a threading.Lock
- cleanup() called outside the lock to avoid holding it during I/O
- is_server_full() computes wait time inline under lock instead of
  calling get_wait_time() to avoid nested lock acquisition
- Added concurrent thread safety tests for add/remove and get operations
CI: expand test suite coverage
  audio: add support for raw pcm input via server flag
Add thread safety to client manager with threading lock
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.