SoafAii SAS Confidential Phase 1 Proof of Concept

SyncSpeak Engine

Audio generates its own synchronization data — the closed loop between speech synthesis and self-transcription.

Family 1 · TTS Self-Sync
Endpoint · /tts-sync
Worker · csit-brain
IP before demo

Input · Text to Synthesize

Step 1 of 5
01
TTS synth
02
Whisper ASR
03
Word timestamps
04
Audio playback
05
Synced reveal

Stage · Progressive Reveal

requestAnimationFrame @ audio.currentTime
Awaiting synthesis
Idle
0.00 / 0.00 s

Pipeline Log

Instrumentation
00:00.000SyncSpeak demo initialized. Ready.
The novelty, made visible. Words on the stage are not rendered from the input text above — they are rendered from Whisper's self-transcription of the AI audio, and they appear exactly when audio.currentTime reaches each word's start. No pre-existing alignment. No human timing. The audio creates its own sync data. Family 1 claim scope.

Latency Budget

Per request
TTS synth
ms
Whisper ASR
ms
Round trip
ms
Audio duration
s
Word count
Sync offset
ms

Whisper Timestamps

Generated by audio
No data — run synthesis to populate