Original production audio
Useful baseline: noisy, harder to parse, but preserves the true performance.
Short, focused A/B tests for the ugliest dialogue moments we have processed so far. Every section includes the original/raw production audio alongside the strongest local rescue, AI variants where relevant, and the final recommended choice.
The page pauses other players automatically when you play a new one. Use headphones if possible. The waveforms are for shape/energy reference, not loudness-normalized mastering judgments.
This clip is the clearest example of local cleanup beating AI. Two hard moments are shown here: an early dialogue section and a later noisier phrase. In both cases, the local winner held up better than ElevenLabs.
Useful baseline: noisy, harder to parse, but preserves the true performance.
More intelligible without collapsing the voice. This is the recommended result for this phrase.
Technically isolated, but less convincing and less successful than the local pass on this material.
The later phrase is noisier and more unstable, which makes it a strong stress test.
Still the strongest compromise of noise reduction and intelligibility.
This is the type of phrase where full confidence in AI can backfire. It cleans the wrong things and loses speech quality.
This is the strongest proof that phrase-level AI can still be useful. The local pass wins overall on the full clip, but the opening 18 seconds benefit significantly from phrase-level ElevenLabs isolation. The hybrid patch is therefore the recommended practical result.
The opening is rough enough that it benefits from direct phrase-level testing.
Strong overall cleanup, and still the best backbone for the complete take.
On this exact opening phrase, ElevenLabs does help. This is the rare case where the AI phrase pass beats the local version.
ElevenLabs for the first 18 seconds, then local cleanup for the rest. This is the smartest production version so far.
This clip pushed us back toward disciplined local processing. The strongest result came from the local full rescue on track 2. A smoother alternate exists, but the final remains the most powerful cleanup. The original is included for direct OG comparison.
This opening phrase was rough enough to test both local and AI rescue options.
The best full rescue we got for 3084. Stronger and clearer than the raw or alternate version.
Smoother and slightly less aggressive. Good if editorial prefers less processing character.
The longer reference makes it easier to hear room tone and pacing differences.
This is still the recommended version when the goal is maximum rescue.
Useful if the director prefers a less assertive cleanup profile.
The compare page exists to make one thing obvious: there is no universal winner. But there is a universal method.
Start with the original. Build the strongest local full-clip rescue. Then test models only on the ugliest tiny phrases. Patch back only the clear winners.
Without the raw reference, it is too easy to confuse “different” with “better.” Keeping the original on the page forces honest listening and makes the tradeoffs visible to everyone on the team.