Timestamped Audio Captioning
Sonal Kumar et al.
Large Audio Language Models struggle to disentangle overlapping events in complex acoustic scenes, yielding temporally inconsistent captions and frequent hallucinations. We introduce Timestamped Audio Captioner (TAC), a model that produces temporally grounded audio descriptions at varying degrees of detail and resolution. TAC is trained with a synthetic data pipeline that constructs challenging and dynamic mixtures from real-world audio sources, enabling robust learning under realistic polyphonic conditions. Across event detection and dense captioning, TAC outperforms all competing methods, with a low hallucination rate and accurate temporal grounding. We also introduce TAC-V, an audio-visual pipeline to generate semantically rich audio-visual descriptions. We then show that TAC and TAC-V serves as a "semantic bridge" for a text-only reasoner: a simple TAC→LLM and TAC-V→LLM cascade achieves state-of-the-art scores on benchmarks for both audio (MMAU-Pro, MMSU, MMAR) and audio-visual (Daily-Omni, VideoHolmes) understanding and reasoning respectively.
https://lnkd.in/e3SMEU9X
I’m Zaki from FileMarket Labs Inc. — a company that specializes in delivering AI solutions, speech & video data collection, and custom machine-learning projects tailored to each client’s needs. We efficiently handle speech and video recording projects for AI training with strict quality controls and timely delivery. Would you be open to a quick chat about how we can support your AI goals? https://data.filemarket.ai/