Papers
arxiv:2601.01554

MOSS Transcribe Diarize: Accurate Transcription with Speaker Diarization

Published on Jan 4
Β· Submitted by
Zhaoye Fei
on Jan 7
#3 Paper of the day
Authors:
,
,
,
,
,
,
,
,
,
,
,
,
Zhe Xu ,
,
,
,

Abstract

A unified multimodal large language model for end-to-end speaker-attributed, time-stamped transcription with extended context window and strong generalization across benchmarks.

AI-generated summary

Speaker-Attributed, Time-Stamped Transcription (SATS) aims to transcribe what is said and to precisely determine the timing of each speaker, which is particularly valuable for meeting transcription. Existing SATS systems rarely adopt an end-to-end formulation and are further constrained by limited context windows, weak long-range speaker memory, and the inability to output timestamps. To address these limitations, we present MOSS Transcribe Diarize, a unified multimodal large language model that jointly performs Speaker-Attributed, Time-Stamped Transcription in an end-to-end paradigm. Trained on extensive real wild data and equipped with a 128k context window for up to 90-minute inputs, MOSS Transcribe Diarize scales well and generalizes robustly. Across comprehensive evaluations, it outperforms state-of-the-art commercial systems on multiple public and in-house benchmarks.

Community

Paper author Paper submitter

MOSS Transcribe Diarize πŸŽ™οΈ

We introduce MOSS Transcribe Diarize β€” a unified multimodal model for Speaker-Attributed, Time-Stamped Transcription (SATS).

πŸ” End-to-end SATS in a single pass (transcription + speaker attribution + timestamps)
🧠 128k context window for up to ~90-minute audio without chunking (strong long-range speaker memory)
🌍 Trained on extensive in-the-wild conversations + controllable simulated mixtures (robust to overlap/noise/domain shift)
πŸ“Š Strong results on AISHELL-4 / Podcast / Movies benchmarks (best cpCER / Ξ”cp among evaluated systems)

Paper: [2601.01554] MOSS Transcribe Diarize: Accurate Transcription with Speaker Diarization
Homepage: https://mosi.cn/models/moss-transcribe-diarize
Online Demo: https://moss-transcribe-diarize-demo.mosi.cn

Β·

Some very interesting work. Very much hoping for at least an open-weight release 🀞.

P.s. In case you were not already aware, your website is currently down.

Verycool & useful work! Is the model going to be open source?

Β·

Thank you for your interest, and we plan to open-source it in the coming months.

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2601.01554 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2601.01554 in a dataset README.md to link it from this page.

Spaces citing this paper 1

Collections including this paper 2