Multimodal Weekly 47: SpiRit-LM, an Interleaved Spoken and Written Language Model
In the 47th session of Multimodal Weekly, we welcome Benjamin Muller, Tu-Anh Nguyen, and Bokai Yu from Meta AI to discuss their work Spirit-LM - which is designed to freely mix text and speech - allowing for a seamless integration of both modalities.
Abstract
We introduce SpiRit-LM, a foundation multimodal language model that freely mixes text and speech. Our model is based on a pretrained text language model that we extend to the speech modality by continuously training it on text and speech units. Speech and text sequences are concatenated as a single set of tokens, and trained with a word-level interleaving method using a small automatically-curated speech-text parallel corpus. SpiRit-LM comes in two versions: a Base version that uses speech semantic units and an Expressive version that models expressivity using pitch and style units in addition to the semantic units. For both versions, the text is encoded with subword BPE tokens. The resulting model displays both the semantic abilities of text models and the expressive abilities of speech models. Additionally, we demonstrate that SpiRit-LM is able to learn new tasks in a few-shot fashion across modalities (i.e. ASR, TTS, Speech Classification).
Resources
Project Page: https://speechbot.github.io/spiritlm/
Paper: https://arxiv.org/abs/2402.05755
Join the Multimodal Minds community to connect with the speakers!
Multimodal Weekly is organized by Twelve Labs, a startup building multimodal foundation models for video understanding. Learn more about Twelve Labs here: https://twelvelabs.io/