Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
License: arXiv.org perpetual non-exclusive license
arXiv:2401.11053v1 [eess.AS] 19 Jan 2024

StreamVoice: Streamable Context-Aware Language Modeling for Real-time Zero-Shot Voice Conversion

Zhichao Wang11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT    Yuanzhe Chen22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT    Xinsheng Wang11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT    Zhuo Chen22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT    Lei Xie11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT111Corresponding author   
Yuping Wang22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT
   Yuxuan Wang22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT 11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPTAudio, Speech and Language Processing Group (ASLP@NPU)
School of Computer Science, Northwestern Polytechnical University, Xi’an, China
22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPTByteDance Inc.
Abstract

Recent language model (LM) advancements have showcased impressive zero-shot voice conversion (VC) performance. However, existing LM-based VC models usually apply offline conversion from source semantics to acoustic features, demanding the complete source speech, and limiting their deployment to real-time applications. In this paper, we introduce StreamVoice, a novel streaming LM-based model for zero-shot VC, facilitating real-time conversion given arbitrary speaker prompts and source speech. Specifically, to enable streaming capability, StreamVoice employs a fully causal context-aware LM with a temporal-independent acoustic predictor, while alternately processing semantic and acoustic features at each time step of autoregression which eliminates the dependence on complete source speech. To address the potential performance degradation from the incomplete context in streaming processing, we enhance the context-awareness of the LM through two strategies: 1) teacher-guided context foresight, using a teacher model to summarize the present and future semantic context during training to guide the model’s forecasting for missing context; 2) semantic masking strategy, promoting acoustic prediction from preceding corrupted semantic and acoustic input, enhancing context-learning ability. Notably, StreamVoice is the first LM-based streaming zero-shot VC model without any future look-ahead. Experimental results demonstrate StreamVoice’s streaming conversion capability while maintaining zero-shot performance comparable to non-streaming VC systems.

Refer to caption
Figure 1: The concept of the streaming zero-shot VC employing the widely used recognition-synthesis framework. StreamVoice is built on this popular paradigm.

1 Introduction

Voice conversion (VC) aims to transfer a speaker’s voice to that of another speaker without changing the linguistic content. Such technique has been deployed in many real-world applications, e.g., movie dubbing, privacy protection, and pronunciation correction etc. With the help of neural semantic features, such as the bottleneck feature (BNF) from an automatic speech recognition (ASR) system, converting source speech from arbitrary speakers in the wild has been successfully achieved Sun et al. (2016). Meanwhile, converting to an arbitrary target speaker with only one utterance of this speaker, which is the so-called zero-shot VC, also has been researched recently Qian et al. (2019); Wang et al. (2023c). However, most existing zero-shot VC models are designed for offline systems, which are insufficient to meet the recent growing demands of streaming capability in real-time VC applications, such as live broadcasting and real-time communication (RTC). In this study, we focus on the streaming zero-shot VC as illuminated in Fig. 1.

Disentangling speech into different components, e.g., semantic content and speaker timbre, plays an important role in the zero-shot VC task Chou and Lee (2019); Wang et al. (2023d, 2021); Qian et al. (2019). Recently, benefiting from the powerful LM framework and the scaling up of training data, LM-based VC models Wang et al. (2023c); Yang et al. (2023); Zhu et al. (2023) with built-in in-context learning ability can sense the context relations between source and target speaker’s utterances to capture fine-grained speaker timbre, achieving impressive zero-shot VC performance. However, demanding the complete source speech utterance limits these LM-based VC models in real-time scenarios, and thus these models can only be used in offline applications. While several non-LM-based methods Yang et al. (2022); Wang et al. (2023a) have been proposed for streaming zero-shot VC, the performance fails to generalize well to unseen speakers with high speaker similarity and speech naturalness, mainly due to the limited model capacity to scale up training data, and also the performance degradation caused by the missing future information in streaming scenario.

Inspired by the success of LM-based models in zero-shot VC, we aim to explore the feasibility of LMs for the streaming VC scenario. An intuitive way is to follow the popular recognize-synthesis framework as shown in Fig. 1, in which, speech is represented in semantic BNF and acoustic features respectively extracted by a streaming ASR and an audio codec. Then the LM-based VC model undertakes the transformation of semantic information into acoustic features with the target speaker’s timbre. However, the development of the LM-based model in streaming zero-shot VC is hampered by two primary challenges.

  • Streamable architecture: streaming models typically produce immediate output upon receiving current input without reliance on future time steps. Current LM-based VC models perform the conversion only when get a full-utterance of source speech, which fails to meet the demands of streaming applications. The widely adopted multi-stage language modeling for multi-layer codec prediction introduces complexity to system design, thereby posing a potential risk of cumulative errors. Additionally, the dependency models of the streaming pipeline also impact the design and performance of the VC model.

  • Performance gap: unlike non-streaming models, streaming models must process frame-wise or chunked input causally on the fly without future information, facing missing context and potential performance degradation. This missing hinders the streaming VC model from achieving high-quality conversion. In addition, as shown in Fig. 1, the VC model relies on the semantic feature BNF from ASR to achieve conversion, which makes semantic features very important. However, streaming ASR exhibits inferior performance compared to its non-streaming counterpart, leading to the extracted BNF carrying low-quality semantic information but more speaker information. In addition to the inherent unavailable future reception, this low-quality semantic input makes achieving high-quality conversion more difficult. The goal of zero-shot VC amplifies the challenges faced by our streaming VC model.

In this work, we propose StreamVoice, a streaming LM-based model for high-quality zero-shot VC, in which teacher-guided semantic foresight and semantic masking are integrated to enhance the context awareness of the model for improving conversion quality. Specifically, StreamVoice creates a streamable architecture by integrating a single-stage language model that casually generates acoustic codecs with the collaboration of an acoustic predictor. Alternating input of semantic and acoustic features at each time step ensures seamless streaming behavior. To mitigate the performance gap caused by missing contextual information, two methods are introduced to enhance the context-awareness of the LM. 1) We incorporate a teacher-guided semantic foresight, where the VC model is taught by a teacher non-streaming ASR model to infer the present and future semantic information summarized by the teacher, which is then used to enhance the acoustic prediction. 2) To enhance the context learning from the input history, semantic masking is used to encourage acoustic prediction from the preceding acoustic and corrupted semantic input, which also implicitly creates an information bottleneck to reduce the source speaker’s information.

Experiments demonstrate StreamVoice’s ability to convert speech in a streaming manner with high speaker similarity for both seen and unseen speakers while maintaining performance comparable to non-streaming VC systems. As the first LM-based zero-shot VC model without any future look-ahead, the total pipeline of StreamVoice only has 124 ms latency to perform the conversion, 2.4x faster than real-time on a single A100 GPU without engineering optimizations.

2 Related Works

2.1 Zero-shot Voice Conversion

Zero-shot VC imposes stringent demands on speech decoupling and capturing speaker timbre. Many studies specifically design many disentanglement approaches, incorporating intricate structures Chou and Lee (2019), loss functions Wang et al. (2021), and training strategies Ebbers et al. (2021), to achieve speech decoupling. Rather than embedding explicit disentanglement designs in VC training, some approaches Gu et al. (2021) leverage a speaker verification (SV) model for speaker representation, while linguistic content is extracted using ASR or self-supervised learning (SSL) models Sun et al. (2016); Choi et al. (2021). To enhance speaker timbre capturing, some fine-grained speaker modeling methods also have been explored Yin et al. (2021); Wang et al. (2023d). Recent successes of language models in generative tasks have prompted the exploration of LM-based models in zero-shot VC, yielding impressive results. Using the pre-trained model to decouple speech, the LM-based VC model Wang et al. (2023c); Yang et al. (2023); Zhu et al. (2023) can capture fine-grained speaker timbre from the speaker prompt and then perform the conversion. However, current LM-based VC models are inapplicable to streaming scenarios, constraining their real-world utility. This paper addresses this gap by investigating the zero-shot capabilities of language models specifically tailored for streaming scenarios.

2.2 Streaming Voice Conversion

Despite the high-quality conversion achieved by non-streaming VC models, their non-streamable structure and reliance on full-utterance input hamper them for real-time streaming applications. For streaming, causal processing and the structure of the streaming pipeline are crucial considerations. Streaming models are compelled to process frame-wise or chunked input on the fly, devoid of access to future information, leading to performance degradation compared to non-streaming counterparts. To address this, a common approach Hayashi et al. (2022); Kameoka et al. (2021); Ning et al. (2023) involves the integration of a teacher model to guide the training of the streaming model or the distillation of knowledge from a non-streaming model. Chen et al. Chen et al. (2023b) focus on selecting BNF with minimal semantic information loss through layer-wise analysis, while Chen et al. Chen et al. (2022) incorporate adversarial training to enhance the quality of semantic features. Beyond streaming VC, recently, there have also been some efforts towards streaming zero-shot VC. For instance, VQMIVC Wang et al. (2021), which is designed for the non-streaming application, is modified to be streamable by Yang et al. Yang et al. (2022). ALO-VC Wang et al. (2023a) constructs a streaming system using an SV model, a streaming PPG extractor, and a pitch extractor. However, current streaming zero-shot VC, designed for low-resource devices, has limited model capacity with poor generalization to unseen speakers, leading to inferior similarity and naturalness. Motivated by LM’s successes in zero-shot VC, we design a streamable LM in streaming scenarios. To tackle distinctive challenges in streaming VC, we introduce teacher-guided semantic foresight and semantic masking to enhance LM’s context awareness and improve conversion quality.

2.3 Language Model-based Speech Generation

In recent years, advancements in language models (LMs) within natural language processing have showcased potent generation capabilities, influencing the development of LMs in speech generation. By employing codec Zeghidour et al. (2021) or other SSL models Chung et al. (2021), speech and audio can be efficiently tokenized into discrete units, facilitating low-bitrate audio representation and semantic information extraction. This progress allows speech generation to seamlessly utilize LM frameworks. Taking audio generation as a conditional language modeling task, AudioLM Borsos et al. (2023) and MusicLM Agostinelli et al. (2023) employ hierarchical language modeling for acoustic prediction from coarse to fine units. VALL-E Wang et al. (2023b) and SpearTTS Kharitonov et al. (2023) extend LMs for zero shot-TTS, which can clone a human’s voice with prompt tokens from a short recording. For zero-shot VC, LM-VC Wang et al. (2023c) employs task-oriented optimizations to this task. And some studies Zhu et al. (2023); Yang et al. (2023) leverage multitask objectives and datasets, achieving high-quality conversion. Despite this progress, existing LM-based VC models usually apply offline processing, demanding complete utterance from the source speech, which hinders their suitability for real-time streaming applications. In contrast to prior studies, we explore the zero-shot capability of the LM-based VC for streaming scenarios. With the enhancement of context awareness, the proposed LM-based VC model achieves results comparable to non-streaming LM-based VC.

Refer to caption
Figure 2: The overall architecture for StreamVoice.

3 StreamVoice

3.1 Overview

As shown in Fig. 2, the development of StreamVoice follows the recognition-synthesis framework. In this framework, speech is first represented as semantic features 𝐬={s1,s2,sTs}𝐬subscript𝑠1subscript𝑠2subscript𝑠subscript𝑇𝑠\mathbf{s}=\{s_{1},s_{2},...s_{T_{s}}\}bold_s = { italic_s start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_s start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , … italic_s start_POSTSUBSCRIPT italic_T start_POSTSUBSCRIPT italic_s end_POSTSUBSCRIPT end_POSTSUBSCRIPT } and acoustic features 𝐚={a1,a2,,aTa}𝐚subscript𝑎1subscript𝑎2subscript𝑎subscript𝑇𝑎\mathbf{a}=\{a_{1},a_{2},...,a_{T_{a}}\}bold_a = { italic_a start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , … , italic_a start_POSTSUBSCRIPT italic_T start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT end_POSTSUBSCRIPT } by a pre-trained streaming ASR model and a speech codec model respectively. Here, Tssubscript𝑇𝑠T_{s}italic_T start_POSTSUBSCRIPT italic_s end_POSTSUBSCRIPT and Tasubscript𝑇𝑎T_{a}italic_T start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT denote the sequence length. Before inputting to StreamVoice, 𝐬𝐬\mathbf{s}bold_s and 𝐚𝐚\mathbf{a}bold_a are aligned to the same length T𝑇Titalic_T. StreamVoice incorporates a context-aware language model and an acoustic predictor to perform a single language modeling process. With the semantic and acoustic features {𝐬~,𝐚~}~𝐬~𝐚\{\mathbf{\tilde{s}},\mathbf{\tilde{a}}\}{ over~ start_ARG bold_s end_ARG , over~ start_ARG bold_a end_ARG } of speech from the target speaker as speaker prompt, the LM leverages the semantic information 𝐚1:tsubscript𝐚:1𝑡\mathbf{a}_{1:t}bold_a start_POSTSUBSCRIPT 1 : italic_t end_POSTSUBSCRIPT of source speech to autoregressively predict the hidden output 𝐡csuperscript𝐡𝑐{}^{c}\mathbf{h}start_FLOATSUPERSCRIPT italic_c end_FLOATSUPERSCRIPT bold_h. In each autoregression time-step of the LM, the acoustic predictor transforms the hidden output 𝐡csuperscript𝐡𝑐{}^{c}\mathbf{h}start_FLOATSUPERSCRIPT italic_c end_FLOATSUPERSCRIPT bold_h to the codec feature 𝐚^^𝐚\mathbf{\hat{a}}over^ start_ARG bold_a end_ARG of the converted speech. Finally, the decoder of the codec model reconstructs the waveform from the predicted codec feature. In the following sections, we will introduce how to build a streamable LM for VC and how to ensure the high-quality conversation of this streaming VC.

Refer to caption
Figure 3: The architecture for context-aware language model.

3.2 Streamable Architecture

To perform streaming voice conversion, a streamable architecture is necessary. In StreamVoice, the LM is carefully designed to perform a fully causal processing in the VC task, and the acoustic predictor is designed to achieve frame-wise prediction without dependency on temporal information.

3.2.1 Fully Casual Language Model

As shown in Fig. 3, inspired by the success of the LM-based VC model, we intend to achieve streaming zero-shot VC by language models. In previous LM-based VC models Wang et al. (2023c), the demand of the complete semantic feature 𝐬𝐬\mathbf{s}bold_s from source speech to achieve conversion hinders the deployment for real-time application, which can be formulated as p(at|𝐬1:Ts,𝐚1:t1)𝑝conditionalsubscript𝑎𝑡subscript𝐬:1subscript𝑇𝑠subscript𝐚:1𝑡1p(a_{t}|\mathbf{s}_{1:T_{s}},\mathbf{a}_{1:t-1})italic_p ( italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT | bold_s start_POSTSUBSCRIPT 1 : italic_T start_POSTSUBSCRIPT italic_s end_POSTSUBSCRIPT end_POSTSUBSCRIPT , bold_a start_POSTSUBSCRIPT 1 : italic_t - 1 end_POSTSUBSCRIPT ) for each time step. To achieve streaming, any components of the LM cannot rely on future information. As shown in Fig. 3, decoder-only LM with unidirectional attention can easily fit the requirement of casual generation. To eliminate the dependency of the complete semantic input, semantic and acoustic features {𝐬,𝐚}𝐬𝐚\{\mathbf{s},\mathbf{a}\}{ bold_s , bold_a } are first aligned with each other to the same sequence length T𝑇Titalic_T and then they are alternatively inputted to the LM, forming a cross-embedding like {s1,a1,s2,a2,,sT,aT}subscript𝑠1subscript𝑎1subscript𝑠2subscript𝑎2subscript𝑠𝑇subscript𝑎𝑇\{s_{1},a_{1},s_{2},a_{2},...,s_{T},a_{T}\}{ italic_s start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_s start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , … , italic_s start_POSTSUBSCRIPT italic_T end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT italic_T end_POSTSUBSCRIPT }. With these modifications, the LM can achieve streaming processing, modeling p(at|𝐬1:t,𝐚1:t1)𝑝conditionalsubscript𝑎𝑡subscript𝐬:1𝑡subscript𝐚:1𝑡1p(a_{t}|\mathbf{s}_{1:t},\mathbf{a}_{1:t-1})italic_p ( italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT | bold_s start_POSTSUBSCRIPT 1 : italic_t end_POSTSUBSCRIPT , bold_a start_POSTSUBSCRIPT 1 : italic_t - 1 end_POSTSUBSCRIPT ).

Refer to caption
Figure 4: The architecture for acoustic predictor. Our system can support continuous or discrete codec projection.

3.2.2 Acoustic Predictor

As the preceding LM has essentially encoded content and speaker into its output 𝐡csuperscript𝐡𝑐{}^{c}\mathbf{h}start_FLOATSUPERSCRIPT italic_c end_FLOATSUPERSCRIPT bold_h, the acoustic predictor can be designed in temporal irrelevant to transform 𝐡csuperscript𝐡𝑐{}^{c}\mathbf{h}start_FLOATSUPERSCRIPT italic_c end_FLOATSUPERSCRIPT bold_h into acoustic codec space, which means the predictor can be easily applied in the streaming scenario. Given that the speech can be represented in acoustic features by neural codec in either continuous or discrete forms, we investigate the incorporation of both features in StreamVoice, which are performed by continuous projection and discrete projection respectively.

Continuous Projection: Following Shen et al. Shen et al. (2023), the D𝐷Ditalic_D-dimensional quantized latent vector 𝐚T×D𝐚superscript𝑇𝐷\mathbf{a}\in\mathcal{R}^{T\times D}bold_a ∈ caligraphic_R start_POSTSUPERSCRIPT italic_T × italic_D end_POSTSUPERSCRIPT encoded by the codec model is used as the continuous acoustic representation. The prediction of the continuous representation involves employing a stack of linear layers, as shown in Fig. 4. The continuous projection loss is calculated as the L2 distance between the predicted acoustic feature 𝐚^^𝐚\hat{\mathbf{a}}over^ start_ARG bold_a end_ARG and the ground-truth acoustic feature 𝐚𝐚\mathbf{a}bold_a, which is defined as:

Cont=𝐚𝐚^22.subscript𝐶𝑜𝑛𝑡subscriptsuperscriptnorm𝐚^𝐚22\mathcal{L}_{Cont}=||\mathbf{a}-\hat{\mathbf{a}}||^{2}_{2}.caligraphic_L start_POSTSUBSCRIPT italic_C italic_o italic_n italic_t end_POSTSUBSCRIPT = | | bold_a - over^ start_ARG bold_a end_ARG | | start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT . (1)

Discrete Projection: In general, the codec is designed with multi-layer quantizers to compress original speech into L𝐿Litalic_L-layer discrete indices 𝐚T×L𝐚superscript𝑇𝐿\mathbf{a}\in\mathcal{R}^{T\times L}bold_a ∈ caligraphic_R start_POSTSUPERSCRIPT italic_T × italic_L end_POSTSUPERSCRIPT at a low bitrate. Most LM-based work Wang et al. (2023b, c) stacks multiple LMs to predict discrete features, making the pipeline complicated and unsuitable for the streaming scenario. In contrast, StreamVoice adopts a streamlined multi-layer codec prediction method inspired by MQTTS Chen et al. (2023a). This method, free from temporal dependencies, can seamlessly integrate into the streaming process of the language model. To be specific, a single-layer transformer is used to model the heretical conditional distribution of codecs. As depicted in the right of Fig. 4, at time t𝑡titalic_t, the transformer employs the 𝐡csuperscript𝐡𝑐{}^{c}\mathbf{h}start_FLOATSUPERSCRIPT italic_c end_FLOATSUPERSCRIPT bold_h as the starting condition and sequentially generates atlsubscriptsuperscript𝑎𝑙𝑡a^{l}_{t}italic_a start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT from layer 1 to L. Remarkably, this generation process is independent of the preceding or the future 𝐡𝐜superscript𝐡𝐜\mathbf{{}^{c}\mathbf{h}}start_FLOATSUPERSCRIPT bold_c end_FLOATSUPERSCRIPT bold_h, rendering it well-suited for the demands of a streaming scenario. Notably, in the proposed StreamVoice, we mainly incorporate the discrete projection to achieve acoustic prediction. The discrete projection loss can be described as:

Disc=logt=0T1p(at|𝐚1:t1,m𝐬1:t,t).\mathcal{L}_{Disc}=-\log{\prod^{T-1}_{t=0}p(a_{t}|\mathbf{a}_{1:t-1},^{m}% \mathbf{s}_{1:t},t)}.caligraphic_L start_POSTSUBSCRIPT italic_D italic_i italic_s italic_c end_POSTSUBSCRIPT = - roman_log ∏ start_POSTSUPERSCRIPT italic_T - 1 end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_t = 0 end_POSTSUBSCRIPT italic_p ( italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT | bold_a start_POSTSUBSCRIPT 1 : italic_t - 1 end_POSTSUBSCRIPT , start_POSTSUPERSCRIPT italic_m end_POSTSUPERSCRIPT bold_s start_POSTSUBSCRIPT 1 : italic_t end_POSTSUBSCRIPT , italic_t ) . (2)

3.3 Context-aware Enhancement

Due to the disadvantage of the causality in the streaming framework, streaming models face missing future reception and potential performance degradation compared to the non-streaming model, while the low-quality semantic input from the streaming ASR, as we mentioned in Section 1, makes the streaming VC for high-quality conversion more challenging. To make the proposed streamable LM-based VC model perform conversion with high quality, a context-aware enhancement method is proposed, which can alleviate incomplete contextual information arising from the semantic input and the absence of future information. Specifically, we introduce context-masked autoregressive prediction in the LM to enhance the capture of historical context from the given semantic input, and meanwhile, a teacher-guided context foresight is proposed to ensure the model can imagine the future context based on that of its historical context.

3.3.1 Context-masked Autoregressive Prediction

As shown in the left of Fig. 3, the LM is achieved by the multi-layer LLaMA with unidirectional attention. To enhance contextual awareness from the given semantic input, semantic masking is introduced in the LM to encourage acoustic prediction from the corrupted semantic. Specifically, within a sequence of semantic tokens 𝐬={s1,s2,sT}𝐬subscript𝑠1subscript𝑠2subscript𝑠𝑇\mathbf{s}=\{s_{1},s_{2},...s_{T}\}bold_s = { italic_s start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_s start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , … italic_s start_POSTSUBSCRIPT italic_T end_POSTSUBSCRIPT }, we randomly select several indices as start indices at a ratio r𝑟ritalic_r, and spans of l𝑙litalic_l steps are masked by [M]. After masking, LM takes the corrupted semantic feature 𝐬msuperscript𝐬𝑚{}^{m}\mathbf{s}start_FLOATSUPERSCRIPT italic_m end_FLOATSUPERSCRIPT bold_s as input and performs autoregression. With this method, an information bottleneck is also implicitly created in the semantic feature to reduce speaker information. Moreover, during training, we do not explicitly use a speech clip as the speaker prompt. Instead, LM leverages the previous sequence {𝐬1:t1,𝐚1:t1,st}subscript𝐬:1𝑡1subscript𝐚:1𝑡1subscript𝑠𝑡\{\mathbf{s}_{1:t-1},\mathbf{a}_{1:t-1},s_{t}\}{ bold_s start_POSTSUBSCRIPT 1 : italic_t - 1 end_POSTSUBSCRIPT , bold_a start_POSTSUBSCRIPT 1 : italic_t - 1 end_POSTSUBSCRIPT , italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT } as prompts to autoregressively generate hidden representation htsubscript𝑡h_{t}italic_h start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT for further acoustic prediction. Notably, during training, when the input of the current step is atsubscript𝑎𝑡a_{t}italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT, the corresponding output is skipped and does not involve further steps.

3.3.2 Teacher-guided Context Foresight

As previously discussed, the absence of future information resulting in the loss of contextual information leads to a decline in the conversion performance. Inspired by the effective representation learning exhibited by autoregressive predictive coding Chung et al. (2019) (APC), we introduce teacher-guided context foresight guided by a non-streaming ASR to enhance the autoregression output, as presented in the right of Fig. 3. This allows the model to learn a context vector containing envisioned future information. Specifically, the context representation 𝐜𝐜\mathbf{c}bold_c is first derived by linear prediction from the hidden features 𝐡𝐡\mathbf{h}bold_h, which is generated by the LM through historical context. Subsequently, this ctsubscript𝑐𝑡c_{t}italic_c start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT is encouraged to discover more general future context information by minimizing the L2 distance not only with k𝑘kitalic_k semantic features from future time steps s¯t+1,,s¯t+ksubscript¯𝑠𝑡1subscript¯𝑠𝑡𝑘{\overline{s}_{t+1},...,\overline{s}_{t+k}}over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t + 1 end_POSTSUBSCRIPT , … , over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t + italic_k end_POSTSUBSCRIPT but also with the current semantic feature st¯¯subscript𝑠𝑡\overline{s_{t}}over¯ start_ARG italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT end_ARG. This dual minimization approach contributes to precise content delivery and enhances the model’s ability to forecast future context. The loss can be summarized as

TF=1Tk1TkctConcat(s¯t,s¯t+1,,s¯t+k)22subscript𝑇𝐹1𝑇𝑘superscriptsubscript1𝑇𝑘subscriptsuperscriptnormsubscript𝑐𝑡𝐶𝑜𝑛𝑐𝑎𝑡subscript¯𝑠𝑡subscript¯𝑠𝑡1subscript¯𝑠𝑡𝑘22\mathcal{L}_{TF}=\frac{1}{T-k}\sum_{1}^{T-k}\left\|c_{t}-Concat(\overline{s}_{% t},\overline{s}_{t+1},...,\overline{s}_{t+k})\right\|^{2}_{2}caligraphic_L start_POSTSUBSCRIPT italic_T italic_F end_POSTSUBSCRIPT = divide start_ARG 1 end_ARG start_ARG italic_T - italic_k end_ARG ∑ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_T - italic_k end_POSTSUPERSCRIPT ∥ italic_c start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT - italic_C italic_o italic_n italic_c italic_a italic_t ( over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t + 1 end_POSTSUBSCRIPT , … , over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t + italic_k end_POSTSUBSCRIPT ) ∥ start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT (3)

where Concat()𝐶𝑜𝑛𝑐𝑎𝑡Concat(\cdot)italic_C italic_o italic_n italic_c italic_a italic_t ( ⋅ ) denotes the concatenation of features along the dimensional axe. Unlike the original APC, which operates between the input and output of an autoregressive model, our approach employs a non-streaming ASR model as a teacher to provide semantic information 𝐬¯¯𝐬\mathbf{\overline{s}}over¯ start_ARG bold_s end_ARG for guiding this foresight process. This is done to tackle the inherent challenge of obtaining high-quality semantic features from the streaming ASR. After dimensional transformations, the context representation 𝐜𝐜\mathbf{c}bold_c is then combined with 𝐡𝐡\mathbf{h}bold_h to form the context-enhanced 𝐡csuperscript𝐡𝑐{}^{c}\mathbf{h}start_FLOATSUPERSCRIPT italic_c end_FLOATSUPERSCRIPT bold_h, which is then fed into the acoustic predictor.

Furthermore, since the semantic feature {𝐬,𝐬¯}𝐬¯𝐬\{\mathbf{s},\mathbf{\overline{s}}\}{ bold_s , over¯ start_ARG bold_s end_ARG } from ASR still may contain speaker-related information. To further ensure the speech decoupling, the bottleneck regulator Qian et al. (2019), which squeezes out speaker information by reducing dimension size with a linear layer, is applied in 𝐬𝐬\mathbf{s}bold_s and 𝐜𝐜\mathbf{c}bold_c.

3.4 Training & Inference Procedure

3.4.1 Training

During the training of StreamVoice, the context-enhanced language model and acoustic predictor are trained together. Both the autoregressive processes in these two parts are performed with a teacher-forcing strategy. The total loss can be described as total=TF+Contsubscript𝑡𝑜𝑡𝑎𝑙subscript𝑇𝐹subscript𝐶𝑜𝑛𝑡\mathcal{L}_{total}=\mathcal{L}_{TF}+\mathcal{L}_{Cont}caligraphic_L start_POSTSUBSCRIPT italic_t italic_o italic_t italic_a italic_l end_POSTSUBSCRIPT = caligraphic_L start_POSTSUBSCRIPT italic_T italic_F end_POSTSUBSCRIPT + caligraphic_L start_POSTSUBSCRIPT italic_C italic_o italic_n italic_t end_POSTSUBSCRIPT for using continuous codec feature or total=TF+Discsubscript𝑡𝑜𝑡𝑎𝑙subscript𝑇𝐹subscript𝐷𝑖𝑠𝑐\mathcal{L}_{total}=\mathcal{L}_{TF}+\mathcal{L}_{Disc}caligraphic_L start_POSTSUBSCRIPT italic_t italic_o italic_t italic_a italic_l end_POSTSUBSCRIPT = caligraphic_L start_POSTSUBSCRIPT italic_T italic_F end_POSTSUBSCRIPT + caligraphic_L start_POSTSUBSCRIPT italic_D italic_i italic_s italic_c end_POSTSUBSCRIPT for discrete version.

3.4.2 Streaming Inference

In inference, we use the semantic and acoustic features from a short speech clip of the target speaker as the speaker prompt. Since this clip is randomly selected, which may contain unfinished pronunciation at the end of the clip, we pad a silence clip after the speaker recording before the conversion process to prevent the unexpected continuation. With this prompt, StreamVoice can stream convert the source speech. In discrete projection, we use greedy decoding to choose the codec token with the highest probability. Besides, to ensure the real-time streaming inference of StreamVoice, we employ a commonly used key-value cache in LM to reduce redundant calculations. In our pipeline, since the beginning and end of the source speech can be determined by ASR or voice activity detection (VAD), we don’t employ techniques, such as window attention or slide attention, to handle the input. It is noteworthy that these techniques can be easily integrated into our framework, providing flexibility for future extensions.

Table 1: Zero-shot performance (unseen speakers)
Table 2: In-datatset performance (seen speakers)
Method Quality Similarity
NMOS \uparrow WVMOS \uparrow CER \downarrow SMOS \uparrow SSIM \uparrow
GT - 3.61 6.29 - 0.803
Non-streaming Topline
LM-VC 3.80±plus-or-minus\pm±0.09 3.74 8.93 3.78±plus-or-minus\pm±0.08 0.742
NS-StreamVoice 3.87±plus-or-minus\pm±0.07 3.68 8.51 3.73±plus-or-minus\pm±0.11 0.755
Streaming Model
C-StreamVoice 3.72±plus-or-minus\pm±0.10 3.49 10.2 3.67±plus-or-minus\pm±0.09 0.729
StreamVoice 3.83±plus-or-minus\pm±0.09 3.63 9.43 3.74±plus-or-minus\pm±0.08 0.740
Method Chunk (ms) Quality Similarity
NMOS \uparrow WVMOS \uparrow CER \downarrow SMOS \uparrow SSIM \uparrow
GT - - 3.65 6.29 - 0.729
Non-streaming Topline
NS-VC - 3.85±plus-or-minus\pm±0.09 3.71 8.39 3.92±plus-or-minus\pm±0.08 0.744
Streaming Model
IBF-VC 160 3.71±plus-or-minus\pm±0.09 3.48 9.52 3.67±plus-or-minus\pm±0.10 0.687
DualVC2 160 3.80±plus-or-minus\pm±0.10 3.57 10.2 3.85±plus-or-minus\pm±0.09 0.703
StreamVoice 80 3.82±plus-or-minus\pm±0.09 3.50 10.0 3.82±plus-or-minus\pm±0.10 0.694
+\quad++ Tuning 80 3.78±plus-or-minus\pm±0.08 3.52 10.4 3.87±plus-or-minus\pm±0.10 0.714
Table 2: In-datatset performance (seen speakers)

4 Experiments

We first present the experimental setup in our work. Next, we provide subjective and objective evaluations and the ablation study conducted on StreamVoice. A detailed analysis of dependency in the streaming pipeline is also provided.

4.1 Experimental Setup

4.1.1 Corpus

A mixed dataset comprising 1,500 hours of Aishell3 Shi et al. (2020) and an internal Chinese dataset are used to train StreamVoice and Audiodec Wu et al. (2023). To extract semantic features, we incorporate a streaming ASR Fast-U2++ Liang et al. (2023), which is implemented by WeNet and trained on WenetSpeech Zhang et al. (2022). For zero-shot testing, a set of 400 testing pairs is selected from DIDISpeech Guo et al. (2021) and EMIME Wester (2010), each with a source and target speaker utterance. For evaluation on seen speakers, eight speakers from Aishell3 are selected to form 160 conversion pairs.

4.1.2 Implement Details

We use open-sourced code222https://github.com/facebookresearch/AudioDec of Audiodec, which has 4 quantizer layers with a 1024 codebook size and 64 codebook dimension, representing a 24kHz waveform in 20ms frame length. The Fast-U2++ uses an 80ms chunk size to perform streaming inference and compresses a 16kHz waveform into a semantic feature with a 40ms frame length. For StreamVoice, we employ the LLaMA Touvron et al. (2023) architecture for context-enhanced LM, with 6 layers and 8 heads. The hidden and intermediate sizes are 1024 and 4096. We use the officially released code333https://github.com/b04901014/MQTTS to implement the acoustic predictor, which uses a layer Transformer decoder with a hidden size 256, feed-forward hidden size 1024, and 4 heads. In semantic masking, mask ratio r𝑟ritalic_r ranges from 0.010.010.010.01 to 0.020.020.020.02, and span l𝑙litalic_l is set to 10. The foresight step k𝑘kitalic_k is set to 4. The bottleneck regulator compresses feature dimensions by 6 times. During training, the max training length is set to 12s. StreamVoice is trained using 8 V100 GPUs with a batch size of 7 per GPU for 700k steps. We use the AdamW optimizer with a learning rate of 5×1045superscript1045\times 10^{-4}5 × 10 start_POSTSUPERSCRIPT - 4 end_POSTSUPERSCRIPT. Exponential decay updates the learning rate after each epoch, using a decay ratio of 0.986.

4.1.3 Evaluation metrics

The mean opinion score (MOS) subjectively measures speech naturalness (NMOS) and speaker similarity (SMOS), which are calculated with 95%percent\%% confidence intervals. We randomly select 120 testing pairs for subjective evaluations, involving a group of 15 listeners. For objective evaluations, a neural network-based system with open-source implementation444https://github.com/AndreevP/wvmos is used to measure speech quality (WV-MOS). Character error rate (CER) measured by an ASR model555https://github.com/wenet-e2e/wenet/tree/main/examples/wenetspeech indicates the speech intelligibility. Speaker similarity (SSIM) is calculated by an SV model Desplanques et al. (2020) to determine if the converted speech matches the target speaker. Real-time factor (RTF) and latency are used to evaluate the streaming performance. Converted samples can be found in https://kerwinchao.github.io/streamvoice/.

4.2 Experiments Results

4.2.1 Zero-shot evaluation

To evaluate the zero-shot VC performance, one recent LM-based zero-shot VC system, LM-VC Wang et al. (2023c), is selected as the topline system. Besides, a variant of StreamVoice, referred to as NS-StreamVoice, using a non-streaming ASR for semantic extraction, is also compared. We implement the proposed system StreamVoice integration discrete projection, while C-StreamVoice also involves the evaluation since speech can represented in continuous form by codec model. Table. 3.4.2 presents both subjective and objective results. Compared with the non-streaming topline LM-VC, our proposed StreamVoice can achieve close results regarding subjective NMOS and SMOS, while a performance gap still exists. Similar results are also observed in objective results. The non-streaming StreamVoice even surpasses the topline model in certain aspects, indicating the effectiveness of our well-designed streamable architecture for zero-shot VC. As can be observed, C-StreamVoice exhibits inferior performance compared to the discrete version, which can contribute to the reported over-smoothness in speech generation Ren et al. (2022) and the mismatch between the ground-truth and predicted features.

Furthermore, as illustrated in Table. 3, the RTF of the entire pipeline is below 1, which meets the real-time requirement. Consisting of chunk-waiting latency (80ms) and model inference latency, the overall pipeline latency is 124.3 ms. If using a V100 GPU, StreamVoice can obtain an RTF of 0.56 and the overall latency reaches 137.2ms. Importantly, unlike previous streaming VC, our VC model is entirely causal without any future look-ahead, highlighting its powerful modeling capability. These results demonstrate that StreamVoice can achieve high-quality zero-shot VC in streaming scenarios.

Table 3: Speed of StreamVoice tested on a single A100 80G GPU. Latency is obtained by multiplying the RTF by the 80ms chunk size.
RTF Latency (ms)
ASR Encoder 0.13  10.4
Codec Decoder 0.004  0.3
StreamVoice 0.42  33.6
Overall 0.554  44.3+80=124.3

4.2.2 In-dataset Evaluation

To get further insight into StreamVoice, we conducted an in-dataset evaluation on eight seen speakers, as shown in Table. 3.4.2. A non-streaming VC system Tian et al. (2020) achieving any2many VC, is selected, referred to as NS-VC. Also, IBF-VC Chen et al. (2022) and DualVC2 Ning et al. (2023) are recently proposed streaming models for any2many VC. As observed, a performance gap exists between the strong non-streaming topline and streaming models. Among the streaming models, StreamVoice, designed for the zero-shot scenario, delivers similar results to systems designed for in-dataset speakers. Even though StreamVoice uses a smaller chunk size of 80ms in streaming ASR which achieves lower ASR performance, indicating the good conversion ability of StreamVoice. With available utterances of target speakers, fine-tuning the StreamVoice yields superior performance. This indicates our system can be easily applied to various scenarios with or without the utterances of target speakers.

Table 4: Results of ablation studies.
Method WVMOS \uparrow CER \downarrow SSIM \uparrow
StreamVoice 3.63 9.43 0.740
    w/o Teacher-guided Context Foresight
        w/o TF(s¯t)subscript𝑇𝐹subscript¯𝑠𝑡\mathcal{L}_{TF}(\overline{s}_{t})caligraphic_L start_POSTSUBSCRIPT italic_T italic_F end_POSTSUBSCRIPT ( over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) 2.56 76.8 0.59
        w/o TF(s¯t+1:t+k)subscript𝑇𝐹subscript¯𝑠:𝑡1𝑡𝑘\mathcal{L}_{TF}(\overline{s}_{t+1:t+k})caligraphic_L start_POSTSUBSCRIPT italic_T italic_F end_POSTSUBSCRIPT ( over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t + 1 : italic_t + italic_k end_POSTSUBSCRIPT ) 3.39 13.7 0.728
   w/o Semantic Masking 3.47 13.0 0.715
   w/o Bottleneck Regulator 3.59 9.21 0.718

4.3 Ablation study

To further validate the effectiveness of StreamVoice, we conducted several ablations studies on teacher-guided context foresight, semantic masking, and the bottleneck regulator. Table 4 presents the results.

In w/o teacher-guided context foresight, we discard the prediction of current and future semantic information, forming two ablations w/o TF(s¯t)subscript𝑇𝐹subscript¯𝑠𝑡\mathcal{L}_{TF}(\overline{s}_{t})caligraphic_L start_POSTSUBSCRIPT italic_T italic_F end_POSTSUBSCRIPT ( over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) and w/o TF(s¯t+1:t+k)subscript𝑇𝐹subscript¯𝑠:𝑡1𝑡𝑘\mathcal{L}_{TF}(\overline{s}_{t+1:t+k})caligraphic_L start_POSTSUBSCRIPT italic_T italic_F end_POSTSUBSCRIPT ( over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t + 1 : italic_t + italic_k end_POSTSUBSCRIPT ), respectively. As can be seen, a noticeable decrease occurs in all evaluation metrics when the TF(s¯t+1:t+k)subscript𝑇𝐹subscript¯𝑠:𝑡1𝑡𝑘\mathcal{L}_{TF}(\overline{s}_{t+1:t+k})caligraphic_L start_POSTSUBSCRIPT italic_T italic_F end_POSTSUBSCRIPT ( over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t + 1 : italic_t + italic_k end_POSTSUBSCRIPT ) is discarded, especially in WVMOS and CER. This indicates that this foresight improves performance in capturing the linguistic content. But when only integrating context from future semantics, the model w/o TF(s¯t)subscript𝑇𝐹subscript¯𝑠𝑡\mathcal{L}_{TF}(\overline{s}_{t})caligraphic_L start_POSTSUBSCRIPT italic_T italic_F end_POSTSUBSCRIPT ( over¯ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) faces severe performance loss. It shows that only using future information interferes with delivering current linguistic content.

In w/o semantic masking, the semantic masking does not apply to the semantic feature during the training of StreamVoice. We observe a performance decrease in all evaluation metrics when the masking is discarded. This indicates that StreamVoice, trained with semantic masking, effectively enhances contextual learning from the preceding input while improving speaker timbre capturing.

We also evaluated the effect of the bottleneck regulator by dropping this usage, referred to as w/o bottleneck regulator. The results show that the integration of the bottleneck regulator is effective in preventing the source speaker information contained in the semantic feature from leaking into the converted speech, with little effect on the speech quality.

Table 5: Analysis of dependency on ASR. The results represent the performance of StreamVoice integrating different ASR.
Type of ASR WVMOS \uparrow CER \downarrow SSIM \uparrow
Non-streaming ASR 3.68 8.51 0.755
Streaming ASR ( Moritz et al. (2019))
+\quad++ 0ms Future Look-ahead 3.19 91.7 0.674
+\quad++ 160ms Future Look-ahead 3.48 10.6 0.727
Streaming ASR (Fast-U2++ Liang et al. (2023))
    Chunk (80ms) 3.63 9.43 0.740
    Chunk (160ms) 3.69 9.16 0.744

4.4 Discussion: Dependency Analysis

As shown in Fig. 2, in the streaming pipeline, additional ASR and codec are needed to extract semantic and acoustic information. In this section, we will explore the dependency relations between the selection of ASR and codec and the performance of StreamVoice.

4.4.1 ASR

To investigate the impact of ASR on StreamVoice, three representative ASR systems, including non-streaming ASR 5, widely used CTC-based streaming ASR Moritz et al. (2019), and the recently proposed streaming Fast-U2++ Liang et al. (2023), are selected to perform semantic extraction. As can be seen in Table 5, StreamVoice using semantic features of non-streaming ASR outperforms those using streaming ASR. This discrepancy may be attributed to the inherent performance gap between non-streaming and streaming ASR models, resulting in different semantic extraction abilities. Besides, without future look-ahead in StreamVoice, using semantic features from Moritz et al. (2019) cannot achieve reasonable conversion, while we introduce a 160ms future look-ahead in StreamVoice, i.e., modeling p(at|𝐚1:t1,𝐬1:t+m,t)𝑝conditionalsubscript𝑎𝑡subscript𝐚:1𝑡1subscript𝐬:1𝑡𝑚𝑡p(a_{t}|\mathbf{a}_{1:t-1},\mathbf{s}_{1:t+m},t)italic_p ( italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT | bold_a start_POSTSUBSCRIPT 1 : italic_t - 1 end_POSTSUBSCRIPT , bold_s start_POSTSUBSCRIPT 1 : italic_t + italic_m end_POSTSUBSCRIPT , italic_t ) with m𝑚mitalic_m future look-ahead, yield good conversion results. This issue may arise from delayed CTC spike distributions and token emission latency existing in streaming ASR Liang et al. (2023), leading to semantic information shifting. Benefiting from the low emission latency of Fast-U2++, StreamVoice can perform conversion without future look-ahead. With a longer chunk size employed in Fast-U2++, StreamVoice can obtain better results, while reaching a larger latency of 270ms. A trade-off still exists between performance and speed within the streaming pipeline.

Table 6: Analysis of dependency on Audiodec with various bitrate.
Type of Audiodec WVMOS \uparrow CER \downarrow SSIM \uparrow RTF
w/ 2kbps2𝑘𝑏𝑝𝑠2kbps2 italic_k italic_b italic_p italic_s Audiodec 3.63 9.43 0.740 0.42
w/ 8kbps8𝑘𝑏𝑝𝑠8kbps8 italic_k italic_b italic_p italic_s Audiodec 3.61 9.38 0.738 0.61
Large w/ 8kbps8𝑘𝑏𝑝𝑠8kbps8 italic_k italic_b italic_p italic_s Audiodec 3.68 9.12 0.751 0.90

4.4.2 Codec

In StreamVoice, we employ a low-latency streaming codec Audiodec Wu et al. (2023). As presented in Table. 6, we validate the performance of StreamVoice using codecs with different bitrates, including 2kbps and 8kbps, where higher bitrate codecs achieve superior reconstruction quality to lower bitrate ones. The 2kbps Audiodec utilizes 4 layers of quantization and represents audio with a frame length of 20ms, while the 8kbps Audiodec employs 8 layers with a frame length of 10ms. Using the configuration of StreamVoice mentioned in Section 4.1.2, the results in different bitrates of codec models show no obvious differences. When increasing the number of transformer layers in the codec predictor, forming Large w/ 8kbps Audiodec, conversion performance using 8kbps codec improves noticeably, but resulting in slower inference. This result shows that the design of StreamVoice depends on the codec configuration, affecting both conversion quality and inference speed.

4.5 Conclusions

In this paper, we introduce StreamVoice, a novel LM-based zero-shot VC system designed for streaming scenarios. Specifically, StreamVoice employs a streamlined, single-stage framework that encompasses a context-aware LM and an acoustic predictor. The casual design of the model’s input and structure ensures compliance with streaming behavior. To address performance degradation caused by missing complete contextual information in streaming scenarios, context-aware LM adopts teacher-guided context foresight to make the model have the ability to forecast the current and future information given by a teacher. Besides, semantic masking is introduced in LM to enhance context learning from historical input and facilitate better disentanglement. Finally, an acoustic predictor collaborates with the LM to generate the target speech. Experiments demonstrate that StreamVoice achieves streaming zero-shot VC while maintaining performance comparable to non-streaming VC systems.

Limitations and future work. We have to point out that StreamVoice still has limitations. In our configuration, StreamVoice needs a GPU, such as V100 and A100, to achieve real-time streaming inference. The design of streaming VC heavily relies on the ASR and codec as mentioned in Section 4.4. Besides, StreamVoice also faces the out-of-domain problem, which causes performance degradation for utterances with accents, strong emotions, or unseen recording environments. Our future work will first use more training data to explore the modeling ability of StreamVoice. Also, we will focus on optimizing our streaming pipeline, such as high fidelity codec with low bitrate and unified streaming model.

Ethics Statement

Since StreamVoice can convert source speech to desired speakers, it may carry potential risks of misuse for various purposes, such as spreading fake information or phone fraud. To prevent the abuse of the VC technology, many studies have focused on synthetic speech detection Yi et al. (2022). Meanwhile, we also encourage the public to report the illegal usage of VC to the appropriate authorities.

Acknowledgments

We would like to thank the Speech Group from ByteDance for their invaluable discussions and providing the essential computing resources. We express our special gratitude to Ao Zhang and Ziqiang Ning for their various insights on streaming ASR and VC.

References

  • Agostinelli et al. [2023] Andrea Agostinelli, Timo I Denk, Zalán Borsos, Jesse Engel, Mauro Verzetti, Antoine Caillon, Qingqing Huang, Aren Jansen, Adam Roberts, Marco Tagliasacchi, et al. Musiclm: Generating music from text. Arxiv, 2023.
  • Borsos et al. [2023] Zalán Borsos, Raphaël Marinier, Damien Vincent, Eugene Kharitonov, Olivier Pietquin, Matt Sharifi, Dominik Roblek, Olivier Teboul, David Grangier, Marco Tagliasacchi, and Neil Zeghidour. Audiolm: A language modeling approach to audio generation. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 31:2523–2533, 2023.
  • Chen et al. [2022] Ziyi Chen, Haoran Miao, and Pengyuan Zhang. Streaming non-autoregressive model for any-to-many voice conversion. Arxiv, 2022.
  • Chen et al. [2023a] Li-Wei Chen, Shinji Watanabe, and Alexander Rudnicky. A vector quantized approach for text to speech synthesis on real-world spontaneous speech. In Association for the Advancement of Artificial Intelligence, 2023.
  • Chen et al. [2023b] Yuanzhe Chen, Ming Tu, Tang Li, Xin Li, Qiuqiang Kong, Jiaxin Li, Zhichao Wang, Qiao Tian, Yuping Wang, and Yuxuan Wang. Streaming voice conversion via intermediate bottleneck features and non-streaming teacher guidance. In International Conference on Acoustics, Speech and Signal Processing, pages 1–5, 2023.
  • Choi et al. [2021] Hyeong-Seok Choi, Juheon Lee, Wansoo Kim, Jie Lee, Hoon Heo, and Kyogu Lee. Neural analysis and synthesis: Reconstructing speech from self-supervised representations. In Neural Information Processing Systems, pages 16251–16265, 2021.
  • Chou and Lee [2019] Juchieh Chou and Hungyi Lee. One-shot voice conversion by separating speaker and content representations with instance normalization. In International Speech Communication Association, pages 664–668, 2019.
  • Chung et al. [2019] Yu-An Chung, Wei-Ning Hsu, Hao Tang, and James Glass. An Unsupervised Autoregressive Model for Speech Representation Learning. In International Speech Communication Association, pages 146–150, 2019.
  • Chung et al. [2021] Yu-An Chung, Yu Zhang, Wei Han, Chung-Cheng Chiu, James Qin, Ruoming Pang, and Yonghui Wu. w2v-bert: Combining contrastive learning and masked language modeling for self-supervised speech pre-training. In Automatic Speech Recognition and Understanding Workshop (ASRU), pages 244–250, 2021.
  • Desplanques et al. [2020] Brecht Desplanques, Jenthe Thienpondt, and Kris Demuynck. ECAPA-TDNN: Emphasized channel attention, propagation and aggregation in TDNN based speaker verification. In International Speech Communication Association, pages 3830–3834, 2020.
  • Ebbers et al. [2021] Janek Ebbers, Michael Kuhlmann, Tobias Cord-Landwehr, and Reinhold Haeb-Umbach. Contrastive predictive coding supported factorized variational autoencoder for unsupervised learning of disentangled speech representations. In International Conference on Acoustics, Speech and Signal Processing, pages 3860–3864, 2021.
  • Gu et al. [2021] Yewei Gu, Zhenyu Zhang, Xiaowei Yi, and Xianfeng Zhao. MediumVC: Any-to-any voice conversion using synthetic specific-speaker speeches as intermedium features. Arxiv, 2021.
  • Guo et al. [2021] Tingwei Guo, Cheng Wen, Dongwei Jiang, Ne Luo, Ruixiong Zhang, Shuaijiang Zhao, Wubo Li, Cheng Gong, Wei Zou, Kun Han, and Xiangang Li. Didispeech: A large scale mandarin speech corpus. In International Conference on Acoustics, Speech and Signal Processing, pages 6968–6972, 2021.
  • Hayashi et al. [2022] Tomoki Hayashi, Kazuhiro Kobayashi, and Tomoki Toda. An investigation of streaming non-autoregressive sequence-to-sequence voice conversion. In International Conference on Acoustics, Speech and Signal Processing, pages 6802–6806, 2022.
  • Kameoka et al. [2021] Hirokazu Kameoka, Kou Tanaka, and Takuhiro Kaneko. Fasts2s-vc: Streaming non-autoregressive sequence-to-sequence voice conversion. Arxiv, 2021.
  • Kharitonov et al. [2023] Eugene Kharitonov, Damien Vincent, Zalán Borsos, Raphaël Marinier, Sertan Girgin, Olivier Pietquin, Matt Sharifi, Marco Tagliasacchi, and Neil Zeghidour. Speak, read and prompt: High-fidelity text-to-speech with minimal supervision. Arxiv, 2023.
  • Liang et al. [2023] Chengdong Liang, Xiao-Lei Zhang, BinBin Zhang, Di Wu, Shengqiang Li, Xingchen Song, Zhendong Peng, and Fuping Pan. Fast-u2++: Fast and accurate end-to-end speech recognition in joint ctc/attention frames. In International Conference on Acoustics, Speech and Signal Processing, pages 1–5. IEEE, 2023.
  • Moritz et al. [2019] Niko Moritz, Takaaki Hori, and Jonathan Le Roux. Streaming end-to-end speech recognition with joint ctc-attention based models. In Automatic Speech Recognition and Understanding Workshop, pages 936–943, 2019.
  • Ning et al. [2023] Ziqian Ning, Yuepeng Jiang, Pengcheng Zhu, Shuai Wang, Jixun Yao, Lei Xie, and Mengxiao Bi. Dualvc 2: Dynamic masked convolution for unified streaming and non-streaming voice conversion. Arxiv, 2023.
  • Qian et al. [2019] Kaizhi Qian, Yang Zhang, Shiyu Chang, Xuesong Yang, and Mark Hasegawa-Johnson. Autovc: Zero-shot voice style transfer with only autoencoder loss. In International Conference on Machine Learning, pages 5210–5219, 2019.
  • Ren et al. [2022] Yi Ren, Xu Tan, Tao Qin, Zhou Zhao, and Tie-Yan Liu. Revisiting over-smoothness in text to speech. In Association for Computational Linguistics, pages 8197–8213, 2022.
  • Shen et al. [2023] Kai Shen, Zeqian Ju, Xu Tan, Yanqing Liu, Yichong Leng, Lei He, Tao Qin, Sheng Zhao, and Jiang Bian. Naturalspeech 2: Latent diffusion models are natural and zero-shot speech and singing synthesizers. Arxiv, 2023.
  • Shi et al. [2020] Yao Shi, Hui Bu, Xin Xu, Shaoji Zhang, and Ming Li. Aishell-3: A multi-speaker mandarin tts corpus and the baselines. Arxiv, 2020.
  • Sun et al. [2016] Lifa Sun, K. Li, Hao Wang, Shiyin Kang, and H. Meng. Phonetic posteriorgrams for many-to-one voice conversion without parallel data training. In International Conference on Multimedia and Expo, pages 1–6, 2016.
  • Tian et al. [2020] Xiaohai Tian, Zhichao Wang, Shan Yang, Xinyong Zhou, Hongqiang Du, Yi Zhou, Mingyang Zhang, Kun Zhou, Berrak Sisman, Lei Xie, and Haizhou Li. The NUS & NWPU system for Voice Conversion Challenge 2020. In Joint Workshop for the Blizzard Challenge and Voice Conversion Challenge 2020, pages 170–174, 2020.
  • Touvron et al. [2023] Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, et al. Llama: Open and efficient foundation language models. Arxiv, 2023.
  • Wang et al. [2021] Disong Wang, Liqun Deng, Yu Ting Yeung, Xiao Chen, Xunying Liu, and Helen Meng. Vqmivc: Vector quantization and mutual information-based unsupervised speech representation disentanglement for one-shot voice conversion. In International Speech Communication Association, pages 1344–1348, 2021.
  • Wang et al. [2023a] Bohan Wang, Damien Ronssin, and Milos Cernak. Alo-vc: Any-to-any low-latency one-shot voice conversion. In International Speech Communication Association, pages 2073–2077, 2023.
  • Wang et al. [2023b] Chengyi Wang, Sanyuan Chen, Yu Wu, Ziqiang Zhang, Long Zhou, Shujie Liu, Zhuo Chen, Yanqing Liu, Huaming Wang, Jinyu Li, et al. Neural codec language models are zero-shot text to speech synthesizers. Arxiv, 2023.
  • Wang et al. [2023c] Zhichao Wang, Yuanzhe Chen, Lei Xie, Qiao Tian, and Yuping Wang. Lm-vc: Zero-shot voice conversion via speech generation based on language models. IEEE Signal Processing Letters, pages 1157–1161, 2023.
  • Wang et al. [2023d] Zhichao Wang, Liumeng Xue, Qiuqiang Kong, Lei Xie, Yuanzhe Chen, Qiao Tian, and Yuping Wang. Multi-level temporal-channel speaker retrieval for robust zero-shot voice conversion. Arxiv, 2023.
  • Wester [2010] Mirjam Wester. The EMIME bilingual database. Technical report, The University of Edinburgh, 2010.
  • Wu et al. [2023] Yi-Chiao Wu, Israel D Gebru, Dejan Marković, and Alexander Richard. Audiodec: An open-source streaming high-fidelity neural audio codec. In International Conference on Acoustics, Speech and Signal Processing, pages 1–5, 2023.
  • Yang et al. [2022] Haoquan Yang, Liqun Deng, Yu Ting Yeung, Nianzu Zheng, and Yong Xu. Streamable speech representation disentanglement and multi-level prosody modeling for live one-shot voice conversion. In International Speech Communication Association, pages 2578–2582, 2022.
  • Yang et al. [2023] Dongchao Yang, Jinchuan Tian, Xu Tan, Rongjie Huang, Songxiang Liu, Xuankai Chang, Jiatong Shi, Sheng Zhao, Jiang Bian, Xixin Wu, et al. Uniaudio: An audio foundation model toward universal audio generation. Arxiv, 2023.
  • Yi et al. [2022] Jiangyan Yi, Ruibo Fu, Jianhua Tao, Shuai Nie, Haoxin Ma, Chenglong Wang, Tao Wang, Zhengkun Tian, Ye Bai, Cunhan Fan, Shan Liang, Shiming Wang, Shuai Zhang, Xinrui Yan, Le Xu, Zhengqi Wen, and Haizhou Li. Add 2022: the first audio deep synthesis detection challenge. In International Conference on Acoustics, Speech and Signal Processing. IEEE, 2022.
  • Yin et al. [2021] Dacheng Yin, Xuanchi Ren, Chong Luo, Yuwang Wang, Zhiwei Xiong, and Wenjun Zeng. Retriever: Learning content-style representation as a token-level bipartite graph. In International Conference on Learning Representations, 2021.
  • Zeghidour et al. [2021] Neil Zeghidour, Alejandro Luebs, Ahmed Omran, Jan Skoglund, and Marco Tagliasacchi. SoundStream: An end-to-end neural audio codec. Transactions on Audio, Speech, and Language Processing, 30:495–507, 2021.
  • Zhang et al. [2022] Binbin Zhang, Hang Lv, Pengcheng Guo, Qijie Shao, Chao Yang, Lei Xie, Xin Xu, Hui Bu, Xiaoyu Chen, Chenchen Zeng, et al. Wenetspeech: A 10000+ hours multi-domain mandarin corpus for speech recognition. In International Conference on Acoustics, Speech and Signal Processing, pages 6182–6186, 2022.
  • Zhu et al. [2023] Xinfa Zhu, Yuanjun Lv, Yi Lei, Tao Li, Wendi He, Hongbin Zhou, Heng Lu, and Lei Xie. Vec-tok speech: speech vectorization and tokenization for neural speech generation, 2023.