FIX:When a prompt is present, concatenate two patches as the context for VAE decoding
This commit is contained in:
@ -452,7 +452,6 @@ class VoxCPMModel(nn.Module):
|
|||||||
patch_len = self.patch_size * self.chunk_size
|
patch_len = self.patch_size * self.chunk_size
|
||||||
for latent_pred, _ in inference_result:
|
for latent_pred, _ in inference_result:
|
||||||
decode_audio = self.audio_vae.decode(latent_pred.to(torch.float32))
|
decode_audio = self.audio_vae.decode(latent_pred.to(torch.float32))
|
||||||
print(decode_audio.shape)
|
|
||||||
decode_audio = decode_audio[..., -patch_len:].squeeze(1).cpu()
|
decode_audio = decode_audio[..., -patch_len:].squeeze(1).cpu()
|
||||||
yield decode_audio
|
yield decode_audio
|
||||||
break
|
break
|
||||||
|
|||||||
Reference in New Issue
Block a user