File size: 3,557 Bytes
17ed7d8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
from typing import Callable, Tuple

import torch
import torch.nn as nn  # pylint: disable=consider-using-from-import

from TTS.tts.layers.delightful_tts.variance_predictor import VariancePredictor
from TTS.tts.utils.helpers import average_over_durations


class PitchAdaptor(nn.Module):  # pylint: disable=abstract-method
    """Module to get pitch embeddings via pitch predictor



    Args:

        n_input (int): Number of pitch predictor input channels.

        n_hidden (int): Number of pitch predictor hidden channels.

        n_out (int): Number of pitch predictor out channels.

        kernel size (int): Size of the kernel for conv layers.

        emb_kernel_size (int): Size the kernel for the pitch embedding.

        p_dropout (float): Probability of dropout.

        lrelu_slope (float): Slope for the leaky relu.



    Inputs: inputs, mask

        - **inputs** (batch, time1, dim): Tensor containing input vector

        - **target** (batch, 1, time2): Tensor containing the pitch target

        - **dr** (batch, time1): Tensor containing aligner durations vector

        - **mask** (batch, time1): Tensor containing indices to be masked

    Returns:

        - **pitch prediction** (batch, 1, time1): Tensor produced by pitch predictor

        - **pitch embedding** (batch, channels, time1): Tensor produced pitch pitch adaptor

        - **average pitch target(train only)** (batch, 1, time1): Tensor produced after averaging over durations

    """

    def __init__(

        self,

        n_input: int,

        n_hidden: int,

        n_out: int,

        kernel_size: int,

        emb_kernel_size: int,

        p_dropout: float,

        lrelu_slope: float,

    ):
        super().__init__()
        self.pitch_predictor = VariancePredictor(
            channels_in=n_input,
            channels=n_hidden,
            channels_out=n_out,
            kernel_size=kernel_size,
            p_dropout=p_dropout,
            lrelu_slope=lrelu_slope,
        )
        self.pitch_emb = nn.Conv1d(
            1,
            n_input,
            kernel_size=emb_kernel_size,
            padding=int((emb_kernel_size - 1) / 2),
        )

    def get_pitch_embedding_train(

        self, x: torch.Tensor, target: torch.Tensor, dr: torch.IntTensor, mask: torch.Tensor

    ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]:
        """

        Shapes:

            x: :math: `[B, T_src, C]`

            target: :math: `[B, 1, T_max2]`

            dr: :math: `[B, T_src]`

            mask: :math: `[B, T_src]`

        """
        pitch_pred = self.pitch_predictor(x, mask)  # [B, T_src, C_hidden], [B, T_src] --> [B, T_src]
        pitch_pred.unsqueeze_(1)  # --> [B, 1, T_src]
        avg_pitch_target = average_over_durations(target, dr)  # [B, 1, T_mel], [B, T_src] --> [B, 1, T_src]
        pitch_emb = self.pitch_emb(avg_pitch_target)  # [B, 1, T_src] --> [B, C_hidden, T_src]
        return pitch_pred, avg_pitch_target, pitch_emb

    def get_pitch_embedding(

        self,

        x: torch.Tensor,

        mask: torch.Tensor,

        pitch_transform: Callable,

        pitch_mean: torch.Tensor,

        pitch_std: torch.Tensor,

    ) -> torch.Tensor:
        pitch_pred = self.pitch_predictor(x, mask)
        if pitch_transform is not None:
            pitch_pred = pitch_transform(pitch_pred, (~mask).sum(), pitch_mean, pitch_std)
        pitch_pred.unsqueeze_(1)
        pitch_emb_pred = self.pitch_emb(pitch_pred)
        return pitch_emb_pred, pitch_pred