Hecheng0625's picture
Upload 409 files
c968fc3 verified
from typing import List, Optional, Tuple
import torch
from torch import nn
from modules.wenet_extractor.utils.common import get_activation, get_rnn
def ApplyPadding(input, padding, pad_value) -> torch.Tensor:
"""
Args:
input: [bs, max_time_step, dim]
padding: [bs, max_time_step]
"""
return padding * pad_value + input * (1 - padding)
class PredictorBase(torch.nn.Module):
# NOTE(Mddct): We can use ABC abstract here, but
# keep this class simple enough for now
def __init__(self) -> None:
super().__init__()
def init_state(
self, batch_size: int, device: torch.device, method: str = "zero"
) -> List[torch.Tensor]:
_, _, _ = batch_size, method, device
raise NotImplementedError("this is a base precictor")
def batch_to_cache(self, cache: List[torch.Tensor]) -> List[List[torch.Tensor]]:
_ = cache
raise NotImplementedError("this is a base precictor")
def cache_to_batch(self, cache: List[List[torch.Tensor]]) -> List[torch.Tensor]:
_ = cache
raise NotImplementedError("this is a base precictor")
def forward(
self,
input: torch.Tensor,
cache: Optional[List[torch.Tensor]] = None,
):
(
_,
_,
) = (
input,
cache,
)
raise NotImplementedError("this is a base precictor")
def forward_step(
self, input: torch.Tensor, padding: torch.Tensor, cache: List[torch.Tensor]
) -> Tuple[torch.Tensor, List[torch.Tensor]]:
(
_,
_,
_,
) = (
input,
padding,
cache,
)
raise NotImplementedError("this is a base precictor")
class RNNPredictor(PredictorBase):
def __init__(
self,
voca_size: int,
embed_size: int,
output_size: int,
embed_dropout: float,
hidden_size: int,
num_layers: int,
bias: bool = True,
rnn_type: str = "lstm",
dropout: float = 0.1,
) -> None:
super().__init__()
self.n_layers = num_layers
self.hidden_size = hidden_size
# disable rnn base out projection
self.embed = nn.Embedding(voca_size, embed_size)
self.dropout = nn.Dropout(embed_dropout)
# NOTE(Mddct): rnn base from torch not support layer norm
# will add layer norm and prune value in cell and layer
# ref: https://github.com/Mddct/neural-lm/blob/main/models/gru_cell.py
self.rnn = get_rnn(rnn_type=rnn_type)(
input_size=embed_size,
hidden_size=hidden_size,
num_layers=num_layers,
bias=bias,
batch_first=True,
dropout=dropout,
)
self.projection = nn.Linear(hidden_size, output_size)
def forward(
self,
input: torch.Tensor,
cache: Optional[List[torch.Tensor]] = None,
) -> torch.Tensor:
"""
Args:
input (torch.Tensor): [batch, max_time).
padding (torch.Tensor): [batch, max_time]
cache : rnn predictor cache[0] == state_m
cache[1] == state_c
Returns:
output: [batch, max_time, output_size]
"""
# NOTE(Mddct): we don't use pack input format
embed = self.embed(input) # [batch, max_time, emb_size]
embed = self.dropout(embed)
states: Optional[Tuple[torch.Tensor, torch.Tensor]] = None
if cache is None:
state = self.init_state(batch_size=input.size(0), device=input.device)
states = (state[0], state[1])
else:
assert len(cache) == 2
states = (cache[0], cache[1])
out, (m, c) = self.rnn(embed, states)
out = self.projection(out)
# NOTE(Mddct): Although we don't use staate in transducer
# training forward, we need make it right for padding value
# so we create forward_step for infering, forward for training
_, _ = m, c
return out
def batch_to_cache(self, cache: List[torch.Tensor]) -> List[List[torch.Tensor]]:
"""
Args:
cache: [state_m, state_c]
state_ms: [1*n_layers, bs, ...]
state_cs: [1*n_layers, bs, ...]
Returns:
new_cache: [[state_m_1, state_c_1], [state_m_2, state_c_2]...]
"""
assert len(cache) == 2
state_ms = cache[0]
state_cs = cache[1]
assert state_ms.size(1) == state_cs.size(1)
new_cache: List[List[torch.Tensor]] = []
for state_m, state_c in zip(
torch.split(state_ms, 1, dim=1), torch.split(state_cs, 1, dim=1)
):
new_cache.append([state_m, state_c])
return new_cache
def cache_to_batch(self, cache: List[List[torch.Tensor]]) -> List[torch.Tensor]:
"""
Args:
cache : [[state_m_1, state_c_1], [state_m_1, state_c_1]...]
Returns:
new_caceh: [state_ms, state_cs],
state_ms: [1*n_layers, bs, ...]
state_cs: [1*n_layers, bs, ...]
"""
state_ms = torch.cat([states[0] for states in cache], dim=1)
state_cs = torch.cat([states[1] for states in cache], dim=1)
return [state_ms, state_cs]
def init_state(
self,
batch_size: int,
device: torch.device,
method: str = "zero",
) -> List[torch.Tensor]:
assert batch_size > 0
# TODO(Mddct): xavier init method
_ = method
return [
torch.zeros(1 * self.n_layers, batch_size, self.hidden_size, device=device),
torch.zeros(1 * self.n_layers, batch_size, self.hidden_size, device=device),
]
def forward_step(
self, input: torch.Tensor, padding: torch.Tensor, cache: List[torch.Tensor]
) -> Tuple[torch.Tensor, List[torch.Tensor]]:
"""
Args:
input (torch.Tensor): [batch_size, time_step=1]
padding (torch.Tensor): [batch_size,1], 1 is padding value
cache : rnn predictor cache[0] == state_m
cache[1] == state_c
"""
assert len(cache) == 2
state_m, state_c = cache[0], cache[1]
embed = self.embed(input) # [batch, 1, emb_size]
embed = self.dropout(embed)
out, (m, c) = self.rnn(embed, (state_m, state_c))
out = self.projection(out)
m = ApplyPadding(m, padding.unsqueeze(0), state_m)
c = ApplyPadding(c, padding.unsqueeze(0), state_c)
return (out, [m, c])
class EmbeddingPredictor(PredictorBase):
"""Embedding predictor
Described in:
https://arxiv.org/pdf/2109.07513.pdf
embed-> proj -> layer norm -> swish
"""
def __init__(
self,
voca_size: int,
embed_size: int,
embed_dropout: float,
n_head: int,
history_size: int = 2,
activation: str = "swish",
bias: bool = False,
layer_norm_epsilon: float = 1e-5,
) -> None:
super().__init__()
# multi head
self.num_heads = n_head
self.embed_size = embed_size
self.context_size = history_size + 1
self.pos_embed = torch.nn.Linear(
embed_size * self.context_size, self.num_heads, bias=bias
)
self.embed = nn.Embedding(voca_size, self.embed_size)
self.embed_dropout = nn.Dropout(p=embed_dropout)
self.ffn = nn.Linear(self.embed_size, self.embed_size)
self.norm = nn.LayerNorm(self.embed_size, eps=layer_norm_epsilon)
self.activatoin = get_activation(activation)
def init_state(
self, batch_size: int, device: torch.device, method: str = "zero"
) -> List[torch.Tensor]:
assert batch_size > 0
_ = method
return [
torch.zeros(
batch_size, self.context_size - 1, self.embed_size, device=device
),
]
def batch_to_cache(self, cache: List[torch.Tensor]) -> List[List[torch.Tensor]]:
"""
Args:
cache : [history]
history: [bs, ...]
Returns:
new_ache : [[history_1], [history_2], [history_3]...]
"""
assert len(cache) == 1
cache_0 = cache[0]
history: List[List[torch.Tensor]] = []
for h in torch.split(cache_0, 1, dim=0):
history.append([h])
return history
def cache_to_batch(self, cache: List[List[torch.Tensor]]) -> List[torch.Tensor]:
"""
Args:
cache : [[history_1], [history_2], [history3]...]
Returns:
new_caceh: [history],
history: [bs, ...]
"""
history = torch.cat([h[0] for h in cache], dim=0)
return [history]
def forward(self, input: torch.Tensor, cache: Optional[List[torch.Tensor]] = None):
"""forward for training"""
input = self.embed(input) # [bs, seq_len, embed]
input = self.embed_dropout(input)
if cache is None:
zeros = self.init_state(input.size(0), device=input.device)[0]
else:
assert len(cache) == 1
zeros = cache[0]
input = torch.cat(
(zeros, input), dim=1
) # [bs, context_size-1 + seq_len, embed]
input = input.unfold(1, self.context_size, 1).permute(
0, 1, 3, 2
) # [bs, seq_len, context_size, embed]
# multi head pos: [n_head, embed, context_size]
multi_head_pos = self.pos_embed.weight.view(
self.num_heads, self.embed_size, self.context_size
)
# broadcast dot attenton
input_expand = input.unsqueeze(2) # [bs, seq_len, 1, context_size, embed]
multi_head_pos = multi_head_pos.permute(
0, 2, 1
) # [num_heads, context_size, embed]
# [bs, seq_len, num_heads, context_size, embed]
weight = input_expand * multi_head_pos
weight = weight.sum(dim=-1, keepdim=False).unsqueeze(
3
) # [bs, seq_len, num_heads, 1, context_size]
output = weight.matmul(input_expand).squeeze(
dim=3
) # [bs, seq_len, num_heads, embed]
output = output.sum(dim=2) # [bs, seq_len, embed]
output = output / (self.num_heads * self.context_size)
output = self.ffn(output)
output = self.norm(output)
output = self.activatoin(output)
return output
def forward_step(
self,
input: torch.Tensor,
padding: torch.Tensor,
cache: List[torch.Tensor],
) -> Tuple[torch.Tensor, List[torch.Tensor]]:
"""forward step for inference
Args:
input (torch.Tensor): [batch_size, time_step=1]
padding (torch.Tensor): [batch_size,1], 1 is padding value
cache: for embedding predictor, cache[0] == history
"""
assert input.size(1) == 1
assert len(cache) == 1
history = cache[0]
assert history.size(1) == self.context_size - 1
input = self.embed(input) # [bs, 1, embed]
input = self.embed_dropout(input)
context_input = torch.cat((history, input), dim=1)
input_expand = context_input.unsqueeze(1).unsqueeze(
2
) # [bs, 1, 1, context_size, embed]
# multi head pos: [n_head, embed, context_size]
multi_head_pos = self.pos_embed.weight.view(
self.num_heads, self.embed_size, self.context_size
)
multi_head_pos = multi_head_pos.permute(
0, 2, 1
) # [num_heads, context_size, embed]
# [bs, 1, num_heads, context_size, embed]
weight = input_expand * multi_head_pos
weight = weight.sum(dim=-1, keepdim=False).unsqueeze(
3
) # [bs, 1, num_heads, 1, context_size]
output = weight.matmul(input_expand).squeeze(dim=3) # [bs, 1, num_heads, embed]
output = output.sum(dim=2) # [bs, 1, embed]
output = output / (self.num_heads * self.context_size)
output = self.ffn(output)
output = self.norm(output)
output = self.activatoin(output)
new_cache = context_input[:, 1:, :]
# TODO(Mddct): we need padding new_cache in future
# new_cache = ApplyPadding(history, padding, new_cache)
return (output, [new_cache])
class ConvPredictor(PredictorBase):
def __init__(
self,
voca_size: int,
embed_size: int,
embed_dropout: float,
history_size: int = 2,
activation: str = "relu",
bias: bool = False,
layer_norm_epsilon: float = 1e-5,
) -> None:
super().__init__()
assert history_size >= 0
self.embed_size = embed_size
self.context_size = history_size + 1
self.embed = nn.Embedding(voca_size, self.embed_size)
self.embed_dropout = nn.Dropout(p=embed_dropout)
self.conv = nn.Conv1d(
in_channels=embed_size,
out_channels=embed_size,
kernel_size=self.context_size,
padding=0,
groups=embed_size,
bias=bias,
)
self.norm = nn.LayerNorm(embed_size, eps=layer_norm_epsilon)
self.activatoin = get_activation(activation)
def init_state(
self, batch_size: int, device: torch.device, method: str = "zero"
) -> List[torch.Tensor]:
assert batch_size > 0
assert method == "zero"
return [
torch.zeros(
batch_size, self.context_size - 1, self.embed_size, device=device
)
]
def cache_to_batch(self, cache: List[List[torch.Tensor]]) -> List[torch.Tensor]:
"""
Args:
cache : [[history_1], [history_2], [history3]...]
Returns:
new_caceh: [history],
history: [bs, ...]
"""
history = torch.cat([h[0] for h in cache], dim=0)
return [history]
def batch_to_cache(self, cache: List[torch.Tensor]) -> List[List[torch.Tensor]]:
"""
Args:
cache : [history]
history: [bs, ...]
Returns:
new_ache : [[history_1], [history_2], [history_3]...]
"""
assert len(cache) == 1
cache_0 = cache[0]
history: List[List[torch.Tensor]] = []
for h in torch.split(cache_0, 1, dim=0):
history.append([h])
return history
def forward(self, input: torch.Tensor, cache: Optional[List[torch.Tensor]] = None):
"""forward for training"""
input = self.embed(input) # [bs, seq_len, embed]
input = self.embed_dropout(input)
if cache is None:
zeros = self.init_state(input.size(0), device=input.device)[0]
else:
assert len(cache) == 1
zeros = cache[0]
input = torch.cat(
(zeros, input), dim=1
) # [bs, context_size-1 + seq_len, embed]
input = input.permute(0, 2, 1)
out = self.conv(input).permute(0, 2, 1)
out = self.activatoin(self.norm(out))
return out
def forward_step(
self, input: torch.Tensor, padding: torch.Tensor, cache: List[torch.Tensor]
) -> Tuple[torch.Tensor, List[torch.Tensor]]:
"""forward step for inference
Args:
input (torch.Tensor): [batch_size, time_step=1]
padding (torch.Tensor): [batch_size,1], 1 is padding value
cache: for embedding predictor, cache[0] == history
"""
assert input.size(1) == 1
assert len(cache) == 1
history = cache[0]
assert history.size(1) == self.context_size - 1
input = self.embed(input) # [bs, 1, embed]
input = self.embed_dropout(input)
context_input = torch.cat((history, input), dim=1)
input = context_input.permute(0, 2, 1)
out = self.conv(input).permute(0, 2, 1)
out = self.activatoin(self.norm(out))
new_cache = context_input[:, 1:, :]
# TODO(Mddct): apply padding in future
return (out, [new_cache])