WebashalarForML's picture
Upload 7 files
fcd0a70 verified
raw
history blame
10.4 kB
import torch
import torch.nn.functional as F
from torch import nn
def create_projection_layer(hidden_size: int, dropout: float, out_dim: int = None) -> nn.Sequential:
"""
Creates a projection layer with specified configurations.
"""
if out_dim is None:
out_dim = hidden_size
return nn.Sequential(
nn.Linear(hidden_size, out_dim * 4),
nn.ReLU(),
nn.Dropout(dropout),
nn.Linear(out_dim * 4, out_dim)
)
class SpanQuery(nn.Module):
def __init__(self, hidden_size, max_width, trainable=True):
super().__init__()
self.query_seg = nn.Parameter(torch.randn(hidden_size, max_width))
nn.init.uniform_(self.query_seg, a=-1, b=1)
if not trainable:
self.query_seg.requires_grad = False
self.project = nn.Sequential(
nn.Linear(hidden_size, hidden_size),
nn.ReLU()
)
def forward(self, h, *args):
# h of shape [B, L, D]
# query_seg of shape [D, max_width]
span_rep = torch.einsum('bld, ds->blsd', h, self.query_seg)
return self.project(span_rep)
class SpanMLP(nn.Module):
def __init__(self, hidden_size, max_width):
super().__init__()
self.mlp = nn.Linear(hidden_size, hidden_size * max_width)
def forward(self, h, *args):
# h of shape [B, L, D]
# query_seg of shape [D, max_width]
B, L, D = h.size()
span_rep = self.mlp(h)
span_rep = span_rep.view(B, L, -1, D)
return span_rep.relu()
class SpanCAT(nn.Module):
def __init__(self, hidden_size, max_width):
super().__init__()
self.max_width = max_width
self.query_seg = nn.Parameter(torch.randn(128, max_width))
self.project = nn.Sequential(
nn.Linear(hidden_size + 128, hidden_size),
nn.ReLU()
)
def forward(self, h, *args):
# h of shape [B, L, D]
# query_seg of shape [D, max_width]
B, L, D = h.size()
h = h.view(B, L, 1, D).repeat(1, 1, self.max_width, 1)
q = self.query_seg.view(1, 1, self.max_width, -1).repeat(B, L, 1, 1)
span_rep = torch.cat([h, q], dim=-1)
span_rep = self.project(span_rep)
return span_rep
class SpanConvBlock(nn.Module):
def __init__(self, hidden_size, kernel_size, span_mode='conv_normal'):
super().__init__()
if span_mode == 'conv_conv':
self.conv = nn.Conv1d(hidden_size, hidden_size,
kernel_size=kernel_size)
# initialize the weights
nn.init.kaiming_uniform_(self.conv.weight, nonlinearity='relu')
elif span_mode == 'conv_max':
self.conv = nn.MaxPool1d(kernel_size=kernel_size, stride=1)
elif span_mode == 'conv_mean' or span_mode == 'conv_sum':
self.conv = nn.AvgPool1d(kernel_size=kernel_size, stride=1)
self.span_mode = span_mode
self.pad = kernel_size - 1
def forward(self, x):
x = torch.einsum('bld->bdl', x)
if self.pad > 0:
x = F.pad(x, (0, self.pad), "constant", 0)
x = self.conv(x)
if self.span_mode == "conv_sum":
x = x * (self.pad + 1)
return torch.einsum('bdl->bld', x)
class SpanConv(nn.Module):
def __init__(self, hidden_size, max_width, span_mode):
super().__init__()
kernels = [i + 2 for i in range(max_width - 1)]
self.convs = nn.ModuleList()
for kernel in kernels:
self.convs.append(SpanConvBlock(hidden_size, kernel, span_mode))
self.project = nn.Sequential(
nn.ReLU(),
nn.Linear(hidden_size, hidden_size)
)
def forward(self, x, *args):
span_reps = [x]
for conv in self.convs:
h = conv(x)
span_reps.append(h)
span_reps = torch.stack(span_reps, dim=-2)
return self.project(span_reps)
class SpanEndpointsBlock(nn.Module):
def __init__(self, kernel_size):
super().__init__()
self.kernel_size = kernel_size
def forward(self, x):
B, L, D = x.size()
span_idx = torch.LongTensor(
[[i, i + self.kernel_size - 1] for i in range(L)]).to(x.device)
x = F.pad(x, (0, 0, 0, self.kernel_size - 1), "constant", 0)
# endrep
start_end_rep = torch.index_select(x, dim=1, index=span_idx.view(-1))
start_end_rep = start_end_rep.view(B, L, 2, D)
return start_end_rep
class ConvShare(nn.Module):
def __init__(self, hidden_size, max_width):
super().__init__()
self.max_width = max_width
self.conv_weigth = nn.Parameter(
torch.randn(hidden_size, hidden_size, max_width))
nn.init.kaiming_uniform_(self.conv_weigth, nonlinearity='relu')
self.project = nn.Sequential(
nn.ReLU(),
nn.Linear(hidden_size, hidden_size)
)
def forward(self, x, *args):
span_reps = []
x = torch.einsum('bld->bdl', x)
for i in range(self.max_width):
pad = i
x_i = F.pad(x, (0, pad), "constant", 0)
conv_w = self.conv_weigth[:, :, :i + 1]
out_i = F.conv1d(x_i, conv_w)
span_reps.append(out_i.transpose(-1, -2))
out = torch.stack(span_reps, dim=-2)
return self.project(out)
def extract_elements(sequence, indices):
B, L, D = sequence.shape
K = indices.shape[1]
# Expand indices to [B, K, D]
expanded_indices = indices.unsqueeze(2).expand(-1, -1, D)
# Gather the elements
extracted_elements = torch.gather(sequence, 1, expanded_indices)
return extracted_elements
class SpanMarker(nn.Module):
def __init__(self, hidden_size, max_width, dropout=0.4):
super().__init__()
self.max_width = max_width
self.project_start = nn.Sequential(
nn.Linear(hidden_size, hidden_size * 2, bias=True),
nn.ReLU(),
nn.Dropout(dropout),
nn.Linear(hidden_size * 2, hidden_size, bias=True),
)
self.project_end = nn.Sequential(
nn.Linear(hidden_size, hidden_size * 2, bias=True),
nn.ReLU(),
nn.Dropout(dropout),
nn.Linear(hidden_size * 2, hidden_size, bias=True),
)
self.out_project = nn.Linear(hidden_size * 2, hidden_size, bias=True)
def forward(self, h, span_idx):
# h of shape [B, L, D]
# query_seg of shape [D, max_width]
B, L, D = h.size()
# project start and end
start_rep = self.project_start(h)
end_rep = self.project_end(h)
start_span_rep = extract_elements(start_rep, span_idx[:, :, 0])
end_span_rep = extract_elements(end_rep, span_idx[:, :, 1])
# concat start and end
cat = torch.cat([start_span_rep, end_span_rep], dim=-1).relu()
# project
cat = self.out_project(cat)
# reshape
return cat.view(B, L, self.max_width, D)
class SpanMarkerV0(nn.Module):
"""
Marks and projects span endpoints using an MLP.
"""
def __init__(self, hidden_size: int, max_width: int, dropout: float = 0.4):
super().__init__()
self.max_width = max_width
self.project_start = create_projection_layer(hidden_size, dropout)
self.project_end = create_projection_layer(hidden_size, dropout)
self.out_project = create_projection_layer(hidden_size * 2, dropout, hidden_size)
def forward(self, h: torch.Tensor, span_idx: torch.Tensor) -> torch.Tensor:
B, L, D = h.size()
start_rep = self.project_start(h)
end_rep = self.project_end(h)
start_span_rep = extract_elements(start_rep, span_idx[:, :, 0])
end_span_rep = extract_elements(end_rep, span_idx[:, :, 1])
cat = torch.cat([start_span_rep, end_span_rep], dim=-1).relu()
return self.out_project(cat).view(B, L, self.max_width, D)
class ConvShareV2(nn.Module):
def __init__(self, hidden_size, max_width):
super().__init__()
self.max_width = max_width
self.conv_weigth = nn.Parameter(
torch.randn(hidden_size, hidden_size, max_width)
)
nn.init.xavier_normal_(self.conv_weigth)
def forward(self, x, *args):
span_reps = []
x = torch.einsum('bld->bdl', x)
for i in range(self.max_width):
pad = i
x_i = F.pad(x, (0, pad), "constant", 0)
conv_w = self.conv_weigth[:, :, :i + 1]
out_i = F.conv1d(x_i, conv_w)
span_reps.append(out_i.transpose(-1, -2))
out = torch.stack(span_reps, dim=-2)
return out
class SpanRepLayer(nn.Module):
"""
Various span representation approaches
"""
def __init__(self, hidden_size, max_width, span_mode, **kwargs):
super().__init__()
if span_mode == 'marker':
self.span_rep_layer = SpanMarker(hidden_size, max_width, **kwargs)
elif span_mode == 'markerV0':
self.span_rep_layer = SpanMarkerV0(hidden_size, max_width, **kwargs)
elif span_mode == 'query':
self.span_rep_layer = SpanQuery(
hidden_size, max_width, trainable=True)
elif span_mode == 'mlp':
self.span_rep_layer = SpanMLP(hidden_size, max_width)
elif span_mode == 'cat':
self.span_rep_layer = SpanCAT(hidden_size, max_width)
elif span_mode == 'conv_conv':
self.span_rep_layer = SpanConv(
hidden_size, max_width, span_mode='conv_conv')
elif span_mode == 'conv_max':
self.span_rep_layer = SpanConv(
hidden_size, max_width, span_mode='conv_max')
elif span_mode == 'conv_mean':
self.span_rep_layer = SpanConv(
hidden_size, max_width, span_mode='conv_mean')
elif span_mode == 'conv_sum':
self.span_rep_layer = SpanConv(
hidden_size, max_width, span_mode='conv_sum')
elif span_mode == 'conv_share':
self.span_rep_layer = ConvShare(hidden_size, max_width)
else:
raise ValueError(f'Unknown span mode {span_mode}')
def forward(self, x, *args):
return self.span_rep_layer(x, *args)