| 1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980 |
- from typing import Optional
- import torch
- import torch.nn as nn
- import numpy as np
- class MakePadMask(nn.Module):
- def __init__(self, max_seq_len=512, flip=True):
- super().__init__()
- if flip:
- self.mask_pad = torch.Tensor(1 - np.tri(max_seq_len)).type(torch.bool)
- else:
- self.mask_pad = torch.Tensor(np.tri(max_seq_len)).type(torch.bool)
-
- def forward(self, lengths, xs=None, length_dim=-1, maxlen=None):
- """Make mask tensor containing indices of padded part.
- This implementation creates the same mask tensor with original make_pad_mask,
- which can be converted into onnx format.
- Dimension length of xs should be 2 or 3.
- """
- if length_dim == 0:
- raise ValueError("length_dim cannot be 0: {}".format(length_dim))
- if xs is not None and len(xs.shape) == 3:
- if length_dim == 1:
- lengths = lengths.unsqueeze(1).expand(
- *xs.transpose(1, 2).shape[:2])
- else:
- lengths = lengths.unsqueeze(1).expand(*xs.shape[:2])
- if maxlen is not None:
- m = maxlen
- elif xs is not None:
- m = xs.shape[-1]
- else:
- m = torch.max(lengths)
- mask = self.mask_pad[lengths - 1][..., :m].type(torch.float32)
- if length_dim == 1:
- return mask.transpose(1, 2)
- else:
- return mask
- class sequence_mask(nn.Module):
- def __init__(self, max_seq_len=512, flip=True):
- super().__init__()
-
- def forward(self, lengths, max_seq_len=None, dtype=torch.float32, device=None):
- if max_seq_len is None:
- max_seq_len = lengths.max()
- row_vector = torch.arange(0, max_seq_len, 1).to(lengths.device)
- matrix = torch.unsqueeze(lengths, dim=-1)
- mask = row_vector < matrix
-
- return mask.type(dtype).to(device) if device is not None else mask.type(dtype)
- def normalize(input: torch.Tensor, p: float = 2.0, dim: int = 1, out: Optional[torch.Tensor] = None) -> torch.Tensor:
- if out is None:
- denom = input.norm(p, dim, keepdim=True).expand_as(input)
- return input / denom
- else:
- denom = input.norm(p, dim, keepdim=True).expand_as(input)
- return torch.div(input, denom, out=out)
- def subsequent_mask(size: torch.Tensor):
- return torch.ones(size, size).tril()
- def MakePadMask_test():
- feats_length = torch.tensor([10]).type(torch.long)
- mask_fn = MakePadMask()
- mask = mask_fn(feats_length)
- print(mask)
- if __name__ == '__main__':
- MakePadMask_test()
|