positionwise_feed_forward.py 1.9 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758
  1. #!/usr/bin/env python3
  2. # -*- coding: utf-8 -*-
  3. # Copyright 2019 Shigeki Karita
  4. # Apache 2.0 (http://www.apache.org/licenses/LICENSE-2.0)
  5. """Positionwise feed forward layer definition."""
  6. import torch
  7. from funasr.modules.layer_norm import LayerNorm
  8. class PositionwiseFeedForward(torch.nn.Module):
  9. """Positionwise feed forward layer.
  10. Args:
  11. idim (int): Input dimenstion.
  12. hidden_units (int): The number of hidden units.
  13. dropout_rate (float): Dropout rate.
  14. """
  15. def __init__(self, idim, hidden_units, dropout_rate, activation=torch.nn.ReLU()):
  16. """Construct an PositionwiseFeedForward object."""
  17. super(PositionwiseFeedForward, self).__init__()
  18. self.w_1 = torch.nn.Linear(idim, hidden_units)
  19. self.w_2 = torch.nn.Linear(hidden_units, idim)
  20. self.dropout = torch.nn.Dropout(dropout_rate)
  21. self.activation = activation
  22. def forward(self, x):
  23. """Forward function."""
  24. return self.w_2(self.dropout(self.activation(self.w_1(x))))
  25. class PositionwiseFeedForwardDecoderSANM(torch.nn.Module):
  26. """Positionwise feed forward layer.
  27. Args:
  28. idim (int): Input dimenstion.
  29. hidden_units (int): The number of hidden units.
  30. dropout_rate (float): Dropout rate.
  31. """
  32. def __init__(self, idim, hidden_units, dropout_rate, adim=None, activation=torch.nn.ReLU()):
  33. """Construct an PositionwiseFeedForward object."""
  34. super(PositionwiseFeedForwardDecoderSANM, self).__init__()
  35. self.w_1 = torch.nn.Linear(idim, hidden_units)
  36. self.w_2 = torch.nn.Linear(hidden_units, idim if adim is None else adim, bias=False)
  37. self.dropout = torch.nn.Dropout(dropout_rate)
  38. self.activation = activation
  39. self.norm = LayerNorm(hidden_units)
  40. def forward(self, x):
  41. """Forward function."""
  42. return self.w_2(self.norm(self.dropout(self.activation(self.w_1(x)))))