gpt_and_prejudice / activation.py
mariamkhmahran
refactoring
b41eca1
raw
history blame contribute delete
369 Bytes
import torch
import torch.nn as nn
class GELU(nn.Module):
'''
GELU (Gausian Error Linear Unit) activation function.
'''
def __init__(self):
super().__init__()
def forward(self, x):
return 0.5 * x * (1 + torch.tanh(
torch.sqrt(torch.tensor(2.0 / torch.pi)) *
(x + 0.044715 * torch.pow(x, 3))
))