activations module

Activation functions Element-wise transformations without trainable parameters

class BiLU(alpha=1.0, delta=0.9)[source]

Bases: torch.nn.modules.module.Module

Bijective linear unit f(x) = alpha + delta*sign(x)

Initializes internal Module state, shared by both nn.Module and ScriptModule.

forward(input)[source]

Output of the BiLU activation function

training: bool[source]
class NormBiTanh(alpha=0.3)[source]

Bases: torch.nn.modules.module.Module

Bijective Normalized Tanh layer f(x) = alpha*Tanh + (1-alpha)*x activation

Initializes internal Module state, shared by both nn.Module and ScriptModule.

forward(input)[source]

Output of the NormBiTanh activation function

training: bool[source]