ReactSeq / onmt /modules /__init__.py
Oopstom's picture
Upload 313 files
c668e80 verified
raw
history blame
1.09 kB
""" Attention and normalization modules """
from onmt.modules.util_class import Elementwise
from onmt.modules.gate import context_gate_factory, ContextGate
from onmt.modules.global_attention import GlobalAttention
from onmt.modules.conv_multi_step_attention import ConvMultiStepAttention
from onmt.modules.copy_generator import CopyGenerator, CopyGeneratorLoss
from onmt.modules.multi_headed_attn import MultiHeadedAttention
from onmt.modules.embeddings import Embeddings, PositionalEncoding
from onmt.modules.weight_norm import WeightNormConv2d
from onmt.modules.average_attn import AverageAttention
from onmt.modules.alibi_position_bias import AlibiPositionalBias
from onmt.modules.rmsnorm import RMSNorm
__all__ = [
"Elementwise",
"context_gate_factory",
"ContextGate",
"GlobalAttention",
"ConvMultiStepAttention",
"CopyGenerator",
"CopyGeneratorLoss",
"CopyGeneratorLMLossCompute",
"MultiHeadedAttention",
"Embeddings",
"PositionalEncoding",
"AlibiPositionalBias",
"WeightNormConv2d",
"AverageAttention",
"RMSNorm",
]