Skip to content

Commit

Permalink
Add dataclass decorate for praxis layers to make link happy
Browse files Browse the repository at this point in the history
Signed-off-by: Reese Wang <rewang@nvidia.com>
  • Loading branch information
zlsh80826 committed Dec 10, 2024
1 parent 0e912ad commit 09ca03e
Show file tree
Hide file tree
Showing 2 changed files with 12 additions and 2 deletions.
8 changes: 7 additions & 1 deletion transformer_engine/jax/praxis/module.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@
"""
Praxis Modules
"""
from dataclasses import field
from dataclasses import dataclass, field
from functools import partial
from typing import Callable, Iterable, Sequence, Tuple, Union

Expand All @@ -28,6 +28,7 @@ def _generate_ln_scale_init(scale_init):
return scale_init


@dataclass
class TransformerEngineBaseLayer(BaseLayer):
"""TransformerEngineBaseLayer"""

Expand Down Expand Up @@ -67,6 +68,7 @@ def create_layer(self, name, flax_module_cls):
self.create_child(name, flax_module_p.clone())


@dataclass
class LayerNorm(TransformerEngineBaseLayer):
"""LayerNorm"""

Expand Down Expand Up @@ -103,6 +105,7 @@ def __call__(self, x: JTensor) -> JTensor:
return self.layer_norm(x)


@dataclass
class FusedSoftmax(TransformerEngineBaseLayer):
"""FusedSoftmax"""

Expand All @@ -124,6 +127,7 @@ def __call__(self, x: JTensor, mask: JTensor = None, bias: JTensor = None) -> JT
return self.fused_softmax(x, mask, bias)


@dataclass
class Linear(TransformerEngineBaseLayer):
"""Linear"""

Expand Down Expand Up @@ -165,6 +169,7 @@ def __call__(self, x: JTensor) -> JTensor:
return self.linear(x)


@dataclass
class LayerNormLinear(TransformerEngineBaseLayer):
"""LayerNormLinear"""

Expand Down Expand Up @@ -228,6 +233,7 @@ def __call__(self, x: JTensor) -> JTensor:
return self.ln_linear(x)


@dataclass
class LayerNormMLP(TransformerEngineBaseLayer):
"""LayerNormMLP"""

Expand Down
6 changes: 5 additions & 1 deletion transformer_engine/jax/praxis/transformer.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@
"""
Praxis Modules related Transformer
"""
from dataclasses import field
from dataclasses import dataclass, field
from functools import partial
from typing import Optional, Sequence, Tuple
import warnings
Expand All @@ -22,6 +22,7 @@
from ..attention import AttnBiasType, AttnMaskType


@dataclass
class RelativePositionBiases(TransformerEngineBaseLayer):
"""RelativePositionBiases"""

Expand Down Expand Up @@ -67,6 +68,7 @@ def __call__(self, q_seqlen: JTensor, k_seqlen: JTensor, bidirectional: bool = T
return self.relative_position_bias(q_seqlen, k_seqlen, bidirectional)


@dataclass
class DotProductAttention(TransformerEngineBaseLayer):
"""DotProductAttention"""

Expand Down Expand Up @@ -125,6 +127,7 @@ def __call__(
)


@dataclass
class MultiHeadAttention(TransformerEngineBaseLayer):
"""MultiHeadAttention"""

Expand Down Expand Up @@ -258,6 +261,7 @@ def __call__(
)


@dataclass
class TransformerLayer(TransformerEngineBaseLayer):
"""TransformerLayer"""

Expand Down

0 comments on commit 09ca03e

Please sign in to comment.