diff --git a/build/torch26-cxx11-cu118-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch26-cxx11-cu118-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index 67eeec1217118899a04322a83fafa12857ebcff4..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu118-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b2fdb7378a1c907c3ff3ad0a5134a0a8ce4a464196404436470d7b4eb77ec305 -size 2957296 diff --git a/build/torch26-cxx11-cu118-x86_64-linux/activation/_ops.py b/build/torch26-cxx11-cu118-x86_64-linux/activation/_ops.py deleted file mode 100644 index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu118-x86_64-linux/activation/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu124-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch26-cxx11-cu124-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index c7b50b41f57eeef90b1f0a959e07e295ac6d1308..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu124-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:5baac6228e04fbb209cbc90a24702c14f4eb52d2698cea12a766d77412622096 -size 2981880 diff --git a/build/torch26-cxx11-cu124-x86_64-linux/activation/_ops.py b/build/torch26-cxx11-cu124-x86_64-linux/activation/_ops.py deleted file mode 100644 index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu124-x86_64-linux/activation/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu126-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch26-cxx11-cu126-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index d30d11725ae8ec2ee197902f1acf9c6eb16c1f08..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu126-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2d5c0095b931923008435d361c1871e97ff2ef04100e93205f09e65316f307f3 -size 2994704 diff --git a/build/torch26-cxx11-cu126-x86_64-linux/activation/_ops.py b/build/torch26-cxx11-cu126-x86_64-linux/activation/_ops.py deleted file mode 100644 index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu126-x86_64-linux/activation/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-rocm62-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch26-cxx11-rocm62-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index 21c1a09ee3c8d9b1e90ff323d1d25332a1d0bbe3..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-rocm62-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:5c29fccf3f62ac3e3b7ff59e898d31ae38f3484bfe762f6767b8bc8cedf1af01 -size 2660632 diff --git a/build/torch26-cxx11-rocm62-x86_64-linux/activation/_ops.py b/build/torch26-cxx11-rocm62-x86_64-linux/activation/_ops.py deleted file mode 100644 index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-rocm62-x86_64-linux/activation/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx98-cu118-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch26-cxx98-cu118-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index 5e6e767041018f2157d2a53d7479cf80cbdbcaef..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu118-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:176b8610ed2b9650c68347ec2f1d9e99b653170b4fd4f6f3540731f3fd78e98b -size 2949936 diff --git a/build/torch26-cxx98-cu118-x86_64-linux/activation/_ops.py b/build/torch26-cxx98-cu118-x86_64-linux/activation/_ops.py deleted file mode 100644 index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu118-x86_64-linux/activation/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx98-cu124-x86_64-linux/activation/__init__.py b/build/torch26-cxx98-cu124-x86_64-linux/activation/__init__.py deleted file mode 100644 index fadce68f2fa0f130463f00a59c3436b822835e24..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu124-x86_64-linux/activation/__init__.py +++ /dev/null @@ -1,30 +0,0 @@ -import torch - -from . import layers -from ._ops import ops -from .poly_norm import PolyNormFunction -from .rms_norm import RMSNormFunction - - -def poly_norm( - x: torch.Tensor, - weight: torch.Tensor, - bias: torch.Tensor, - eps: float = 1e-6, -) -> None: - return PolyNormFunction.apply(x, weight, bias, eps) - - -def rms_norm( - x: torch.Tensor, - weight: torch.Tensor, - eps: float = 1e-6, -) -> None: - return RMSNormFunction.apply(x, weight, eps) - - -__all__ = [ - "poly_norm", - "layers", - "ops", -] diff --git a/build/torch26-cxx98-cu124-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch26-cxx98-cu124-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index c65d99af91dd3aaaffe08d2c572cf46f1a671fee..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu124-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:8075bbb5b339e0305d353003eb86a2b6a4d8a468907d821cefbed29e6e439c19 -size 2974640 diff --git a/build/torch26-cxx98-cu124-x86_64-linux/activation/_ops.py b/build/torch26-cxx98-cu124-x86_64-linux/activation/_ops.py deleted file mode 100644 index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu124-x86_64-linux/activation/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx98-cu124-x86_64-linux/activation/layers.py b/build/torch26-cxx98-cu124-x86_64-linux/activation/layers.py deleted file mode 100644 index 3824e5de50583f385215cf90adc03aff91653e2e..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu124-x86_64-linux/activation/layers.py +++ /dev/null @@ -1,46 +0,0 @@ -import torch -import torch.nn as nn -from torch.nn import init - -from .poly_norm import PolyNormFunction -from .rms_norm import RMSNormFunction - - -class PolyNorm(nn.Module): - def __init__(self, eps=1e-6, dtype: torch.dtype = torch.float32): - super().__init__() - self.weight = torch.nn.Parameter(torch.ones(3, dtype=dtype) / 3) - self.bias = torch.nn.Parameter(torch.zeros(1, dtype=dtype)) - self.eps = eps - - def forward( - self, - x: torch.Tensor, - ): - return PolyNormFunction.apply(x, self.weight, self.bias, self.eps) - - def reset_parameters(self) -> None: - """ - Resets parameters based on their initialization used in __init__. - """ - init.ones_(self.weight) - init.zeros_(self.bias) - - -class RMSNorm(nn.Module): - def __init__(self, dim: int, eps=1e-6, dtype: torch.dtype = torch.float32): - super().__init__() - self.weight = torch.nn.Parameter(torch.ones(dim, dtype=dtype)) - self.eps = eps - - def forward( - self, - x: torch.Tensor, - ): - return RMSNormFunction.apply(x, self.weight, self.eps) - - def reset_parameters(self) -> None: - """ - Resets parameters based on their initialization used in __init__. - """ - init.ones_(self.weight) diff --git a/build/torch26-cxx98-cu124-x86_64-linux/activation/poly_norm.py b/build/torch26-cxx98-cu124-x86_64-linux/activation/poly_norm.py deleted file mode 100644 index ce14e5cd8078de06d964775e34e1e668df32493e..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu124-x86_64-linux/activation/poly_norm.py +++ /dev/null @@ -1,41 +0,0 @@ -import torch - -from ._ops import ops - - -# Inherit from Function -class PolyNormFunction(torch.autograd.Function): - # Note that forward, setup_context, and backward are @staticmethods - @staticmethod - def forward(input, weight, bias, eps): - output = torch.empty_like(input) - ops.poly_norm(output, input, weight, bias, eps) - return output - - @staticmethod - # inputs is a Tuple of all of the inputs passed to forward. - # output is the output of the forward(). - def setup_context(ctx, inputs, output): - input, weight, bias, eps = inputs - ctx.save_for_backward(input, weight) - ctx.eps = eps - - # This function has only a single output, so it gets only one gradient - @staticmethod - def backward(ctx, output_grad): - input, weight = ctx.saved_tensors - eps = ctx.eps - - input_grad = torch.empty_like(input) if ctx.needs_input_grad[0] else None - weight_grad = torch.empty_like(weight) if ctx.needs_input_grad[1] else None - bias_grad = ( - torch.empty(1, dtype=weight.dtype, device=weight.device) - if ctx.needs_input_grad[2] - else None - ) - - ops.poly_norm_backward( - input_grad, weight_grad, bias_grad, output_grad, input, weight, eps - ) - - return input_grad, weight_grad, bias_grad, None diff --git a/build/torch26-cxx98-cu124-x86_64-linux/activation/rms_norm.py b/build/torch26-cxx98-cu124-x86_64-linux/activation/rms_norm.py deleted file mode 100644 index 53df35e855d5d1591bc6fceff50ba81afdb2c873..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu124-x86_64-linux/activation/rms_norm.py +++ /dev/null @@ -1,34 +0,0 @@ -import torch - -from ._ops import ops - - -# Inherit from Function -class RMSNormFunction(torch.autograd.Function): - # Note that forward, setup_context, and backward are @staticmethods - @staticmethod - def forward(input, weight, eps): - output = torch.empty_like(input) - ops.rms_norm(output, input, weight, eps) - return output - - @staticmethod - # inputs is a Tuple of all of the inputs passed to forward. - # output is the output of the forward(). - def setup_context(ctx, inputs, output): - input, weight, eps = inputs - ctx.save_for_backward(input, weight) - ctx.eps = eps - - # This function has only a single output, so it gets only one gradient - @staticmethod - def backward(ctx, output_grad): - input, weight = ctx.saved_tensors - eps = ctx.eps - - input_grad = torch.empty_like(input) if ctx.needs_input_grad[0] else None - weight_grad = torch.empty_like(weight) if ctx.needs_input_grad[1] else None - - ops.rms_norm_backward(input_grad, weight_grad, output_grad, input, weight, eps) - - return input_grad, weight_grad, None diff --git a/build/torch26-cxx98-cu126-x86_64-linux/activation/__init__.py b/build/torch26-cxx98-cu126-x86_64-linux/activation/__init__.py deleted file mode 100644 index fadce68f2fa0f130463f00a59c3436b822835e24..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-x86_64-linux/activation/__init__.py +++ /dev/null @@ -1,30 +0,0 @@ -import torch - -from . import layers -from ._ops import ops -from .poly_norm import PolyNormFunction -from .rms_norm import RMSNormFunction - - -def poly_norm( - x: torch.Tensor, - weight: torch.Tensor, - bias: torch.Tensor, - eps: float = 1e-6, -) -> None: - return PolyNormFunction.apply(x, weight, bias, eps) - - -def rms_norm( - x: torch.Tensor, - weight: torch.Tensor, - eps: float = 1e-6, -) -> None: - return RMSNormFunction.apply(x, weight, eps) - - -__all__ = [ - "poly_norm", - "layers", - "ops", -] diff --git a/build/torch26-cxx98-cu126-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch26-cxx98-cu126-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index c8dbf12bc93a1955e673270baa58f88c9189ed2d..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:614ef2bf9867f65bf8e09d861def1c554d384676aa58dfbfd73bf96241cb7171 -size 2987456 diff --git a/build/torch26-cxx98-cu126-x86_64-linux/activation/_ops.py b/build/torch26-cxx98-cu126-x86_64-linux/activation/_ops.py deleted file mode 100644 index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-x86_64-linux/activation/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx98-cu126-x86_64-linux/activation/layers.py b/build/torch26-cxx98-cu126-x86_64-linux/activation/layers.py deleted file mode 100644 index 3824e5de50583f385215cf90adc03aff91653e2e..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-x86_64-linux/activation/layers.py +++ /dev/null @@ -1,46 +0,0 @@ -import torch -import torch.nn as nn -from torch.nn import init - -from .poly_norm import PolyNormFunction -from .rms_norm import RMSNormFunction - - -class PolyNorm(nn.Module): - def __init__(self, eps=1e-6, dtype: torch.dtype = torch.float32): - super().__init__() - self.weight = torch.nn.Parameter(torch.ones(3, dtype=dtype) / 3) - self.bias = torch.nn.Parameter(torch.zeros(1, dtype=dtype)) - self.eps = eps - - def forward( - self, - x: torch.Tensor, - ): - return PolyNormFunction.apply(x, self.weight, self.bias, self.eps) - - def reset_parameters(self) -> None: - """ - Resets parameters based on their initialization used in __init__. - """ - init.ones_(self.weight) - init.zeros_(self.bias) - - -class RMSNorm(nn.Module): - def __init__(self, dim: int, eps=1e-6, dtype: torch.dtype = torch.float32): - super().__init__() - self.weight = torch.nn.Parameter(torch.ones(dim, dtype=dtype)) - self.eps = eps - - def forward( - self, - x: torch.Tensor, - ): - return RMSNormFunction.apply(x, self.weight, self.eps) - - def reset_parameters(self) -> None: - """ - Resets parameters based on their initialization used in __init__. - """ - init.ones_(self.weight) diff --git a/build/torch26-cxx98-cu126-x86_64-linux/activation/poly_norm.py b/build/torch26-cxx98-cu126-x86_64-linux/activation/poly_norm.py deleted file mode 100644 index ce14e5cd8078de06d964775e34e1e668df32493e..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-x86_64-linux/activation/poly_norm.py +++ /dev/null @@ -1,41 +0,0 @@ -import torch - -from ._ops import ops - - -# Inherit from Function -class PolyNormFunction(torch.autograd.Function): - # Note that forward, setup_context, and backward are @staticmethods - @staticmethod - def forward(input, weight, bias, eps): - output = torch.empty_like(input) - ops.poly_norm(output, input, weight, bias, eps) - return output - - @staticmethod - # inputs is a Tuple of all of the inputs passed to forward. - # output is the output of the forward(). - def setup_context(ctx, inputs, output): - input, weight, bias, eps = inputs - ctx.save_for_backward(input, weight) - ctx.eps = eps - - # This function has only a single output, so it gets only one gradient - @staticmethod - def backward(ctx, output_grad): - input, weight = ctx.saved_tensors - eps = ctx.eps - - input_grad = torch.empty_like(input) if ctx.needs_input_grad[0] else None - weight_grad = torch.empty_like(weight) if ctx.needs_input_grad[1] else None - bias_grad = ( - torch.empty(1, dtype=weight.dtype, device=weight.device) - if ctx.needs_input_grad[2] - else None - ) - - ops.poly_norm_backward( - input_grad, weight_grad, bias_grad, output_grad, input, weight, eps - ) - - return input_grad, weight_grad, bias_grad, None diff --git a/build/torch26-cxx98-cu126-x86_64-linux/activation/rms_norm.py b/build/torch26-cxx98-cu126-x86_64-linux/activation/rms_norm.py deleted file mode 100644 index 53df35e855d5d1591bc6fceff50ba81afdb2c873..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-x86_64-linux/activation/rms_norm.py +++ /dev/null @@ -1,34 +0,0 @@ -import torch - -from ._ops import ops - - -# Inherit from Function -class RMSNormFunction(torch.autograd.Function): - # Note that forward, setup_context, and backward are @staticmethods - @staticmethod - def forward(input, weight, eps): - output = torch.empty_like(input) - ops.rms_norm(output, input, weight, eps) - return output - - @staticmethod - # inputs is a Tuple of all of the inputs passed to forward. - # output is the output of the forward(). - def setup_context(ctx, inputs, output): - input, weight, eps = inputs - ctx.save_for_backward(input, weight) - ctx.eps = eps - - # This function has only a single output, so it gets only one gradient - @staticmethod - def backward(ctx, output_grad): - input, weight = ctx.saved_tensors - eps = ctx.eps - - input_grad = torch.empty_like(input) if ctx.needs_input_grad[0] else None - weight_grad = torch.empty_like(weight) if ctx.needs_input_grad[1] else None - - ops.rms_norm_backward(input_grad, weight_grad, output_grad, input, weight, eps) - - return input_grad, weight_grad, None diff --git a/build/torch27-cxx11-cu118-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch27-cxx11-cu118-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index 22f58c584491d644d7e0a5f52938366b2bfc014b..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu118-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:654d16d66565185dfd1a6f16e0b24d8fff83e12558c8862c322734e6b52e5cc0 -size 2957448 diff --git a/build/torch27-cxx11-cu118-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so b/build/torch27-cxx11-cu118-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..fee453a5f3d827219ce25b2491d12dd3883ef044 --- /dev/null +++ b/build/torch27-cxx11-cu118-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1655e52503ce7d0b7dabd55b97c1bd7d11071cbe0f80b9e810c443523638fd9b +size 2994312 diff --git a/build/torch27-cxx11-cu118-x86_64-linux/activation/_ops.py b/build/torch27-cxx11-cu118-x86_64-linux/activation/_ops.py index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..ecb351fdb0e6990bab29e42a4aecd1896b77a1b2 100644 --- a/build/torch27-cxx11-cu118-x86_64-linux/activation/_ops.py +++ b/build/torch27-cxx11-cu118-x86_64-linux/activation/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty +from . import _activation_cf68df1_dirty +ops = torch.ops._activation_cf68df1_dirty def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file + return f"_activation_cf68df1_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-cu126-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch27-cxx11-cu126-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index 1d4298b396205534b39df4f6014a0cab01c8ac06..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:eaf16ec2d17feb812fd485f168fbdd938880122d882edcdfb7fd8efcf3ed77af -size 2994736 diff --git a/build/torch27-cxx11-cu126-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so b/build/torch27-cxx11-cu126-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..1a2c10a5ac919201397afa3d9ffa2f8c49f434b9 --- /dev/null +++ b/build/torch27-cxx11-cu126-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:174dbe4375aa22fb34d9d23630b3bec4eeb95635ef681b665db0985e78cf5af3 +size 3027504 diff --git a/build/torch27-cxx11-cu126-x86_64-linux/activation/_ops.py b/build/torch27-cxx11-cu126-x86_64-linux/activation/_ops.py index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..ecb351fdb0e6990bab29e42a4aecd1896b77a1b2 100644 --- a/build/torch27-cxx11-cu126-x86_64-linux/activation/_ops.py +++ b/build/torch27-cxx11-cu126-x86_64-linux/activation/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty +from . import _activation_cf68df1_dirty +ops = torch.ops._activation_cf68df1_dirty def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file + return f"_activation_cf68df1_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-cu128-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch27-cxx11-cu128-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index e893ea914b71ecdba19784492df7b62bca7d20d0..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu128-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4188bd38f2c4d2b19a4a79e2234456fe29c28da064269abc0409de11c725f831 -size 3909704 diff --git a/build/torch27-cxx11-cu128-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so b/build/torch27-cxx11-cu128-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..af7500bce257ce738f3a19d5b1330bc6dae90856 --- /dev/null +++ b/build/torch27-cxx11-cu128-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91d71ca84a19b393c22b269226a7b4ddadbf1feec73a80bd45f655179c7a53f5 +size 3987512 diff --git a/build/torch27-cxx11-cu128-x86_64-linux/activation/_ops.py b/build/torch27-cxx11-cu128-x86_64-linux/activation/_ops.py index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..ecb351fdb0e6990bab29e42a4aecd1896b77a1b2 100644 --- a/build/torch27-cxx11-cu128-x86_64-linux/activation/_ops.py +++ b/build/torch27-cxx11-cu128-x86_64-linux/activation/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty +from . import _activation_cf68df1_dirty +ops = torch.ops._activation_cf68df1_dirty def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file + return f"_activation_cf68df1_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-rocm63-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so b/build/torch27-cxx11-rocm63-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so deleted file mode 100644 index d2a82f4c55f0b615552bf5c38e7ee20a868b9f1f..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-rocm63-x86_64-linux/activation/_activation_605f22e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:90da5f30c084a0316266d4886a052721347d40e775f03b6033894de5890331bf -size 2642880 diff --git a/build/torch27-cxx11-rocm63-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so b/build/torch27-cxx11-rocm63-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..d44ecea9b84cd8373497342429108adfcb3021cb --- /dev/null +++ b/build/torch27-cxx11-rocm63-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab1037bf6b41bf2be1d00a6a0ed01a97a5e4d64dd0abaf509492ad31eea0a576 +size 2642976 diff --git a/build/torch27-cxx11-rocm63-x86_64-linux/activation/_ops.py b/build/torch27-cxx11-rocm63-x86_64-linux/activation/_ops.py index 4f65fa5de7bb717f07527d6846085f48d18b7e1d..ecb351fdb0e6990bab29e42a4aecd1896b77a1b2 100644 --- a/build/torch27-cxx11-rocm63-x86_64-linux/activation/_ops.py +++ b/build/torch27-cxx11-rocm63-x86_64-linux/activation/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _activation_605f22e_dirty -ops = torch.ops._activation_605f22e_dirty +from . import _activation_cf68df1_dirty +ops = torch.ops._activation_cf68df1_dirty def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_activation_605f22e_dirty::{op_name}" \ No newline at end of file + return f"_activation_cf68df1_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu118-x86_64-linux/activation/__init__.py b/build/torch28-cxx11-cu126-x86_64-linux/activation/__init__.py similarity index 100% rename from build/torch26-cxx11-cu118-x86_64-linux/activation/__init__.py rename to build/torch28-cxx11-cu126-x86_64-linux/activation/__init__.py diff --git a/build/torch28-cxx11-cu126-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so b/build/torch28-cxx11-cu126-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..8cb08e411d00e0c33bd32a334b13f032afabf622 --- /dev/null +++ b/build/torch28-cxx11-cu126-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:012788f2064588edf60df24778dff33f8ca95e3b1aaf5243554735cd783dd7ed +size 3032488 diff --git a/build/torch28-cxx11-cu126-x86_64-linux/activation/_ops.py b/build/torch28-cxx11-cu126-x86_64-linux/activation/_ops.py new file mode 100644 index 0000000000000000000000000000000000000000..ecb351fdb0e6990bab29e42a4aecd1896b77a1b2 --- /dev/null +++ b/build/torch28-cxx11-cu126-x86_64-linux/activation/_ops.py @@ -0,0 +1,9 @@ +import torch +from . import _activation_cf68df1_dirty +ops = torch.ops._activation_cf68df1_dirty + +def add_op_namespace_prefix(op_name: str): + """ + Prefix op by namespace. + """ + return f"_activation_cf68df1_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu118-x86_64-linux/activation/layers.py b/build/torch28-cxx11-cu126-x86_64-linux/activation/layers.py similarity index 100% rename from build/torch26-cxx11-cu118-x86_64-linux/activation/layers.py rename to build/torch28-cxx11-cu126-x86_64-linux/activation/layers.py diff --git a/build/torch26-cxx11-cu118-x86_64-linux/activation/poly_norm.py b/build/torch28-cxx11-cu126-x86_64-linux/activation/poly_norm.py similarity index 100% rename from build/torch26-cxx11-cu118-x86_64-linux/activation/poly_norm.py rename to build/torch28-cxx11-cu126-x86_64-linux/activation/poly_norm.py diff --git a/build/torch26-cxx11-cu118-x86_64-linux/activation/rms_norm.py b/build/torch28-cxx11-cu126-x86_64-linux/activation/rms_norm.py similarity index 100% rename from build/torch26-cxx11-cu118-x86_64-linux/activation/rms_norm.py rename to build/torch28-cxx11-cu126-x86_64-linux/activation/rms_norm.py diff --git a/build/torch26-cxx11-cu124-x86_64-linux/activation/__init__.py b/build/torch28-cxx11-cu128-x86_64-linux/activation/__init__.py similarity index 100% rename from build/torch26-cxx11-cu124-x86_64-linux/activation/__init__.py rename to build/torch28-cxx11-cu128-x86_64-linux/activation/__init__.py diff --git a/build/torch28-cxx11-cu128-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so b/build/torch28-cxx11-cu128-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..fe1ba22a5e5b4dd967450df3fb72dcf989ff3c49 --- /dev/null +++ b/build/torch28-cxx11-cu128-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1a65b79b750f550a09e6a1142b5151b03b2a60ec6115a264e6d8de3cac7ee5d +size 4000920 diff --git a/build/torch28-cxx11-cu128-x86_64-linux/activation/_ops.py b/build/torch28-cxx11-cu128-x86_64-linux/activation/_ops.py new file mode 100644 index 0000000000000000000000000000000000000000..ecb351fdb0e6990bab29e42a4aecd1896b77a1b2 --- /dev/null +++ b/build/torch28-cxx11-cu128-x86_64-linux/activation/_ops.py @@ -0,0 +1,9 @@ +import torch +from . import _activation_cf68df1_dirty +ops = torch.ops._activation_cf68df1_dirty + +def add_op_namespace_prefix(op_name: str): + """ + Prefix op by namespace. + """ + return f"_activation_cf68df1_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu124-x86_64-linux/activation/layers.py b/build/torch28-cxx11-cu128-x86_64-linux/activation/layers.py similarity index 100% rename from build/torch26-cxx11-cu124-x86_64-linux/activation/layers.py rename to build/torch28-cxx11-cu128-x86_64-linux/activation/layers.py diff --git a/build/torch26-cxx11-cu124-x86_64-linux/activation/poly_norm.py b/build/torch28-cxx11-cu128-x86_64-linux/activation/poly_norm.py similarity index 100% rename from build/torch26-cxx11-cu124-x86_64-linux/activation/poly_norm.py rename to build/torch28-cxx11-cu128-x86_64-linux/activation/poly_norm.py diff --git a/build/torch26-cxx11-cu124-x86_64-linux/activation/rms_norm.py b/build/torch28-cxx11-cu128-x86_64-linux/activation/rms_norm.py similarity index 100% rename from build/torch26-cxx11-cu124-x86_64-linux/activation/rms_norm.py rename to build/torch28-cxx11-cu128-x86_64-linux/activation/rms_norm.py diff --git a/build/torch26-cxx11-cu126-x86_64-linux/activation/__init__.py b/build/torch28-cxx11-cu129-x86_64-linux/activation/__init__.py similarity index 100% rename from build/torch26-cxx11-cu126-x86_64-linux/activation/__init__.py rename to build/torch28-cxx11-cu129-x86_64-linux/activation/__init__.py diff --git a/build/torch28-cxx11-cu129-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so b/build/torch28-cxx11-cu129-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..eb2ce397183a075e5f4923b2d2f551bf761331ef --- /dev/null +++ b/build/torch28-cxx11-cu129-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd38039c3401b0f6a136f1761c7f396f5954f05e16d78ed1600d8325c1221781 +size 4059256 diff --git a/build/torch28-cxx11-cu129-x86_64-linux/activation/_ops.py b/build/torch28-cxx11-cu129-x86_64-linux/activation/_ops.py new file mode 100644 index 0000000000000000000000000000000000000000..ecb351fdb0e6990bab29e42a4aecd1896b77a1b2 --- /dev/null +++ b/build/torch28-cxx11-cu129-x86_64-linux/activation/_ops.py @@ -0,0 +1,9 @@ +import torch +from . import _activation_cf68df1_dirty +ops = torch.ops._activation_cf68df1_dirty + +def add_op_namespace_prefix(op_name: str): + """ + Prefix op by namespace. + """ + return f"_activation_cf68df1_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu126-x86_64-linux/activation/layers.py b/build/torch28-cxx11-cu129-x86_64-linux/activation/layers.py similarity index 100% rename from build/torch26-cxx11-cu126-x86_64-linux/activation/layers.py rename to build/torch28-cxx11-cu129-x86_64-linux/activation/layers.py diff --git a/build/torch26-cxx11-cu126-x86_64-linux/activation/poly_norm.py b/build/torch28-cxx11-cu129-x86_64-linux/activation/poly_norm.py similarity index 100% rename from build/torch26-cxx11-cu126-x86_64-linux/activation/poly_norm.py rename to build/torch28-cxx11-cu129-x86_64-linux/activation/poly_norm.py diff --git a/build/torch26-cxx11-cu126-x86_64-linux/activation/rms_norm.py b/build/torch28-cxx11-cu129-x86_64-linux/activation/rms_norm.py similarity index 100% rename from build/torch26-cxx11-cu126-x86_64-linux/activation/rms_norm.py rename to build/torch28-cxx11-cu129-x86_64-linux/activation/rms_norm.py diff --git a/build/torch26-cxx11-rocm62-x86_64-linux/activation/__init__.py b/build/torch28-cxx11-rocm63-x86_64-linux/activation/__init__.py similarity index 100% rename from build/torch26-cxx11-rocm62-x86_64-linux/activation/__init__.py rename to build/torch28-cxx11-rocm63-x86_64-linux/activation/__init__.py diff --git a/build/torch28-cxx11-rocm63-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so b/build/torch28-cxx11-rocm63-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..d7e1620ead5192b60900f00994daaca04dd7b6c0 --- /dev/null +++ b/build/torch28-cxx11-rocm63-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8a75fc3e8648bbab973e3021720ed372ec8468f7a28b5b047640fd7198ab369 +size 2647872 diff --git a/build/torch28-cxx11-rocm63-x86_64-linux/activation/_ops.py b/build/torch28-cxx11-rocm63-x86_64-linux/activation/_ops.py new file mode 100644 index 0000000000000000000000000000000000000000..ecb351fdb0e6990bab29e42a4aecd1896b77a1b2 --- /dev/null +++ b/build/torch28-cxx11-rocm63-x86_64-linux/activation/_ops.py @@ -0,0 +1,9 @@ +import torch +from . import _activation_cf68df1_dirty +ops = torch.ops._activation_cf68df1_dirty + +def add_op_namespace_prefix(op_name: str): + """ + Prefix op by namespace. + """ + return f"_activation_cf68df1_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-rocm62-x86_64-linux/activation/layers.py b/build/torch28-cxx11-rocm63-x86_64-linux/activation/layers.py similarity index 100% rename from build/torch26-cxx11-rocm62-x86_64-linux/activation/layers.py rename to build/torch28-cxx11-rocm63-x86_64-linux/activation/layers.py diff --git a/build/torch26-cxx11-rocm62-x86_64-linux/activation/poly_norm.py b/build/torch28-cxx11-rocm63-x86_64-linux/activation/poly_norm.py similarity index 100% rename from build/torch26-cxx11-rocm62-x86_64-linux/activation/poly_norm.py rename to build/torch28-cxx11-rocm63-x86_64-linux/activation/poly_norm.py diff --git a/build/torch26-cxx11-rocm62-x86_64-linux/activation/rms_norm.py b/build/torch28-cxx11-rocm63-x86_64-linux/activation/rms_norm.py similarity index 100% rename from build/torch26-cxx11-rocm62-x86_64-linux/activation/rms_norm.py rename to build/torch28-cxx11-rocm63-x86_64-linux/activation/rms_norm.py diff --git a/build/torch26-cxx98-cu118-x86_64-linux/activation/__init__.py b/build/torch28-cxx11-rocm64-x86_64-linux/activation/__init__.py similarity index 100% rename from build/torch26-cxx98-cu118-x86_64-linux/activation/__init__.py rename to build/torch28-cxx11-rocm64-x86_64-linux/activation/__init__.py diff --git a/build/torch28-cxx11-rocm64-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so b/build/torch28-cxx11-rocm64-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..1e9b12a0c73081a688a6c2da9b54740dd3aeaa61 --- /dev/null +++ b/build/torch28-cxx11-rocm64-x86_64-linux/activation/_activation_cf68df1_dirty.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cccb0567a8f86f1f9e23a653a2e1f7177f4528cb1ecf8cbec42e40c60392eb39 +size 2633232 diff --git a/build/torch28-cxx11-rocm64-x86_64-linux/activation/_ops.py b/build/torch28-cxx11-rocm64-x86_64-linux/activation/_ops.py new file mode 100644 index 0000000000000000000000000000000000000000..ecb351fdb0e6990bab29e42a4aecd1896b77a1b2 --- /dev/null +++ b/build/torch28-cxx11-rocm64-x86_64-linux/activation/_ops.py @@ -0,0 +1,9 @@ +import torch +from . import _activation_cf68df1_dirty +ops = torch.ops._activation_cf68df1_dirty + +def add_op_namespace_prefix(op_name: str): + """ + Prefix op by namespace. + """ + return f"_activation_cf68df1_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx98-cu118-x86_64-linux/activation/layers.py b/build/torch28-cxx11-rocm64-x86_64-linux/activation/layers.py similarity index 100% rename from build/torch26-cxx98-cu118-x86_64-linux/activation/layers.py rename to build/torch28-cxx11-rocm64-x86_64-linux/activation/layers.py diff --git a/build/torch26-cxx98-cu118-x86_64-linux/activation/poly_norm.py b/build/torch28-cxx11-rocm64-x86_64-linux/activation/poly_norm.py similarity index 100% rename from build/torch26-cxx98-cu118-x86_64-linux/activation/poly_norm.py rename to build/torch28-cxx11-rocm64-x86_64-linux/activation/poly_norm.py diff --git a/build/torch26-cxx98-cu118-x86_64-linux/activation/rms_norm.py b/build/torch28-cxx11-rocm64-x86_64-linux/activation/rms_norm.py similarity index 100% rename from build/torch26-cxx98-cu118-x86_64-linux/activation/rms_norm.py rename to build/torch28-cxx11-rocm64-x86_64-linux/activation/rms_norm.py