Kernels
danieldk HF Staff commited on
Commit
ccdc939
·
verified ·
1 Parent(s): 5d54a3e

Build uploaded using `kernels`.

Browse files
.gitattributes CHANGED
@@ -40,3 +40,4 @@ build/torch29-xpu20252-x86_64-windows/rotary/_rotary_66b961a.pyd filter=lfs diff
40
  build/torch210-cu128-x86_64-windows/rotary/_rotary_9f63cc2.pyd filter=lfs diff=lfs merge=lfs -text
41
  build/torch210-xpu20253-x86_64-windows/rotary/_rotary_9f63cc2.pyd filter=lfs diff=lfs merge=lfs -text
42
  build/torch210-cu128-x86_64-windows/_rotary_cuda_07a01e5.pyd filter=lfs diff=lfs merge=lfs -text
 
 
40
  build/torch210-cu128-x86_64-windows/rotary/_rotary_9f63cc2.pyd filter=lfs diff=lfs merge=lfs -text
41
  build/torch210-xpu20253-x86_64-windows/rotary/_rotary_9f63cc2.pyd filter=lfs diff=lfs merge=lfs -text
42
  build/torch210-cu128-x86_64-windows/_rotary_cuda_07a01e5.pyd filter=lfs diff=lfs merge=lfs -text
43
+ build/torch210-xpu20253-x86_64-windows/_rotary_xpu_07a01e5.pyd filter=lfs diff=lfs merge=lfs -text
build/torch210-xpu20253-x86_64-windows/__init__.py ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import Optional, Tuple
2
+ import torch
3
+
4
+ from ._ops import ops
5
+
6
+
7
+ def apply_rotary(
8
+ x1: torch.Tensor,
9
+ x2: torch.Tensor,
10
+ cos: torch.Tensor,
11
+ sin: torch.Tensor,
12
+ out1: torch.Tensor,
13
+ out2: torch.Tensor,
14
+ conj: bool,
15
+ ) -> None:
16
+ ops.apply_rotary(x1, x2, cos, sin, out1, out2, conj)
17
+
18
+
19
+ def apply_rotary_transformers(
20
+ q: torch.Tensor,
21
+ k: torch.Tensor,
22
+ cos: torch.Tensor,
23
+ sin: torch.Tensor,
24
+ unsqueeze_dim: int = 1,
25
+ ) -> Tuple[torch.Tensor, torch.Tensor]:
26
+ """
27
+ Rotary kernel implementation wrapper
28
+ Adapts rotary kernel implementation to match transformers apply_rotary_pos_emb signature
29
+ """
30
+ cos = cos.unsqueeze(unsqueeze_dim)
31
+ sin = sin.unsqueeze(unsqueeze_dim)
32
+
33
+ q_rotated = q.clone()
34
+ k_rotated = k.clone()
35
+
36
+ # Get half dimension for rotation
37
+ half_dim = q.shape[-1] // 2
38
+ q1 = q_rotated[..., :half_dim]
39
+ q2 = q_rotated[..., half_dim:]
40
+ k1 = k_rotated[..., :half_dim]
41
+ k2 = k_rotated[..., half_dim:]
42
+ if cos.shape[-1] != half_dim:
43
+ # Trim cos/sin to match half_dim
44
+ cos = cos[..., :half_dim]
45
+ sin = sin[..., :half_dim]
46
+
47
+ apply_rotary(q1, q2, cos, sin, q1, q2, False)
48
+ apply_rotary(k1, k2, cos, sin, k1, k2, False)
49
+ return q_rotated, k_rotated
50
+
51
+
52
+ __all__ = ["apply_rotary", "apply_rotary_transformers"]
build/torch210-xpu20253-x86_64-windows/{rotary/_ops.py → _ops.py} RENAMED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _rotary_9f63cc2
3
- ops = torch.ops._rotary_9f63cc2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_rotary_9f63cc2::{op_name}"
 
1
  import torch
2
+ from . import _rotary_xpu_07a01e5
3
+ ops = torch.ops._rotary_xpu_07a01e5
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_rotary_xpu_07a01e5::{op_name}"
build/torch210-xpu20253-x86_64-windows/{rotary/_rotary_9f63cc2.pyd → _rotary_xpu_07a01e5.pyd} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8ece9a8c015b26093abe355961e81d14f6a26bf809cce3f465f1125d97039ea
3
- size 202240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34579b75b36dea1b18446ac679fde1e89d16b25350e2cb8213acfc19434bba8b
3
+ size 396288
build/torch210-xpu20253-x86_64-windows/metadata.json CHANGED
@@ -1,4 +1,5 @@
1
  {
2
  "version": 1,
 
3
  "python-depends": []
4
  }
 
1
  {
2
  "version": 1,
3
+ "license": "BSD-3-Clause",
4
  "python-depends": []
5
  }
build/torch210-xpu20253-x86_64-windows/rotary/__init__.py CHANGED
@@ -1,52 +1,26 @@
1
- from typing import Optional, Tuple
2
- import torch
3
-
4
- from ._ops import ops
5
-
6
-
7
- def apply_rotary(
8
- x1: torch.Tensor,
9
- x2: torch.Tensor,
10
- cos: torch.Tensor,
11
- sin: torch.Tensor,
12
- out1: torch.Tensor,
13
- out2: torch.Tensor,
14
- conj: bool,
15
- ) -> None:
16
- ops.apply_rotary(x1, x2, cos, sin, out1, out2, conj)
17
-
18
-
19
- def apply_rotary_transformers(
20
- q: torch.Tensor,
21
- k: torch.Tensor,
22
- cos: torch.Tensor,
23
- sin: torch.Tensor,
24
- unsqueeze_dim: int = 1,
25
- ) -> Tuple[torch.Tensor, torch.Tensor]:
26
- """
27
- Rotary kernel implementation wrapper
28
- Adapts rotary kernel implementation to match transformers apply_rotary_pos_emb signature
29
- """
30
- cos = cos.unsqueeze(unsqueeze_dim)
31
- sin = sin.unsqueeze(unsqueeze_dim)
32
-
33
- q_rotated = q.clone()
34
- k_rotated = k.clone()
35
-
36
- # Get half dimension for rotation
37
- half_dim = q.shape[-1] // 2
38
- q1 = q_rotated[..., :half_dim]
39
- q2 = q_rotated[..., half_dim:]
40
- k1 = k_rotated[..., :half_dim]
41
- k2 = k_rotated[..., half_dim:]
42
- if cos.shape[-1] != half_dim:
43
- # Trim cos/sin to match half_dim
44
- cos = cos[..., :half_dim]
45
- sin = sin[..., :half_dim]
46
-
47
- apply_rotary(q1, q2, cos, sin, q1, q2, False)
48
- apply_rotary(k1, k2, cos, sin, k1, k2, False)
49
- return q_rotated, k_rotated
50
-
51
-
52
- __all__ = ["apply_rotary", "apply_rotary_transformers"]
 
1
+ import ctypes
2
+ import sys
3
+
4
+ import importlib
5
+ from pathlib import Path
6
+ from types import ModuleType
7
+
8
+ def _import_from_path(file_path: Path) -> ModuleType:
9
+ # We cannot use the module name as-is, after adding it to `sys.modules`,
10
+ # it would also be used for other imports. So, we make a module name that
11
+ # depends on the path for it to be unique using the hex-encoded hash of
12
+ # the path.
13
+ path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value)
14
+ module_name = path_hash
15
+ spec = importlib.util.spec_from_file_location(module_name, file_path)
16
+ if spec is None:
17
+ raise ImportError(f"Cannot load spec for {module_name} from {file_path}")
18
+ module = importlib.util.module_from_spec(spec)
19
+ if module is None:
20
+ raise ImportError(f"Cannot load module {module_name} from spec")
21
+ sys.modules[module_name] = module
22
+ spec.loader.exec_module(module) # type: ignore
23
+ return module
24
+
25
+
26
+ globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py")))