diff --git a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc b/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 2025ba81deff31718ff2e720b996338f0d2540fc..0000000000000000000000000000000000000000 Binary files a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc b/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc deleted file mode 100644 index 1f404b1f39ddeb156e04f8b746085c59fa30d3df..0000000000000000000000000000000000000000 Binary files a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc and /dev/null differ diff --git a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc b/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc deleted file mode 100644 index e252c479a0578dbcf1c7121a941302ae0e4fbefc..0000000000000000000000000000000000000000 Binary files a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc and /dev/null differ diff --git a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/_ops.py b/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/_ops.py deleted file mode 100644 index 56d920dbf5b60b10a3444a94a1035d9e72a0df99..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _paged_attention_6677800 -ops = torch.ops._paged_attention_6677800 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_paged_attention_6677800::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so b/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so deleted file mode 100755 index 13b149c786f1319fb27ecb766c03cd47cdd053ce..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4b860f51a86eb13c6c7a6e74f28c59abf0b73245f241744766029fbda53cba87 -size 88169704 diff --git a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc b/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 87bceab15fa2985197513b8700a85860430a13c5..0000000000000000000000000000000000000000 Binary files a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc b/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc deleted file mode 100644 index 1388862d6688734a1139ca43aa03c37cd50188d4..0000000000000000000000000000000000000000 Binary files a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc and /dev/null differ diff --git a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc b/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc deleted file mode 100644 index 69246dbda4e37ee595a8235c57315504c6836eb4..0000000000000000000000000000000000000000 Binary files a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc and /dev/null differ diff --git a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/_ops.py b/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/_ops.py deleted file mode 100644 index 56d920dbf5b60b10a3444a94a1035d9e72a0df99..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _paged_attention_6677800 -ops = torch.ops._paged_attention_6677800 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_paged_attention_6677800::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so b/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so deleted file mode 100755 index c73a38e4bc93a3677a85cbc5a6ec90ef7c06a672..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:385c8fd6816793573ac802c40e1be99cc5703eebb4df05330ce57b2216da8836 -size 88162432 diff --git a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc b/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 908d5c1b3682b7b845e4f28b3ad5cf5679eb3d77..0000000000000000000000000000000000000000 Binary files a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc b/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc deleted file mode 100644 index 2203a53195e0b8962aee36558c1884632ec585c3..0000000000000000000000000000000000000000 Binary files a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc and /dev/null differ diff --git a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc b/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc deleted file mode 100644 index 4ec467145c405dbb040a9a708962635fe78ed179..0000000000000000000000000000000000000000 Binary files a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc and /dev/null differ diff --git a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/_ops.py b/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/_ops.py deleted file mode 100644 index 56d920dbf5b60b10a3444a94a1035d9e72a0df99..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _paged_attention_6677800 -ops = torch.ops._paged_attention_6677800 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_paged_attention_6677800::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so b/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so deleted file mode 100755 index dd409ba6bbe8de6c11328545f31e2daa59ddde60..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2639286d960d9992c40b85c911773825374d8bb218a82eb84cdd4ac41342cb4c -size 88169904 diff --git a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc similarity index 74% rename from build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc rename to build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc index e8d9819be0b7ea63b7b542b6c63e082a1212fb24..4426d89b40d2875c6242745d3a3404dc105dc340 100644 Binary files a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc and b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc differ diff --git a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc deleted file mode 100644 index 24b4d42473c53006da31733d5b941572ab0f3b5c..0000000000000000000000000000000000000000 Binary files a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-312.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc similarity index 95% rename from build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc rename to build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc index 6d63d9402c07b41e05af4e9aff4356760a31e341..76058206a88f8f66740efd7b611f125a5f3e1eee 100644 Binary files a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc and b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc differ diff --git a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc deleted file mode 100644 index 410de667d4102b48056d27c4cabe44385c6ddbde..0000000000000000000000000000000000000000 Binary files a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-312.pyc and /dev/null differ diff --git a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc new file mode 100644 index 0000000000000000000000000000000000000000..0056674c98c61fb9850920022abc1bbebebdb2e2 Binary files /dev/null and b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc differ diff --git a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/_ops.py b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/_ops.py index 56d920dbf5b60b10a3444a94a1035d9e72a0df99..a5e9f56ff02bdcf047fd97ef2a46b10d6fb5e5eb 100644 --- a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/_ops.py +++ b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _paged_attention_6677800 -ops = torch.ops._paged_attention_6677800 +from . import _paged_attention_b4c51e9 +ops = torch.ops._paged_attention_b4c51e9 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_paged_attention_6677800::{op_name}" \ No newline at end of file + return f"_paged_attention_b4c51e9::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so deleted file mode 100755 index 8c38fbecd70d32dddab373dc0fd433f88b09c8d0..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_6677800.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:7bb98e687cc11950f2103990b777e6699c4a7285a53092def0063a9684fa951e -size 120216032 diff --git a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..31455c6c5a159469575f362d673d272a341e4b19 --- /dev/null +++ b/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9cba9e7b3cf9de722c5dc8f56a533542521fdef05a290a2b4db13948b0f1eca4 +size 138172880 diff --git a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__init__.py b/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__init__.py deleted file mode 100644 index 9de56043369487facc1f163df6bd319c9806e5ca..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__init__.py +++ /dev/null @@ -1,21 +0,0 @@ -from ._custom_ops import ( - convert_fp8, - copy_blocks, - paged_attention_v1, - paged_attention_v2, - reshape_and_cache, - reshape_and_cache_flash, - swap_blocks, -) -from ._ops import ops - -__all__ = [ - "convert_fp8", - "copy_blocks", - "ops", - "paged_attention_v1", - "paged_attention_v2", - "reshape_and_cache", - "reshape_and_cache_flash", - "swap_blocks", -] diff --git a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc b/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc deleted file mode 100644 index 80ebb23881a636449fcb50b7310de1fd35892af3..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/_custom_ops.py b/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/_custom_ops.py deleted file mode 100644 index a0c0b8db085468dee5100c98d14106a9ee917bf2..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/_custom_ops.py +++ /dev/null @@ -1,173 +0,0 @@ -from typing import List, Optional - -import torch - -from ._ops import ops - - -# page attention ops -def paged_attention_v1( - out: torch.Tensor, - query: torch.Tensor, - key_cache: torch.Tensor, - value_cache: torch.Tensor, - num_kv_heads: int, - scale: float, - block_tables: torch.Tensor, - seq_lens: torch.Tensor, - block_size: int, - max_seq_len: int, - alibi_slopes: Optional[torch.Tensor], - kv_cache_dtype: str, - k_scale: float, - v_scale: float, - tp_rank: int = 0, - blocksparse_local_blocks: int = 0, - blocksparse_vert_stride: int = 0, - blocksparse_block_size: int = 64, - blocksparse_head_sliding_step: int = 0, -) -> None: - ops.paged_attention_v1( - out, - query, - key_cache, - value_cache, - num_kv_heads, - scale, - block_tables, - seq_lens, - block_size, - max_seq_len, - alibi_slopes, - kv_cache_dtype, - k_scale, - v_scale, - tp_rank, - blocksparse_local_blocks, - blocksparse_vert_stride, - blocksparse_block_size, - blocksparse_head_sliding_step, - ) - - -def paged_attention_v2( - out: torch.Tensor, - exp_sum: torch.Tensor, - max_logits: torch.Tensor, - tmp_out: torch.Tensor, - query: torch.Tensor, - key_cache: torch.Tensor, - value_cache: torch.Tensor, - num_kv_heads: int, - scale: float, - block_tables: torch.Tensor, - seq_lens: torch.Tensor, - block_size: int, - max_seq_len: int, - alibi_slopes: Optional[torch.Tensor], - kv_cache_dtype: str, - k_scale: float, - v_scale: float, - tp_rank: int = 0, - blocksparse_local_blocks: int = 0, - blocksparse_vert_stride: int = 0, - blocksparse_block_size: int = 64, - blocksparse_head_sliding_step: int = 0, -) -> None: - ops.paged_attention_v2( - out, - exp_sum, - max_logits, - tmp_out, - query, - key_cache, - value_cache, - num_kv_heads, - scale, - block_tables, - seq_lens, - block_size, - max_seq_len, - alibi_slopes, - kv_cache_dtype, - k_scale, - v_scale, - tp_rank, - blocksparse_local_blocks, - blocksparse_vert_stride, - blocksparse_block_size, - blocksparse_head_sliding_step, - ) - - -def reshape_and_cache( - key: torch.Tensor, - value: torch.Tensor, - key_cache: torch.Tensor, - value_cache: torch.Tensor, - slot_mapping: torch.Tensor, - kv_cache_dtype: str, - k_scale: float, - v_scale: float, -) -> None: - ops.reshape_and_cache( - key, - value, - key_cache, - value_cache, - slot_mapping, - kv_cache_dtype, - k_scale, - v_scale, - ) - - -def reshape_and_cache_flash( - key: torch.Tensor, - value: torch.Tensor, - key_cache: torch.Tensor, - value_cache: torch.Tensor, - slot_mapping: torch.Tensor, - kv_cache_dtype: str, - k_scale: torch.Tensor, - v_scale: torch.Tensor, -) -> None: - ops.reshape_and_cache_flash( - key, - value, - key_cache, - value_cache, - slot_mapping, - kv_cache_dtype, - k_scale, - v_scale, - ) - - -def copy_blocks( - key_caches: List[torch.Tensor], - value_caches: List[torch.Tensor], - block_mapping: torch.Tensor, -) -> None: - ops.copy_blocks(key_caches, value_caches, block_mapping) - - -def swap_blocks( - src: torch.Tensor, dst: torch.Tensor, block_mapping: torch.Tensor -) -> None: - ops.swap_blocks(src, dst, block_mapping) - - -def convert_fp8( - output: torch.Tensor, input: torch.Tensor, scale: float = 1.0, kv_dtype: str = "fp8" -) -> None: - ops.convert_fp8(output, input, scale, kv_dtype) - - -__all__ = [ - "convert_fp8", - "paged_attention_v1", - "paged_attention_v2", - "reshape_and_cache", - "copy_blocks", -] diff --git a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/_ops.py b/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/_ops.py deleted file mode 100644 index 47f517a56cfeff26ae58451354ac8f77f73da8ad..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _paged_attention_e6ce28c_dirty -ops = torch.ops._paged_attention_e6ce28c_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_paged_attention_e6ce28c_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_e6ce28c_dirty.abi3.so b/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_e6ce28c_dirty.abi3.so deleted file mode 100755 index 80faac87843ff109fac27648f6b4d79be0e763e9..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_e6ce28c_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:97f6ba61bd73001a587126e697ca500dd6e89bb5b2e46a58ea42e83d023f0d0e -size 88171000 diff --git a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/platforms.py b/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/platforms.py deleted file mode 100644 index 6277d5f50ff3ddc265bb39fa1c4d17e0341b7767..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/platforms.py +++ /dev/null @@ -1,92 +0,0 @@ -import os -import random -from abc import ABC, abstractmethod -from functools import lru_cache, wraps -from typing import Callable, ParamSpec, TypeVar - -import numpy as np -import torch - -IS_ROCM = torch.version.hip is not None -IS_MPS = torch.backends.mps.is_available() - - -class Platform(ABC): - @classmethod - def seed_everything(cls, seed: int) -> None: - """ - Set the seed of each random module. - `torch.manual_seed` will set seed on all devices. - - Loosely based on: https://github.com/Lightning-AI/pytorch-lightning/blob/2.4.0/src/lightning/fabric/utilities/seed.py#L20 - """ - random.seed(seed) - np.random.seed(seed) - torch.manual_seed(seed) - - @abstractmethod - def get_device_name(self, device_id: int = 0) -> str: ... - - @abstractmethod - def is_cuda(self) -> bool: ... - - @abstractmethod - def is_rocm(self) -> bool: ... - - @abstractmethod - def is_mps(self) -> bool: ... - - -class CudaPlatform(Platform): - @classmethod - @lru_cache(maxsize=8) - def get_device_name(cls, device_id: int = 0) -> str: - return torch.cuda.get_device_name(0) - - def is_cuda(self) -> bool: - return True - - def is_rocm(self) -> bool: - return False - - def is_mps(self) -> bool: - return False - - -class RocmPlatform(Platform): - @classmethod - @lru_cache(maxsize=8) - def get_device_name(cls, device_id: int = 0) -> str: - return torch.cuda.get_device_name(device_id) - - def is_cuda(self) -> bool: - return False - - def is_rocm(self) -> bool: - return True - - def is_mps(self) -> bool: - return False - - -class MpsPlatform(Platform): - @classmethod - @lru_cache(maxsize=8) - def get_device_name(cls, device_id: int = 0) -> str: - return torch.cuda.get_device_name(device_id) - - def is_cuda(self) -> bool: - return False - - def is_rocm(self) -> bool: - return False - - def is_mps(self) -> bool: - return True - -current_platform = ( - RocmPlatform() if IS_ROCM else - MpsPlatform() if IS_MPS else - CudaPlatform() if torch.cuda.is_available() else - None -) diff --git a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__init__.py b/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__init__.py deleted file mode 100644 index 9de56043369487facc1f163df6bd319c9806e5ca..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__init__.py +++ /dev/null @@ -1,21 +0,0 @@ -from ._custom_ops import ( - convert_fp8, - copy_blocks, - paged_attention_v1, - paged_attention_v2, - reshape_and_cache, - reshape_and_cache_flash, - swap_blocks, -) -from ._ops import ops - -__all__ = [ - "convert_fp8", - "copy_blocks", - "ops", - "paged_attention_v1", - "paged_attention_v2", - "reshape_and_cache", - "reshape_and_cache_flash", - "swap_blocks", -] diff --git a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc b/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc deleted file mode 100644 index 0358df0ee25460b76445ed98a2d8959e41bc6238..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/_custom_ops.py b/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/_custom_ops.py deleted file mode 100644 index a0c0b8db085468dee5100c98d14106a9ee917bf2..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/_custom_ops.py +++ /dev/null @@ -1,173 +0,0 @@ -from typing import List, Optional - -import torch - -from ._ops import ops - - -# page attention ops -def paged_attention_v1( - out: torch.Tensor, - query: torch.Tensor, - key_cache: torch.Tensor, - value_cache: torch.Tensor, - num_kv_heads: int, - scale: float, - block_tables: torch.Tensor, - seq_lens: torch.Tensor, - block_size: int, - max_seq_len: int, - alibi_slopes: Optional[torch.Tensor], - kv_cache_dtype: str, - k_scale: float, - v_scale: float, - tp_rank: int = 0, - blocksparse_local_blocks: int = 0, - blocksparse_vert_stride: int = 0, - blocksparse_block_size: int = 64, - blocksparse_head_sliding_step: int = 0, -) -> None: - ops.paged_attention_v1( - out, - query, - key_cache, - value_cache, - num_kv_heads, - scale, - block_tables, - seq_lens, - block_size, - max_seq_len, - alibi_slopes, - kv_cache_dtype, - k_scale, - v_scale, - tp_rank, - blocksparse_local_blocks, - blocksparse_vert_stride, - blocksparse_block_size, - blocksparse_head_sliding_step, - ) - - -def paged_attention_v2( - out: torch.Tensor, - exp_sum: torch.Tensor, - max_logits: torch.Tensor, - tmp_out: torch.Tensor, - query: torch.Tensor, - key_cache: torch.Tensor, - value_cache: torch.Tensor, - num_kv_heads: int, - scale: float, - block_tables: torch.Tensor, - seq_lens: torch.Tensor, - block_size: int, - max_seq_len: int, - alibi_slopes: Optional[torch.Tensor], - kv_cache_dtype: str, - k_scale: float, - v_scale: float, - tp_rank: int = 0, - blocksparse_local_blocks: int = 0, - blocksparse_vert_stride: int = 0, - blocksparse_block_size: int = 64, - blocksparse_head_sliding_step: int = 0, -) -> None: - ops.paged_attention_v2( - out, - exp_sum, - max_logits, - tmp_out, - query, - key_cache, - value_cache, - num_kv_heads, - scale, - block_tables, - seq_lens, - block_size, - max_seq_len, - alibi_slopes, - kv_cache_dtype, - k_scale, - v_scale, - tp_rank, - blocksparse_local_blocks, - blocksparse_vert_stride, - blocksparse_block_size, - blocksparse_head_sliding_step, - ) - - -def reshape_and_cache( - key: torch.Tensor, - value: torch.Tensor, - key_cache: torch.Tensor, - value_cache: torch.Tensor, - slot_mapping: torch.Tensor, - kv_cache_dtype: str, - k_scale: float, - v_scale: float, -) -> None: - ops.reshape_and_cache( - key, - value, - key_cache, - value_cache, - slot_mapping, - kv_cache_dtype, - k_scale, - v_scale, - ) - - -def reshape_and_cache_flash( - key: torch.Tensor, - value: torch.Tensor, - key_cache: torch.Tensor, - value_cache: torch.Tensor, - slot_mapping: torch.Tensor, - kv_cache_dtype: str, - k_scale: torch.Tensor, - v_scale: torch.Tensor, -) -> None: - ops.reshape_and_cache_flash( - key, - value, - key_cache, - value_cache, - slot_mapping, - kv_cache_dtype, - k_scale, - v_scale, - ) - - -def copy_blocks( - key_caches: List[torch.Tensor], - value_caches: List[torch.Tensor], - block_mapping: torch.Tensor, -) -> None: - ops.copy_blocks(key_caches, value_caches, block_mapping) - - -def swap_blocks( - src: torch.Tensor, dst: torch.Tensor, block_mapping: torch.Tensor -) -> None: - ops.swap_blocks(src, dst, block_mapping) - - -def convert_fp8( - output: torch.Tensor, input: torch.Tensor, scale: float = 1.0, kv_dtype: str = "fp8" -) -> None: - ops.convert_fp8(output, input, scale, kv_dtype) - - -__all__ = [ - "convert_fp8", - "paged_attention_v1", - "paged_attention_v2", - "reshape_and_cache", - "copy_blocks", -] diff --git a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/_ops.py b/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/_ops.py deleted file mode 100644 index 47f517a56cfeff26ae58451354ac8f77f73da8ad..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _paged_attention_e6ce28c_dirty -ops = torch.ops._paged_attention_e6ce28c_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_paged_attention_e6ce28c_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_e6ce28c_dirty.abi3.so b/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_e6ce28c_dirty.abi3.so deleted file mode 100755 index ad45b83eafc75fdfb7ccb3f999af90e7993fcb52..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_e6ce28c_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:80929eb7a9f2561ef3e21c764494ac30f80935b054150b415626e49fe8d52b96 -size 120282936 diff --git a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/platforms.py b/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/platforms.py deleted file mode 100644 index 6277d5f50ff3ddc265bb39fa1c4d17e0341b7767..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/platforms.py +++ /dev/null @@ -1,92 +0,0 @@ -import os -import random -from abc import ABC, abstractmethod -from functools import lru_cache, wraps -from typing import Callable, ParamSpec, TypeVar - -import numpy as np -import torch - -IS_ROCM = torch.version.hip is not None -IS_MPS = torch.backends.mps.is_available() - - -class Platform(ABC): - @classmethod - def seed_everything(cls, seed: int) -> None: - """ - Set the seed of each random module. - `torch.manual_seed` will set seed on all devices. - - Loosely based on: https://github.com/Lightning-AI/pytorch-lightning/blob/2.4.0/src/lightning/fabric/utilities/seed.py#L20 - """ - random.seed(seed) - np.random.seed(seed) - torch.manual_seed(seed) - - @abstractmethod - def get_device_name(self, device_id: int = 0) -> str: ... - - @abstractmethod - def is_cuda(self) -> bool: ... - - @abstractmethod - def is_rocm(self) -> bool: ... - - @abstractmethod - def is_mps(self) -> bool: ... - - -class CudaPlatform(Platform): - @classmethod - @lru_cache(maxsize=8) - def get_device_name(cls, device_id: int = 0) -> str: - return torch.cuda.get_device_name(0) - - def is_cuda(self) -> bool: - return True - - def is_rocm(self) -> bool: - return False - - def is_mps(self) -> bool: - return False - - -class RocmPlatform(Platform): - @classmethod - @lru_cache(maxsize=8) - def get_device_name(cls, device_id: int = 0) -> str: - return torch.cuda.get_device_name(device_id) - - def is_cuda(self) -> bool: - return False - - def is_rocm(self) -> bool: - return True - - def is_mps(self) -> bool: - return False - - -class MpsPlatform(Platform): - @classmethod - @lru_cache(maxsize=8) - def get_device_name(cls, device_id: int = 0) -> str: - return torch.cuda.get_device_name(device_id) - - def is_cuda(self) -> bool: - return False - - def is_rocm(self) -> bool: - return False - - def is_mps(self) -> bool: - return True - -current_platform = ( - RocmPlatform() if IS_ROCM else - MpsPlatform() if IS_MPS else - CudaPlatform() if torch.cuda.is_available() else - None -) diff --git a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc index 5ec2a4901c0d53ef222b2b186cfaa6600d3034aa..f2029d1fa1af1cacf90cbe0e8a2a59b907940e8d 100644 Binary files a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc and b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc differ diff --git a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc index 60dbdc7997f5c407a08bc9f11e0680b7958709a0..4f54c4683b6c5a5d05d435d40f546ac8cbf62e76 100644 Binary files a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc and b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc differ diff --git a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc index 860e06f7e8df8f36f3efe6d07b13af0078394e29..095bc0f18dddbaa40b04468fa6cb9dde32b9285a 100644 Binary files a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc and b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc differ diff --git a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/_ops.py b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/_ops.py index 47f517a56cfeff26ae58451354ac8f77f73da8ad..a5e9f56ff02bdcf047fd97ef2a46b10d6fb5e5eb 100644 --- a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/_ops.py +++ b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _paged_attention_e6ce28c_dirty -ops = torch.ops._paged_attention_e6ce28c_dirty +from . import _paged_attention_b4c51e9 +ops = torch.ops._paged_attention_b4c51e9 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_paged_attention_e6ce28c_dirty::{op_name}" \ No newline at end of file + return f"_paged_attention_b4c51e9::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..f4f21766fa2c547675a1373926d281b4ec7a28fa --- /dev/null +++ b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02238a0a4dacdbbf60eb9bb73a95332b448d5127f71149003c764f40595d9d06 +size 149841048 diff --git a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/_paged_attention_e6ce28c_dirty.abi3.so b/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/_paged_attention_e6ce28c_dirty.abi3.so deleted file mode 100755 index dc661ad01016554257bc6dfa91854131f6c0d12d..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-aarch64-linux/paged_attention/_paged_attention_e6ce28c_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:874c4b927fbbc21dc27ec190beac22b38346b66a080a0e56e4e9a010c5d737b5 -size 130442416 diff --git a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__init__.py b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__init__.py similarity index 100% rename from build/torch26-cxx11-cu126-aarch64-linux/paged_attention/__init__.py rename to build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__init__.py diff --git a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc similarity index 74% rename from build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc rename to build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc index a20cde01b8857717f6eec5bce7029cc120a4ba79..4c35d9c01a28b3d52578f20c0116c30fd3aee5e7 100644 Binary files a/build/torch28-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc and b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc differ diff --git a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc similarity index 95% rename from build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc rename to build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc index e1497ea30ccdc24d2aa6ff587e078666676918dd..a7ee2b17c2afd046a6634541c2a78148adc920b4 100644 Binary files a/build/torch28-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc and b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc differ diff --git a/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc new file mode 100644 index 0000000000000000000000000000000000000000..b05df5ac67a38cdaee7a1c65ae83b0df7476efe8 Binary files /dev/null and b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc differ diff --git a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/_custom_ops.py b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/_custom_ops.py similarity index 100% rename from build/torch26-cxx11-cu126-aarch64-linux/paged_attention/_custom_ops.py rename to build/torch29-cxx11-cu126-aarch64-linux/paged_attention/_custom_ops.py diff --git a/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/_ops.py b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/_ops.py new file mode 100644 index 0000000000000000000000000000000000000000..a5e9f56ff02bdcf047fd97ef2a46b10d6fb5e5eb --- /dev/null +++ b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/_ops.py @@ -0,0 +1,9 @@ +import torch +from . import _paged_attention_b4c51e9 +ops = torch.ops._paged_attention_b4c51e9 + +def add_op_namespace_prefix(op_name: str): + """ + Prefix op by namespace. + """ + return f"_paged_attention_b4c51e9::{op_name}" \ No newline at end of file diff --git a/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..e147e4816316407f5430e44670a08020a4d2d1b6 --- /dev/null +++ b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75a193f66139a77108969642a435eb345b97c24e4a31e53856fd7c65bff6e810 +size 110584256 diff --git a/build/torch26-cxx11-cu126-aarch64-linux/paged_attention/platforms.py b/build/torch29-cxx11-cu126-aarch64-linux/paged_attention/platforms.py similarity index 100% rename from build/torch26-cxx11-cu126-aarch64-linux/paged_attention/platforms.py rename to build/torch29-cxx11-cu126-aarch64-linux/paged_attention/platforms.py diff --git a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__init__.py b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__init__.py similarity index 100% rename from build/torch26-cxx98-cu126-aarch64-linux/paged_attention/__init__.py rename to build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__init__.py diff --git a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc similarity index 55% rename from build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc rename to build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc index 8a177a2be59f2ec3ddb8c5ce19e1316c614b8cfd..d9b411f561235eeebf4831cc5b7fd3af6d6b2b97 100644 Binary files a/build/torch27-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-312.pyc and b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc differ diff --git a/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc new file mode 100644 index 0000000000000000000000000000000000000000..8eb24f6e486563cd362705980c39de6d933f9e49 Binary files /dev/null and b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc differ diff --git a/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc new file mode 100644 index 0000000000000000000000000000000000000000..89df1e0361267bd24e3d4341e77b75daf8a187a9 Binary files /dev/null and b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc differ diff --git a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/_custom_ops.py b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/_custom_ops.py similarity index 100% rename from build/torch26-cxx98-cu126-aarch64-linux/paged_attention/_custom_ops.py rename to build/torch29-cxx11-cu128-aarch64-linux/paged_attention/_custom_ops.py diff --git a/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/_ops.py b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/_ops.py new file mode 100644 index 0000000000000000000000000000000000000000..a5e9f56ff02bdcf047fd97ef2a46b10d6fb5e5eb --- /dev/null +++ b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/_ops.py @@ -0,0 +1,9 @@ +import torch +from . import _paged_attention_b4c51e9 +ops = torch.ops._paged_attention_b4c51e9 + +def add_op_namespace_prefix(op_name: str): + """ + Prefix op by namespace. + """ + return f"_paged_attention_b4c51e9::{op_name}" \ No newline at end of file diff --git a/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..d514c5d024e62151a13638101a3c96d23b0ee198 --- /dev/null +++ b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0955cb5ea9b9714579c81efc8810bba7ce46aad33701c244efcb01d021a944f +size 138174216 diff --git a/build/torch26-cxx98-cu126-aarch64-linux/paged_attention/platforms.py b/build/torch29-cxx11-cu128-aarch64-linux/paged_attention/platforms.py similarity index 100% rename from build/torch26-cxx98-cu126-aarch64-linux/paged_attention/platforms.py rename to build/torch29-cxx11-cu128-aarch64-linux/paged_attention/platforms.py diff --git a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__init__.py b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__init__.py similarity index 100% rename from build/torch27-cxx11-cu126-aarch64-linux/paged_attention/__init__.py rename to build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__init__.py diff --git a/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc new file mode 100644 index 0000000000000000000000000000000000000000..f8fb851276be004995178c84b861a35401203ccc Binary files /dev/null and b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__pycache__/__init__.cpython-313.pyc differ diff --git a/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc new file mode 100644 index 0000000000000000000000000000000000000000..3f7ec2073f8594850264af2ff274d089067cc2b0 Binary files /dev/null and b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__pycache__/_custom_ops.cpython-313.pyc differ diff --git a/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc new file mode 100644 index 0000000000000000000000000000000000000000..389abfdad4679d365116f9341819ed8d89fd1e8a Binary files /dev/null and b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/__pycache__/_ops.cpython-313.pyc differ diff --git a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/_custom_ops.py b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/_custom_ops.py similarity index 100% rename from build/torch27-cxx11-cu126-aarch64-linux/paged_attention/_custom_ops.py rename to build/torch29-cxx11-cu130-aarch64-linux/paged_attention/_custom_ops.py diff --git a/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/_ops.py b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/_ops.py new file mode 100644 index 0000000000000000000000000000000000000000..a5e9f56ff02bdcf047fd97ef2a46b10d6fb5e5eb --- /dev/null +++ b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/_ops.py @@ -0,0 +1,9 @@ +import torch +from . import _paged_attention_b4c51e9 +ops = torch.ops._paged_attention_b4c51e9 + +def add_op_namespace_prefix(op_name: str): + """ + Prefix op by namespace. + """ + return f"_paged_attention_b4c51e9::{op_name}" \ No newline at end of file diff --git a/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..9fe5180086bb402126f2bd617343f8814eecfe2c --- /dev/null +++ b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/_paged_attention_b4c51e9.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6515c6d2e299021620445210f11a918d6b8cd93c8f8912b6904b40a86d646561 +size 76628152 diff --git a/build/torch27-cxx11-cu126-aarch64-linux/paged_attention/platforms.py b/build/torch29-cxx11-cu130-aarch64-linux/paged_attention/platforms.py similarity index 100% rename from build/torch27-cxx11-cu126-aarch64-linux/paged_attention/platforms.py rename to build/torch29-cxx11-cu130-aarch64-linux/paged_attention/platforms.py