danieldk HF staff commited on
Commit
331b7e6
Β·
1 Parent(s): 3dcba92
This view is limited to 50 files because it contains too many changes. Β  See raw diff
Files changed (50) hide show
  1. build/torch25-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/__init__.py +0 -0
  2. build/torch25-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py +0 -0
  3. build/{torch25-cxx11-cu124-x86_64-linux/attention β†’ torch25-cxx11-cu118-x86_64-linux/paged_attention}/_ops.py +3 -3
  4. build/torch25-cxx11-cu118-x86_64-linux/{attention/_attention_6yvgebnqctora.abi3.so β†’ paged_attention/_paged_attention_fao6f4gjjrpl6.abi3.so} +2 -2
  5. build/torch25-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/platforms.py +0 -0
  6. build/torch25-cxx11-cu121-x86_64-linux/{attention β†’ paged_attention}/__init__.py +0 -0
  7. build/torch25-cxx11-cu121-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py +0 -0
  8. build/{torch25-cxx11-cu118-x86_64-linux/attention β†’ torch25-cxx11-cu121-x86_64-linux/paged_attention}/_ops.py +3 -3
  9. build/{torch25-cxx11-cu124-x86_64-linux/attention/_attention_syg6kbhkhc4xk.abi3.so β†’ torch25-cxx11-cu121-x86_64-linux/paged_attention/_paged_attention_eo7ts45r6k64y.abi3.so} +2 -2
  10. build/torch25-cxx11-cu121-x86_64-linux/{attention β†’ paged_attention}/platforms.py +0 -0
  11. build/torch25-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/__init__.py +0 -0
  12. build/torch25-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py +0 -0
  13. build/{torch25-cxx98-cu118-x86_64-linux/attention β†’ torch25-cxx11-cu124-x86_64-linux/paged_attention}/_ops.py +3 -3
  14. build/{torch25-cxx11-cu121-x86_64-linux/attention/_attention_4jg2igd54wzge.abi3.so β†’ torch25-cxx11-cu124-x86_64-linux/paged_attention/_paged_attention_5odgyxqhwqtv2.abi3.so} +2 -2
  15. build/torch25-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/platforms.py +0 -0
  16. build/torch25-cxx98-cu118-x86_64-linux/{attention β†’ paged_attention}/__init__.py +0 -0
  17. build/torch25-cxx98-cu118-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py +0 -0
  18. build/{torch25-cxx11-cu121-x86_64-linux/attention β†’ torch25-cxx98-cu118-x86_64-linux/paged_attention}/_ops.py +3 -3
  19. build/torch25-cxx98-cu118-x86_64-linux/{attention/_attention_hhzgzhvc7zviy.abi3.so β†’ paged_attention/_paged_attention_uy2moinaww2jc.abi3.so} +2 -2
  20. build/torch25-cxx98-cu118-x86_64-linux/{attention β†’ paged_attention}/platforms.py +0 -0
  21. build/torch25-cxx98-cu121-x86_64-linux/attention/_attention_gbi5gm244waic.abi3.so +0 -3
  22. build/torch25-cxx98-cu121-x86_64-linux/attention/_ops.py +0 -9
  23. build/torch25-cxx98-cu121-x86_64-linux/{attention β†’ paged_attention}/__init__.py +0 -0
  24. build/torch25-cxx98-cu121-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py +0 -0
  25. build/torch25-cxx98-cu121-x86_64-linux/paged_attention/_ops.py +9 -0
  26. build/torch25-cxx98-cu121-x86_64-linux/paged_attention/_paged_attention_35dt23tewn2p2.abi3.so +3 -0
  27. build/torch25-cxx98-cu121-x86_64-linux/{attention β†’ paged_attention}/platforms.py +0 -0
  28. build/torch25-cxx98-cu124-x86_64-linux/attention/_attention_ill75rmpj7yds.abi3.so +0 -3
  29. build/torch25-cxx98-cu124-x86_64-linux/attention/_ops.py +0 -9
  30. build/torch25-cxx98-cu124-x86_64-linux/{attention β†’ paged_attention}/__init__.py +0 -0
  31. build/torch25-cxx98-cu124-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py +0 -0
  32. build/torch25-cxx98-cu124-x86_64-linux/paged_attention/_ops.py +9 -0
  33. build/torch25-cxx98-cu124-x86_64-linux/paged_attention/_paged_attention_fhq57q56w3m5o.abi3.so +3 -0
  34. build/torch25-cxx98-cu124-x86_64-linux/{attention β†’ paged_attention}/platforms.py +0 -0
  35. build/torch26-cxx11-cu118-x86_64-linux/attention/_attention_6qe5ft3kiteru.abi3.so +0 -3
  36. build/torch26-cxx11-cu118-x86_64-linux/attention/_ops.py +0 -9
  37. build/torch26-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/__init__.py +0 -0
  38. build/torch26-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py +0 -0
  39. build/torch26-cxx11-cu118-x86_64-linux/paged_attention/_ops.py +9 -0
  40. build/torch26-cxx11-cu118-x86_64-linux/paged_attention/_paged_attention_xvepb4loq5mm2.abi3.so +3 -0
  41. build/torch26-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/platforms.py +0 -0
  42. build/torch26-cxx11-cu124-x86_64-linux/attention/_attention_ftq3cjdxqfw4m.abi3.so +0 -3
  43. build/torch26-cxx11-cu124-x86_64-linux/attention/_ops.py +0 -9
  44. build/torch26-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/__init__.py +0 -0
  45. build/torch26-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py +0 -0
  46. build/torch26-cxx11-cu124-x86_64-linux/paged_attention/_ops.py +9 -0
  47. build/torch26-cxx11-cu124-x86_64-linux/paged_attention/_paged_attention_uyfdujhnc2xoe.abi3.so +3 -0
  48. build/torch26-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/platforms.py +0 -0
  49. build/torch26-cxx11-cu126-x86_64-linux/attention/_attention_lkibbjh726iwm.abi3.so +0 -3
  50. build/torch26-cxx11-cu126-x86_64-linux/attention/_ops.py +0 -9
build/torch25-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/__init__.py RENAMED
File without changes
build/torch25-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py RENAMED
File without changes
build/{torch25-cxx11-cu124-x86_64-linux/attention β†’ torch25-cxx11-cu118-x86_64-linux/paged_attention}/_ops.py RENAMED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _attention_syg6kbhkhc4xk
3
- ops = torch.ops._attention_syg6kbhkhc4xk
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_attention_syg6kbhkhc4xk::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_fao6f4gjjrpl6
3
+ ops = torch.ops._paged_attention_fao6f4gjjrpl6
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_fao6f4gjjrpl6::{op_name}"
build/torch25-cxx11-cu118-x86_64-linux/{attention/_attention_6yvgebnqctora.abi3.so β†’ paged_attention/_paged_attention_fao6f4gjjrpl6.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aee255dc2618e23d4e2076ff3d16c4fbd12d63742fde84252cfb6bfe55c5376e
3
- size 78886392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd3f60d12b061ef8cc4adb9bd2d60539bc233075c007af99b995e4c12bb5df98
3
+ size 78886416
build/torch25-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/platforms.py RENAMED
File without changes
build/torch25-cxx11-cu121-x86_64-linux/{attention β†’ paged_attention}/__init__.py RENAMED
File without changes
build/torch25-cxx11-cu121-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py RENAMED
File without changes
build/{torch25-cxx11-cu118-x86_64-linux/attention β†’ torch25-cxx11-cu121-x86_64-linux/paged_attention}/_ops.py RENAMED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _attention_6yvgebnqctora
3
- ops = torch.ops._attention_6yvgebnqctora
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_attention_6yvgebnqctora::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_eo7ts45r6k64y
3
+ ops = torch.ops._paged_attention_eo7ts45r6k64y
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_eo7ts45r6k64y::{op_name}"
build/{torch25-cxx11-cu124-x86_64-linux/attention/_attention_syg6kbhkhc4xk.abi3.so β†’ torch25-cxx11-cu121-x86_64-linux/paged_attention/_paged_attention_eo7ts45r6k64y.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42a3b2b450b7e284694e8e6d7398627b977d1e5da12bb79d93c6009c192922f9
3
- size 75568320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4f9be8ea04e306483bd77ee84a70be4b1e2bbe1e942766f398cb41afe8cb3f3
3
+ size 75398840
build/torch25-cxx11-cu121-x86_64-linux/{attention β†’ paged_attention}/platforms.py RENAMED
File without changes
build/torch25-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/__init__.py RENAMED
File without changes
build/torch25-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py RENAMED
File without changes
build/{torch25-cxx98-cu118-x86_64-linux/attention β†’ torch25-cxx11-cu124-x86_64-linux/paged_attention}/_ops.py RENAMED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _attention_hhzgzhvc7zviy
3
- ops = torch.ops._attention_hhzgzhvc7zviy
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_attention_hhzgzhvc7zviy::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_5odgyxqhwqtv2
3
+ ops = torch.ops._paged_attention_5odgyxqhwqtv2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_5odgyxqhwqtv2::{op_name}"
build/{torch25-cxx11-cu121-x86_64-linux/attention/_attention_4jg2igd54wzge.abi3.so β†’ torch25-cxx11-cu124-x86_64-linux/paged_attention/_paged_attention_5odgyxqhwqtv2.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22599ebe9d209fcc82068054caf39f93e6828bb3889b344e655fee50e7a98864
3
- size 75398808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fbf3b6cf357130fb9bb6e68cab38cbcbf485147d9dcce1a9a7d7c02fb2e1af8
3
+ size 75568352
build/torch25-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/platforms.py RENAMED
File without changes
build/torch25-cxx98-cu118-x86_64-linux/{attention β†’ paged_attention}/__init__.py RENAMED
File without changes
build/torch25-cxx98-cu118-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py RENAMED
File without changes
build/{torch25-cxx11-cu121-x86_64-linux/attention β†’ torch25-cxx98-cu118-x86_64-linux/paged_attention}/_ops.py RENAMED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _attention_4jg2igd54wzge
3
- ops = torch.ops._attention_4jg2igd54wzge
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_attention_4jg2igd54wzge::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_uy2moinaww2jc
3
+ ops = torch.ops._paged_attention_uy2moinaww2jc
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_uy2moinaww2jc::{op_name}"
build/torch25-cxx98-cu118-x86_64-linux/{attention/_attention_hhzgzhvc7zviy.abi3.so β†’ paged_attention/_paged_attention_uy2moinaww2jc.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ffad04fc3e82be818bafed25c1be1e9e6145f99eb0ef89ab87ef5ab8c8366f9b
3
- size 78850608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:546d3cb8850575fdd3dd198614d7c64894517dfc124ecd0b384f1918940e9db0
3
+ size 78850640
build/torch25-cxx98-cu118-x86_64-linux/{attention β†’ paged_attention}/platforms.py RENAMED
File without changes
build/torch25-cxx98-cu121-x86_64-linux/attention/_attention_gbi5gm244waic.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ed1c9c4c080a10f7d7f8c18e8e96613020851f769a1bf5e2b92bf19b4e01fb6
3
- size 75359216
 
 
 
 
build/torch25-cxx98-cu121-x86_64-linux/attention/_ops.py DELETED
@@ -1,9 +0,0 @@
1
- import torch
2
- from . import _attention_gbi5gm244waic
3
- ops = torch.ops._attention_gbi5gm244waic
4
-
5
- def add_op_namespace_prefix(op_name: str):
6
- """
7
- Prefix op by namespace.
8
- """
9
- return f"_attention_gbi5gm244waic::{op_name}"
 
 
 
 
 
 
 
 
 
 
build/torch25-cxx98-cu121-x86_64-linux/{attention β†’ paged_attention}/__init__.py RENAMED
File without changes
build/torch25-cxx98-cu121-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py RENAMED
File without changes
build/torch25-cxx98-cu121-x86_64-linux/paged_attention/_ops.py ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ from . import _paged_attention_35dt23tewn2p2
3
+ ops = torch.ops._paged_attention_35dt23tewn2p2
4
+
5
+ def add_op_namespace_prefix(op_name: str):
6
+ """
7
+ Prefix op by namespace.
8
+ """
9
+ return f"_paged_attention_35dt23tewn2p2::{op_name}"
build/torch25-cxx98-cu121-x86_64-linux/paged_attention/_paged_attention_35dt23tewn2p2.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45aacf490514529bb7c46955239c5e2f8c07b827184ebdb1bc78ffff6c304b51
3
+ size 75359240
build/torch25-cxx98-cu121-x86_64-linux/{attention β†’ paged_attention}/platforms.py RENAMED
File without changes
build/torch25-cxx98-cu124-x86_64-linux/attention/_attention_ill75rmpj7yds.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f263e022ef503e7fffcbc15ef59e515b84889d4c473b9113f3fea292725b9e37
3
- size 75532912
 
 
 
 
build/torch25-cxx98-cu124-x86_64-linux/attention/_ops.py DELETED
@@ -1,9 +0,0 @@
1
- import torch
2
- from . import _attention_ill75rmpj7yds
3
- ops = torch.ops._attention_ill75rmpj7yds
4
-
5
- def add_op_namespace_prefix(op_name: str):
6
- """
7
- Prefix op by namespace.
8
- """
9
- return f"_attention_ill75rmpj7yds::{op_name}"
 
 
 
 
 
 
 
 
 
 
build/torch25-cxx98-cu124-x86_64-linux/{attention β†’ paged_attention}/__init__.py RENAMED
File without changes
build/torch25-cxx98-cu124-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py RENAMED
File without changes
build/torch25-cxx98-cu124-x86_64-linux/paged_attention/_ops.py ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ from . import _paged_attention_fhq57q56w3m5o
3
+ ops = torch.ops._paged_attention_fhq57q56w3m5o
4
+
5
+ def add_op_namespace_prefix(op_name: str):
6
+ """
7
+ Prefix op by namespace.
8
+ """
9
+ return f"_paged_attention_fhq57q56w3m5o::{op_name}"
build/torch25-cxx98-cu124-x86_64-linux/paged_attention/_paged_attention_fhq57q56w3m5o.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e614a9c07805944a58249f9fbcf8ba6368c1e06e104382eba51ffa168c2d1667
3
+ size 75532944
build/torch25-cxx98-cu124-x86_64-linux/{attention β†’ paged_attention}/platforms.py RENAMED
File without changes
build/torch26-cxx11-cu118-x86_64-linux/attention/_attention_6qe5ft3kiteru.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e66eca8e825e5cee2dc18c1235319a4e5b1372d843cab74660e8d94792e02f7c
3
- size 78857896
 
 
 
 
build/torch26-cxx11-cu118-x86_64-linux/attention/_ops.py DELETED
@@ -1,9 +0,0 @@
1
- import torch
2
- from . import _attention_6qe5ft3kiteru
3
- ops = torch.ops._attention_6qe5ft3kiteru
4
-
5
- def add_op_namespace_prefix(op_name: str):
6
- """
7
- Prefix op by namespace.
8
- """
9
- return f"_attention_6qe5ft3kiteru::{op_name}"
 
 
 
 
 
 
 
 
 
 
build/torch26-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/__init__.py RENAMED
File without changes
build/torch26-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py RENAMED
File without changes
build/torch26-cxx11-cu118-x86_64-linux/paged_attention/_ops.py ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ from . import _paged_attention_xvepb4loq5mm2
3
+ ops = torch.ops._paged_attention_xvepb4loq5mm2
4
+
5
+ def add_op_namespace_prefix(op_name: str):
6
+ """
7
+ Prefix op by namespace.
8
+ """
9
+ return f"_paged_attention_xvepb4loq5mm2::{op_name}"
build/torch26-cxx11-cu118-x86_64-linux/paged_attention/_paged_attention_xvepb4loq5mm2.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9ea33ac4e0a3fb3aeac93525946f52326dd1a84a40e82ed86ff896f3a887b74
3
+ size 78857928
build/torch26-cxx11-cu118-x86_64-linux/{attention β†’ paged_attention}/platforms.py RENAMED
File without changes
build/torch26-cxx11-cu124-x86_64-linux/attention/_attention_ftq3cjdxqfw4m.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:913ba8f5166dc4e84ed8a2da4b1dc44c178a93eeb16aae9782176fb089a459a7
3
- size 75552112
 
 
 
 
build/torch26-cxx11-cu124-x86_64-linux/attention/_ops.py DELETED
@@ -1,9 +0,0 @@
1
- import torch
2
- from . import _attention_ftq3cjdxqfw4m
3
- ops = torch.ops._attention_ftq3cjdxqfw4m
4
-
5
- def add_op_namespace_prefix(op_name: str):
6
- """
7
- Prefix op by namespace.
8
- """
9
- return f"_attention_ftq3cjdxqfw4m::{op_name}"
 
 
 
 
 
 
 
 
 
 
build/torch26-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/__init__.py RENAMED
File without changes
build/torch26-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/_custom_ops.py RENAMED
File without changes
build/torch26-cxx11-cu124-x86_64-linux/paged_attention/_ops.py ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ from . import _paged_attention_uyfdujhnc2xoe
3
+ ops = torch.ops._paged_attention_uyfdujhnc2xoe
4
+
5
+ def add_op_namespace_prefix(op_name: str):
6
+ """
7
+ Prefix op by namespace.
8
+ """
9
+ return f"_paged_attention_uyfdujhnc2xoe::{op_name}"
build/torch26-cxx11-cu124-x86_64-linux/paged_attention/_paged_attention_uyfdujhnc2xoe.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f43b82896ccd22f8860fce338ab01c3eb93fba305c27d1b9a73252151ebcd534
3
+ size 75552144
build/torch26-cxx11-cu124-x86_64-linux/{attention β†’ paged_attention}/platforms.py RENAMED
File without changes
build/torch26-cxx11-cu126-x86_64-linux/attention/_attention_lkibbjh726iwm.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:91380eebc7db2ff85f92e687d388055f210123bac602a6bc273172834bf49012
3
- size 75376640
 
 
 
 
build/torch26-cxx11-cu126-x86_64-linux/attention/_ops.py DELETED
@@ -1,9 +0,0 @@
1
- import torch
2
- from . import _attention_lkibbjh726iwm
3
- ops = torch.ops._attention_lkibbjh726iwm
4
-
5
- def add_op_namespace_prefix(op_name: str):
6
- """
7
- Prefix op by namespace.
8
- """
9
- return f"_attention_lkibbjh726iwm::{op_name}"