Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
11 changes: 9 additions & 2 deletions einx/_src/frontend/impl/jax.py
Original file line number Diff line number Diff line change
@@ -1,3 +1,6 @@
from typing import ParamSpec, TypeVar, Concatenate, cast
from collections.abc import Callable

import einx._src.tracer as tracer
import einx._src.adapter as adapter
from ..api import api
Expand Down Expand Up @@ -42,7 +45,11 @@ def get_shape(tensor):
return {"optimizations": optimizations, "compiler": tracer.compiler.python, "is_supported_tensor": is_supported_tensor, "get_shape": get_shape}


def adapt_with_vmap(op, signature=None):
P = ParamSpec("P")
R = TypeVar("R")


def adapt_with_vmap(op: Callable[P, R], signature=None) -> Callable[Concatenate[str, P], R]:
iskwarg = _make_iskwarg(op)
jax = tracer.signature.jax()

Expand All @@ -55,7 +62,7 @@ def adapt_with_vmap(op, signature=None):
op = adapter.namedtensor_calltensorfactory.op(op, expected_type=jax.numpy.ndarray)
op = adapter.einx_from_namedtensor.op(op, iskwarg=iskwarg, el_op=signature, implicit_output="bijective")

return api(op, backend=types.SimpleNamespace(**_get_backend_kwargs()))
return cast(Callable[Concatenate[str, P], R], api(op, backend=types.SimpleNamespace(**_get_backend_kwargs())))


adapt_with_vmap.__doc__ = _make_doc_adapt_with_vmap("jax", "``jax.vmap``")
Expand Down
11 changes: 9 additions & 2 deletions einx/_src/frontend/impl/mlx.py
Original file line number Diff line number Diff line change
@@ -1,3 +1,6 @@
from typing import ParamSpec, TypeVar, Concatenate, cast
from collections.abc import Callable

import einx._src.tracer as tracer
import einx._src.adapter as adapter
from ..types import Tensor
Expand Down Expand Up @@ -33,7 +36,11 @@ def get_shape(tensor):
return {"optimizations": optimizations, "compiler": tracer.compiler.python, "is_supported_tensor": is_supported_tensor, "get_shape": get_shape}


def adapt_with_vmap(op, signature=None):
P = ParamSpec("P")
R = TypeVar("R")


def adapt_with_vmap(op: Callable[P, R], signature=None) -> Callable[Concatenate[str, P], R]:
iskwarg = _make_iskwarg(op)
mlx = tracer.signature.mlx()

Expand All @@ -46,7 +53,7 @@ def adapt_with_vmap(op, signature=None):
op = adapter.namedtensor_calltensorfactory.op(op, expected_type=mlx.core.array)
op = adapter.einx_from_namedtensor.op(op, iskwarg=iskwarg, el_op=signature, implicit_output="bijective")

return api(op, backend=types.SimpleNamespace(**_get_backend_kwargs()))
return cast(Callable[Concatenate[str, P], R], api(op, backend=types.SimpleNamespace(**_get_backend_kwargs())))


adapt_with_vmap.__doc__ = _make_doc_adapt_with_vmap("mlx", "``mlx.core.vmap``")
Expand Down
11 changes: 9 additions & 2 deletions einx/_src/frontend/impl/torch.py
Original file line number Diff line number Diff line change
@@ -1,3 +1,6 @@
from typing import ParamSpec, TypeVar, Concatenate, cast
from collections.abc import Callable

import einx._src.tracer as tracer
import einx._src.adapter as adapter
from ..api import api
Expand Down Expand Up @@ -62,7 +65,11 @@ def get_shape(tensor):
return {"optimizations": optimizations, "compiler": tracer.compiler.python, "is_supported_tensor": is_supported_tensor, "get_shape": get_shape}


def adapt_with_vmap(op, signature=None):
P = ParamSpec("P")
R = TypeVar("R")


def adapt_with_vmap(op: Callable[P, R], signature=None) -> Callable[Concatenate[str, P], R]:
_raise_on_invalid_version()
iskwarg = _make_iskwarg(op)

Expand All @@ -85,7 +92,7 @@ def adapt_with_vmap(op, signature=None):

torch.compiler.allow_in_graph(op)

return op
return cast(Callable[Concatenate[str, P], R], op)


adapt_with_vmap.__doc__ = _make_doc_adapt_with_vmap("torch", "``torch.vmap``")
Expand Down