You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_higher_order_ops/flex_attention.py", line 703, in flex_attention_autograd
out, logsumexp = FlexAttentionAutogradOp.apply(
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/autograd/function.py", line 575, in apply
return super().apply(*args, **kwargs) # type: ignore[misc]
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_higher_order_ops/flex_attention.py", line 591, in forward
out, logsumexp = flex_attention(
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_higher_order_ops/flex_attention.py", line 109, in __call__
return super().__call__(
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_ops.py", line 433, in __call__
return wrapper()
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 632, in _fn
return fn(*args, **kwargs)
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_ops.py", line 424, in wrapper
return torch.overrides.handle_torch_function(
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/overrides.py", line 1717, in handle_torch_function
result = mode.__torch_function__(public_api, types, args, kwargs)
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_higher_order_ops/flex_attention.py", line 85, in __torch_function__
return func(*args, **(kwargs or {}))
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_higher_order_ops/flex_attention.py", line 109, in __call__
return super().__call__(
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_ops.py", line 433, in __call__
return wrapper()
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 632, in _fn
return fn(*args, **kwargs)
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_ops.py", line 429, in wrapper
return self.dispatch(
File "/lustre/fsw/portfolios/llmservice/users/xiaolwang/software/miniforge3/envs/torchtitan-triton/lib/python3.10/site-packages/torch/_ops.py", line 334, in dispatch
raise NotImplementedError(
NotImplementedError: There was no rule registered for HOP flex_attention and mode <torch.utils.checkpoint._CachingTorchDispatchMode object at 0x14f2d67d1c90>. We recommend filing an issue.
Could you please help me with this? Thank you!
My env:
torch 2.5.0
The text was updated successfully, but these errors were encountered:
From the error message, I speculate it's because I'm using FlexAttention in a gradient checkpointed nn module. I'm wondering is there a way to make FlexAttention compatible with that?
Hi, thank you for releasing this wonderful code base.
When I'm trying to combine causal and cross-document mask, like below
I encountered this error:
Could you please help me with this? Thank you!
My env:
torch 2.5.0
The text was updated successfully, but these errors were encountered: