support attention bwd

Signed-off-by: Jinjie Liu <jjliu@baai.ac.cn>
This commit is contained in:
2026-02-10 17:01:28 +08:00
parent e41ec26329
commit 599957e156
6 changed files with 206 additions and 26 deletions

View File

@@ -1,8 +0,0 @@
cmake_minimum_required(VERSION 3.18)
project(${SKBUILD_PROJECT_NAME})
install(
DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR}/include
DESTINATION ${CMAKE_INSTALL_PREFIX}/triton_tvm_ffi
)

View File

@@ -16,6 +16,7 @@ Extra Credits:
import os
from pathlib import Path
import time
from typing import Sequence
import torch
import triton
@@ -731,13 +732,41 @@ class _attention_triton(torch.autograd.Function):
@triton_tvm_ffi.torch_wrap(
[_attn_fwd],
Path(__file__).parent / "attention.cc",
Path(__file__).parent / "attnfwd.cc",
)
def _attn_fwd_tvm_ffi(
q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, causal: bool, sm_scale: float
) -> torch.Tensor: ...
@triton_tvm_ffi.torch_wrap(
[_attn_bwd_preprocess],
Path(__file__).parent / "attnbwdpre.cc",
)
def _attn_bwd_preprocess_tvm_ffi(
o: torch.Tensor,
do: torch.Tensor,
mshape: Sequence[int],
head_dim: int,
): ...
@triton_tvm_ffi.torch_wrap(
[_attn_bwd],
Path(__file__).parent / "attnbwd.cc",
)
def _attn_bwd_tvm_ffi(
q: torch.Tensor,
k: torch.Tensor,
v: torch.Tensor,
sm_scale: float,
do: torch.Tensor,
m: torch.Tensor,
delta: torch.Tensor,
HEAD_DIM: int,
): ...
class _attention_tvm_ffi(_attention_triton):
@staticmethod
def forward(ctx, q, k, v, causal, sm_scale):
@@ -752,6 +781,31 @@ class _attention_tvm_ffi(_attention_triton):
ctx.causal = causal
return o
@staticmethod
def backward(ctx, do):
q, k, v, o, M = ctx.saved_tensors
delta = _attn_bwd_preprocess_tvm_ffi(
o,
do,
M.shape,
ctx.HEAD_DIM,
)
dq, dk, dv = _attn_bwd_tvm_ffi(
q,
k,
v,
ctx.sm_scale,
do,
M,
delta,
ctx.HEAD_DIM,
)
dq = torch.from_dlpack(dq)
dk = torch.from_dlpack(dk)
dv = torch.from_dlpack(dv)
return dq, dk, dv, None, None, None, None
def attn_torch(q, k, v, causal=False, sm_scale=1.0):
M = torch.tril(torch.ones((N_CTX, N_CTX), device=DEVICE))
@@ -778,7 +832,7 @@ if __name__ == "__main__":
N_CTX = 128
HEAD_DIM = 64
causal = True
mode = "fwd"
mode = "bwd"
dtype = torch.float16
torch.manual_seed(20)
q = (
@@ -811,25 +865,27 @@ if __name__ == "__main__":
ref_dq, q.grad = q.grad.clone(), None
tri_out = attn_triton(q, k, v, causal, sm_scale).half()
tvm_ffi_out = attn_tvm_ffi(q, k, v, causal, sm_scale).half()
warmup = 5
round = 1000
if mode == "fwd":
atol = 1e-2
torch.testing.assert_close(tri_out, ref_out, atol=atol, rtol=0)
torch.testing.assert_close(tvm_ffi_out, ref_out, atol=atol, rtol=0)
for _ in range(5):
attn_torch(q, k, v, causal, sm_scale)
attn_triton(q, k, v, causal, sm_scale)
attn_tvm_ffi(q, k, v, causal, sm_scale)
cp0 = time.perf_counter_ns()
for _ in range(round):
attn_torch(q, k, v, causal, sm_scale)
cp1 = time.perf_counter_ns()
for _ in range(round):
attn_triton(q, k, v, causal, sm_scale)
cp2 = time.perf_counter_ns()
for _ in range(round):
attn_tvm_ffi(q, k, v, causal, sm_scale)
cp3 = time.perf_counter_ns()
with torch.no_grad():
for _ in range(warmup):
attn_torch(q, k, v, causal, sm_scale)
attn_triton(q, k, v, causal, sm_scale)
attn_tvm_ffi(q, k, v, causal, sm_scale)
cp0 = time.perf_counter_ns()
for _ in range(round):
attn_torch(q, k, v, causal, sm_scale)
cp1 = time.perf_counter_ns()
for _ in range(round):
attn_triton(q, k, v, causal, sm_scale)
cp2 = time.perf_counter_ns()
for _ in range(round):
attn_tvm_ffi(q, k, v, causal, sm_scale)
cp3 = time.perf_counter_ns()
print(
f"PyTorch: {(cp1 - cp0) / round * 1e-6:.3f} ms\nTriton: {(cp2 - cp1) / round * 1e-6:.3f} ms\nTVM FFI: {(cp3 - cp2) / round * 1e-6:.3f} ms"
)
@@ -851,3 +907,31 @@ if __name__ == "__main__":
torch.testing.assert_close(tri_dv, ref_dv, atol=1e-2, rtol=rtol)
torch.testing.assert_close(tri_dk, ref_dk, atol=1e-2, rtol=rtol)
torch.testing.assert_close(tri_dq, ref_dq, atol=1e-2, rtol=rtol)
tvm_ffi_out.backward(dout)
tvm_ffi_dv, v.grad = v.grad.clone(), None
tvm_ffi_dk, k.grad = k.grad.clone(), None
tvm_ffi_dq, q.grad = q.grad.clone(), None
# compare
torch.testing.assert_close(tvm_ffi_out, ref_out, atol=1e-2, rtol=0)
rtol = 0.0
torch.testing.assert_close(tvm_ffi_dv, ref_dv, atol=1e-2, rtol=rtol)
torch.testing.assert_close(tvm_ffi_dk, ref_dk, atol=1e-2, rtol=rtol)
torch.testing.assert_close(tvm_ffi_dq, ref_dq, atol=1e-2, rtol=rtol)
for _ in range(warmup):
attn_torch(q, k, v, causal, sm_scale).backward(dout)
attn_triton(q, k, v, causal, sm_scale).backward(dout)
attn_tvm_ffi(q, k, v, causal, sm_scale).backward(dout)
cp0 = time.perf_counter_ns()
for _ in range(round):
attn_torch(q, k, v, causal, sm_scale).backward(dout)
cp1 = time.perf_counter_ns()
for _ in range(round):
attn_triton(q, k, v, causal, sm_scale).backward(dout)
cp2 = time.perf_counter_ns()
for _ in range(round):
attn_tvm_ffi(q, k, v, causal, sm_scale).backward(dout)
cp3 = time.perf_counter_ns()
print(
f"PyTorch: {(cp1 - cp0) / round * 1e-6:.3f} ms\nTriton: {(cp2 - cp1) / round * 1e-6:.3f} ms\nTVM FFI: {(cp3 - cp2) / round * 1e-6:.3f} ms"
)

View File

@@ -0,0 +1,57 @@
#include "ATen/core/ATen_fwd.h"
#include "ATen/ops/empty.h"
#include "c10/core/Device.h"
#include "torch/headeronly/core/DeviceType.h"
#include "tvm/ffi/container/tensor.h"
#include <ATen/DLConvertor.h>
#include <ATen/dlpack.h>
#include <tvm/ffi/extra/cuda/cubin_launcher.h>
#include <tvm/ffi/function.h>
#include <tvm/ffi/tvm_ffi.h>
#ifndef _ATTN_BWD_STUB
#define _ATTN_BWD_STUB(grid, device, stream, args, kwargs)
#endif
#ifndef _ATTN_BWD_TVM_FFI_NAME
#define _ATTN_BWD_TVM_FFI_NAME ""
#endif
tvm::ffi::Tuple<tvm::ffi::Tensor, tvm::ffi::Tensor, tvm::ffi::Tensor>
AttnBwd(tvm::ffi::Tensor q, tvm::ffi::Tensor k, tvm::ffi::Tensor v,
const double smScale, tvm::ffi::Tensor do_, tvm::ffi::Tensor m,
tvm::ffi::Tensor delta, const int32_t kHeadDim) {
tvm::ffi::ShapeView qshape = q.shape(), qstride = q.strides();
const int32_t kBatch = qshape[0], kNHead = qshape[1], kNCtx = qshape[2],
kBlockN1 = 128;
const double kArgKScale = smScale / log(2);
at::Tensor qTorch = at::fromDLPack(q.ToDLPack()),
kTorch = at::fromDLPack(k.ToDLPack()),
vTorch = at::fromDLPack(v.ToDLPack()),
dqTorch = at::empty_like(qTorch), dkTorch = at::empty_like(kTorch),
dvTorch = at::empty_like(vTorch),
argKTorch = at::mul(kTorch, kArgKScale);
tvm::ffi::Tensor dq = tvm::ffi::Tensor::FromDLPack(at::toDLPack(dqTorch)),
dk = tvm::ffi::Tensor::FromDLPack(at::toDLPack(dkTorch)),
dv = tvm::ffi::Tensor::FromDLPack(at::toDLPack(dvTorch)),
argK = tvm::ffi::Tensor::FromDLPack(at::toDLPack(argKTorch));
tvm::ffi::Tuple<int32_t, int32_t, int32_t> grid(kNCtx / kBlockN1, 1,
kBatch * kNHead);
tvm::ffi::Array<tvm::ffi::Any> args = {
q, argK, v, smScale, do_, dq, dk, dv,
m, delta, qstride[0], qstride[1], qstride[2], qstride[3], kNHead, kNCtx};
tvm::ffi::Map<tvm::ffi::String, tvm::ffi::Any> kwargs = {
{"BLOCK_M1", 32}, {"BLOCK_N1", kBlockN1}, {"BLOCK_M2", 128},
{"BLOCK_N2", 32}, {"BLK_SLICE_FACTOR", 2}, {"HEAD_DIM", kHeadDim},
{"num_warps", 4}, {"num_stages", 5},
};
DLDevice device = q.device();
void *stream = TVMFFIEnvGetStream(device.device_type, device.device_id);
_ATTN_BWD_STUB(grid, device.device_id, stream, args, kwargs);
return tvm::ffi::Tuple{dq, dk, dv};
}
TVM_FFI_STATIC_INIT_BLOCK() {
namespace refl = tvm::ffi::reflection;
refl::GlobalDef().def(_ATTN_BWD_TVM_FFI_NAME, AttnBwd);
}

View File

@@ -0,0 +1,45 @@
#include "ATen/core/ATen_fwd.h"
#include "ATen/ops/empty.h"
#include "torch/headeronly/core/DeviceType.h"
#include "tvm/ffi/container/tensor.h"
#include <ATen/DLConvertor.h>
#include <ATen/dlpack.h>
#include <tvm/ffi/extra/cuda/cubin_launcher.h>
#include <tvm/ffi/function.h>
#include <tvm/ffi/tvm_ffi.h>
#ifndef _ATTN_BWD_PREPROCESS_STUB
#define _ATTN_BWD_PREPROCESS_STUB(grid, device, stream, args, kwargs)
#endif
#ifndef _ATTN_BWD_PREPROCESS_TVM_FFI_NAME
#define _ATTN_BWD_PREPROCESS_TVM_FFI_NAME ""
#endif
tvm::ffi::Tensor AttnBwdPreprocess(tvm::ffi::Tensor o, tvm::ffi::Tensor do_,
tvm::ffi::Shape mshape,
const int32_t kHeadDim) {
const int32_t kBatch = mshape[0], kNHead = mshape[1], kNCtx = mshape[2],
kPreBlock = 128;
at::Tensor deltaTorch = at::empty(mshape, at::kFloat, std::nullopt,
at::Device(at::kCUDA, o.device().device_id),
std::nullopt, std::nullopt);
tvm::ffi::Tensor delta =
tvm::ffi::Tensor::FromDLPack(at::toDLPack(deltaTorch));
tvm::ffi::Tuple<int32_t, int32_t, int32_t> grid(kNCtx / kPreBlock,
kBatch * kNHead, 1);
tvm::ffi::Array<tvm::ffi::Any> args = {o, do_, delta, kBatch, kNHead, kNCtx};
tvm::ffi::Map<tvm::ffi::String, tvm::ffi::Any> kwargs = {
{"BLOCK_M", kPreBlock},
{"HEAD_DIM", kHeadDim},
};
DLDevice device = o.device();
void *stream = TVMFFIEnvGetStream(device.device_type, device.device_id);
_ATTN_BWD_PREPROCESS_STUB(grid, device.device_id, stream, args, kwargs);
return delta;
}
TVM_FFI_STATIC_INIT_BLOCK() {
namespace refl = tvm::ffi::reflection;
refl::GlobalDef().def(_ATTN_BWD_PREPROCESS_TVM_FFI_NAME, AttnBwdPreprocess);
}

View File

@@ -1,3 +1,5 @@
#include "ATen/core/ATen_fwd.h"
#include "ATen/ops/empty.h"
#include "tvm/ffi/container/tensor.h"
#include <ATen/DLConvertor.h>
#include <ATen/dlpack.h>

View File

@@ -47,8 +47,8 @@ class TVMFFIJITFunction(object):
):
args: Iterator[Any] = map(self.canonicalize, args)
kwargs: Dict[str, Any] = {
k: v for k, v in zip(self.signature, args) if v is not None
} | {k: self.canonicalize(v) for k, v in kwargs.items()}
k: self.canonicalize(v) for k, v in kwargs.items()
}
kernel: CompiledKernel = self.fn[grid](*args, **kwargs)
self.num_warps, _, self.shmem = kernel.packed_metadata
self.ctypes = [type_canonicalize(v) for v in kernel.src.signature.values()]