Skip to content

Use fused softmax kernel in llama attention layer #3356

Use fused softmax kernel in llama attention layer

Use fused softmax kernel in llama attention layer #3356

Re-run triggered October 23, 2024 17:43
Status Skipped
Total duration 4s
Artifacts

ci_cuda.yaml

on: pull_request
test-cuda
0s
test-cuda
Fit to window
Zoom out
Zoom in