Skip to content

[common/PyTorch] Add FusedAttention support for SWA (left, right) #6446

[common/PyTorch] Add FusedAttention support for SWA (left, right)

[common/PyTorch] Add FusedAttention support for SWA (left, right) #6446

Triggered via pull request December 12, 2024 05:42
Status Success
Total duration 2m 15s
Artifacts

lint.yml

on: pull_request
Fit to window
Zoom out
Zoom in

Annotations

6 warnings
JAX C++
ubuntu-latest pipelines will use ubuntu-24.04 soon. For more details, see https://github.com/actions/runner-images/issues/10636
PyTorch C++
ubuntu-latest pipelines will use ubuntu-24.04 soon. For more details, see https://github.com/actions/runner-images/issues/10636
JAX Python
ubuntu-latest pipelines will use ubuntu-24.04 soon. For more details, see https://github.com/actions/runner-images/issues/10636
PaddlePaddle C++
ubuntu-latest pipelines will use ubuntu-24.04 soon. For more details, see https://github.com/actions/runner-images/issues/10636
PaddlePaddle Python
ubuntu-latest pipelines will use ubuntu-24.04 soon. For more details, see https://github.com/actions/runner-images/issues/10636
PyTorch Python
ubuntu-latest pipelines will use ubuntu-24.04 soon. For more details, see https://github.com/actions/runner-images/issues/10636