Skip to content

Actions: NVIDIA/TransformerEngine

Documentation

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
251 workflow run results
251 workflow run results

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

Implement fused kernel for FP8 scale update
Documentation #2259: Pull request #593 synchronize by timmoon10
January 10, 2024 02:47 51s timmoon10:fused-fp8-scale-update
January 10, 2024 02:47 51s
[PyTorch] cuda graph support
Documentation #2254: Pull request #575 synchronize by ksivaman
January 9, 2024 05:28 47s ksivaman:fp8_cuda_graphs
January 9, 2024 05:28 47s
[Paddle] Support GQA
Documentation #2253: Pull request #595 synchronize by Wong4j
January 9, 2024 05:16 53s Wong4j:jaywan/add_gqa
January 9, 2024 05:16 53s
[Paddle] Add sequence parallel
Documentation #2244: Pull request #561 synchronize by ksivaman
January 8, 2024 15:42 1m 0s Tom-Zheng:gh_add_sp
January 8, 2024 15:42 1m 0s
[Paddle] Add sequence parallel
Documentation #2243: Pull request #561 synchronize by ksivaman
January 8, 2024 15:33 1m 5s Tom-Zheng:gh_add_sp
January 8, 2024 15:33 1m 5s
[Paddle] Add sequence parallel
Documentation #2242: Pull request #561 synchronize by ksivaman
January 8, 2024 15:32 48s Tom-Zheng:gh_add_sp
January 8, 2024 15:32 48s
[JAX][Common] Support GQA
Documentation #2240: Pull request #578 synchronize by zlsh80826
January 8, 2024 09:07 1m 8s zlsh80826:rewang/gqa-clean
January 8, 2024 09:07 1m 8s
[JAX][Common] Support GQA
Documentation #2239: Pull request #578 synchronize by zlsh80826
January 8, 2024 09:06 1m 8s zlsh80826:rewang/gqa-clean
January 8, 2024 09:06 1m 8s
[JAX][Common] Support GQA
Documentation #2237: Pull request #578 synchronize by zlsh80826
January 8, 2024 07:14 1m 6s zlsh80826:rewang/gqa-clean
January 8, 2024 07:14 1m 6s
[Paddle] Support GQA
Documentation #2230: Pull request #595 synchronize by Wong4j
January 7, 2024 14:49 46s Wong4j:jaywan/add_gqa
January 7, 2024 14:49 46s
[PyTorch] cuda graph support
Documentation #2228: Pull request #575 synchronize by ksivaman
January 6, 2024 13:03 53s ksivaman:fp8_cuda_graphs
January 6, 2024 13:03 53s
[Common/PyTorch] Fix FP8 fused attention input args
Documentation #2226: Pull request #592 synchronize by cyanguwa
January 6, 2024 02:01 1m 5s cyanguwa:fused_attn/fp8_dims
January 6, 2024 02:01 1m 5s
NVRTC kernels for cast-transpose
Documentation #2225: Pull request #258 synchronize by timmoon10
January 6, 2024 01:47 51s timmoon10:nvrtc-cast-transpose
January 6, 2024 01:47 51s