-
Notifications
You must be signed in to change notification settings - Fork 653
Pull requests: NVIDIA/TransformerEngine
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
[Common] Remove redundant grad_logits zero-initialization in fused router backward kernels
#2745
opened Mar 7, 2026 by
roycho96
Loading…
5 of 13 tasks
[JAX] Add bias support for v2 grouped GEMM path
#2744
opened Mar 6, 2026 by
jberchtold-nvidia
Loading…
8 of 13 tasks
[Common] Persistent Grouped NVFP4 quantization kernel
#2743
opened Mar 6, 2026 by
Oleg-Goncharov
Loading…
8 of 13 tasks
Add guard at lowest JAX version that still supports triton kernel calling
#2741
opened Mar 6, 2026 by
tdophung
Loading…
6 of 13 tasks
[Common] Persistent Grouped MXFP8 quantization kernel
enhancement
New feature or request
MoE
#2738
opened Mar 5, 2026 by
Oleg-Goncharov
Loading…
9 of 13 tasks
Feat/cp nvshmem enhanced
community-contribution
PRs from external contributor outside the core maintainers, representing community-driven work.
#2737
opened Mar 5, 2026 by
Knight-of-Thunder
Loading…
13 tasks
[PyTorch debug] Fix issue with tp_group=None
#2733
opened Mar 4, 2026 by
pggPL
Loading…
8 of 13 tasks
Feature/unswizzle
community-contribution
PRs from external contributor outside the core maintainers, representing community-driven work.
#2732
opened Mar 4, 2026 by
int-smart
Loading…
9 of 13 tasks
fix: scope get_full_cu_seqlens cache key by device and inference mode
#2728
opened Mar 3, 2026 by
DmCarpe93
Loading…
8 of 13 tasks
Add DCP compatibility for FSDP2-TP sharding in TransformerEngine.
#2713
opened Feb 26, 2026 by
cspades
Loading…
3 of 13 tasks
Enable dequantization from MXFP8 tensor with only columnwise data
#2712
opened Feb 26, 2026 by
ptrendx
Loading…
13 tasks
[Draft] Newton-Schulz via cuSOLVERMp
#2706
opened Feb 25, 2026 by
vcherepanov-nv
Loading…
6 of 13 tasks
Fix Flash Attention 3 API compatibility for window size parameters
2.14.0
#2704
opened Feb 25, 2026 by
jhvmhg
Loading…
3 of 13 tasks
[PyTorch] Zero-initialize learnable softmax_offset in DotProductAttention
#2694
opened Feb 20, 2026 by
fjosw
Loading…
7 of 13 tasks
Enable sm120 support for fused attn if cuDNN is 9.18.1+
#2693
opened Feb 20, 2026 by
KshitijLakhani
•
Draft
13 tasks
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.