🔥
Fire!
- Shanghai, China
Pinned Loading
-
SJTU-IPADS/gnnlab
SJTU-IPADS/gnnlab PublicA Factored System for Sample-based GNN Training over GPUs
-
SJTU-IPADS/fgnn-artifacts
SJTU-IPADS/fgnn-artifacts PublicFGNN's artifact evaluation (EuroSys 2022)
-
PaddlePaddle/Paddle
PaddlePaddle/Paddle PublicPArallel Distributed Deep LEarning: Machine Learning Framework from Industrial Practice (『飞桨』核心框架,深度学习&机器学习高性能单机、分布式训练和跨平台部署)
-
llvm/llvm-project
llvm/llvm-project PublicThe LLVM Project is a collection of modular and reusable compiler and toolchain technologies.
-
-
FlagOpen/FlagPerf
FlagOpen/FlagPerf PublicFlagPerf is an open-source software platform for benchmarking AI chips.
227 contributions in the last year
Day of Week | March Mar | April Apr | May May | June Jun | July Jul | August Aug | September Sep | October Oct | November Nov | December Dec | January Jan | February Feb | March Mar | ||||||||||||||||||||||||||||||||||||||||
Sunday Sun | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Monday Mon | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Tuesday Tue | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Wednesday Wed | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Thursday Thu | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Friday Fri | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Saturday Sat |
Less
No contributions.
Low contributions.
Medium-low contributions.
Medium-high contributions.
High contributions.
More
Contribution activity
March 2025
Created 6 commits in 1 repository
Created a pull request in PaddlePaddle/Paddle that received 6 comments
[XPU] add conv3d_transpose
PR Category Custom Device PR Types New features Description [XPU] add conv3d_transpose
+647
−3
lines changed
•
6
comments
Opened 5 other pull requests in 1 repository
PaddlePaddle/Paddle
5
merged
-
[XPU] bump XCCL to 3.0.2.5
This contribution was made on Mar 12
-
paddle.distributed.all_to_all supports unequal split(#71429)
This contribution was made on Mar 6
-
[XPU] support all_to_all_unequal_split_size
This contribution was made on Mar 4
-
[XPU] add isfinite/isinf support
This contribution was made on Mar 3
-
[XPU] reduce_xxx and broadcast_xxx use int64_t shape
This contribution was made on Mar 3
Reviewed 8 pull requests in 1 repository
PaddlePaddle/Paddle
8 pull requests
-
[XPU] support all_to_all_unequal_split_size
This contribution was made on Mar 11
-
[XPU] feat: add xpu async memory copy to enable zero cost checkpoint
This contribution was made on Mar 7
-
[XPU] Support print runtime error log for xdnn/xfa/xpudnn error
This contribution was made on Mar 7
-
[XPU] fix batch_norm_grad when use global status
This contribution was made on Mar 6
-
[XPU] add quantize_linear and dequantize_linear op
This contribution was made on Mar 6
-
[XPU] add grid sampler support.
This contribution was made on Mar 5
-
[XPU] reduce_xxx and broadcast_xxx use int64_t shape
This contribution was made on Mar 5
-
[xpu] support: each parameter has different lr in merged_momentum
This contribution was made on Mar 5
Created an issue in PaddlePaddle/Paddle that received 1 comment
paddle.distributed.all_to_all不支持unequal_split_size的语义
PyTorch paddle.distributed.all_to_all
unequal split size的例子(https://pytorch.org/docs/stable/distributed.html#torch.distributed.all_to_all):
>>> i…
1
comment