Ruobing Han

Orcid: 0000-0002-3090-3951

According to our database1, Ruobing Han authored at least 14 papers between 2019 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
CuPBoP: Making CUDA a Portable Language.
ACM Trans. Design Autom. Electr. Syst., 2024

Unleashing CPU Potential for Executing GPU Programs Through Compiler/Runtime Optimizations.
Proceedings of the 57th IEEE/ACM International Symposium on Microarchitecture, 2024

Enabling Fine-Grained Incremental Builds by Making Compiler Stateful.
Proceedings of the IEEE/ACM International Symposium on Code Generation and Optimization, 2024

Exponentially Expanding the Phase-Ordering Search Space via Dormant Information.
Proceedings of the 33rd ACM SIGPLAN International Conference on Compiler Construction, 2024

2023
Haplotype-resolved Genome of Sika Deer Reveals Allele-specific Gene Expression and Chromosome Evolution.
Genom. Proteom. Bioinform., June, 2023

CuPBoP: A Framework to Make CUDA Portable.
Proceedings of the 28th ACM SIGPLAN Annual Symposium on Principles and Practice of Parallel Programming, 2023

2022
GradientFlow: Optimizing Network Performance for Large-Scale Distributed DNN Training.
IEEE Trans. Big Data, 2022

COX : Exposing CUDA Warp-level Functions to CPUs.
ACM Trans. Archit. Code Optim., 2022

CuPBoP: CUDA for Parallelized and Broad-range Processors.
CoRR, 2022

2021
COX: CUDA on X86 by Exposing Warp-Level Functions to CPUs.
CoRR, 2021

Supporting CUDA for an extended RISC-V GPU architecture.
CoRR, 2021

Auto-Precision Scaling for Distributed Deep Learning.
Proceedings of the High Performance Computing - 36th International Conference, 2021

Dynamic scaling for low-precision learning.
Proceedings of the PPoPP '21: 26th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming, 2021

2019
Optimizing Network Performance for Distributed DNN Training on GPU Clusters: ImageNet/AlexNet Training in 1.5 Minutes.
CoRR, 2019


  Loading...