Hongyu Wang

Orcid: 0000-0003-1811-3903

Affiliations:
  • Institute of Computing Technology, Chinese Academy of Sciences, China


According to our database1, Hongyu Wang authored at least 8 papers between 2022 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
DeepNet: Scaling Transformers to 1,000 Layers.
IEEE Trans. Pattern Anal. Mach. Intell., October, 2024

Q-Sparse: All Large Language Models can be Fully Sparsely-Activated.
CoRR, 2024

The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits.
CoRR, 2024

2023
BitNet: Scaling 1-bit Transformers for Large Language Models.
CoRR, 2023

Magneto: A Foundation Transformer.
Proceedings of the International Conference on Machine Learning, 2023

2022
TorchScale: Transformers at Scale.
CoRR, 2022

Foundation Transformers.
CoRR, 2022

DeepNet: Scaling Transformers to 1, 000 Layers.
CoRR, 2022


  Loading...