Tao Ge
Affiliations:- Tencent AI Lab, Seattle, WA, USA
Timeline
Legend:
Book In proceedings Article PhD thesis Dataset OtherLinks
On csauthors.net:
Bibliography
2024
Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens.
CoRR, 2024
Router-Tuning: A Simple and Effective Approach for Enabling Dynamic-Depth in Transformers.
CoRR, 2024
Learn Beyond The Answer: Training Language Models with Reflection for Mathematical Reasoning.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024