Chen Liang

Affiliations:
  • Georgia Institute of Technology, Atlanta, USA


According to our database1, Chen Liang authored at least 22 papers between 2020 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
On Parameter Efficiency of Neural Language Models.
PhD thesis, 2024

GRIN: GRadient-INformed MoE.
CoRR, 2024

Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling.
CoRR, 2024

Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone.
CoRR, 2024

LoftQ: LoRA-Fine-Tuning-aware Quantization for Large Language Models.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

2023
HomoDistil: Homotopic Task-Agnostic Distillation of Pre-trained Transformers.
CoRR, 2023

Module-wise Adaptive Distillation for Multimodality Foundation Models.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Less is More: Task-aware Layer-wise Distillation for Language Model Compression.
Proceedings of the International Conference on Machine Learning, 2023

LoSparse: Structured Compression of Large Language Models based on Low-Rank and Sparse Approximation.
Proceedings of the International Conference on Machine Learning, 2023

HomoDistil: Homotopic Task-Agnostic Distillation of Pre-trained Transformers.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

2022
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation.
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2022

Self-Training with Differentiable Teacher.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2022, 2022

PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance.
Proceedings of the International Conference on Machine Learning, 2022

No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models.
Proceedings of the Tenth International Conference on Learning Representations, 2022

CAMERO: Consistency Regularized Ensemble of Perturbed Language Models with Weight Sharing.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022

2021
Adversarial Training as Stackelberg Game: An Unrolled Optimization Approach.
CoRR, 2021

Adversarial Regularization as Stackelberg Game: An Unrolled Optimization Approach.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021

ARCH: Efficient Adversarial Regularized Training with Caching.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2021, 2021

Token-wise Curriculum Learning for Neural Machine Translation.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2021, 2021

Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021

2020
BOND: BERT-Assisted Open-Domain Named Entity Recognition with Distant Supervision.
Proceedings of the KDD '20: The 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2020

Multi-Domain Neural Machine Translation with Word-Level Adaptive Layer-wise Domain Mixing.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020


  Loading...