Tian Xu
Orcid: 0000-0001-9409-448XAffiliations:
- Nanjing University, National Key Laboratory for Novel Software Technology, China
According to our database1,
Tian Xu
authored at least 21 papers
between 2019 and 2024.
Collaborative distances:
Collaborative distances:
Timeline
Legend:
Book In proceedings Article PhD thesis Dataset OtherLinks
Online presence:
-
on orcid.org
On csauthors.net:
Bibliography
2024
Model gradient: unified model and policy learning in model-based reinforcement learning.
Frontiers Comput. Sci., August, 2024
Entropic Distribution Matching in Supervised Fine-tuning of LLMs: Less Overfitting and Better Diversity.
CoRR, 2024
ReMax: A Simple, Effective, and Efficient Reinforcement Learning Method for Aligning Large Language Models.
Proceedings of the Forty-first International Conference on Machine Learning, 2024
Proceedings of the Forty-first International Conference on Machine Learning, 2024
Reward-Consistent Dynamics Models are Strongly Generalizable for Offline Reinforcement Learning.
Proceedings of the Twelfth International Conference on Learning Representations, 2024
Proceedings of the Second Tiny Papers Track at ICLR 2024, 2024
Proceedings of the Twelfth International Conference on Learning Representations, 2024
2023
ReMax: A Simple, Effective, and Efficient Reinforcement Learning Method for Aligning Large Language Models.
CoRR, 2023
Proceedings of the Uncertainty in Artificial Intelligence, 2023
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023
2022
IEEE Trans. Pattern Anal. Mach. Intell., 2022
Understanding Adversarial Imitation Learning in Small Sample Regime: A Stage-coupled Analysis.
CoRR, 2022
CoRR, 2022
2021
Nearly Minimax Optimal Adversarial Imitation Learning with Known and Unknown Transitions.
CoRR, 2021
2020
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020
2019