Yangchen Pan

Orcid: 0009-0000-8297-9045

According to our database1, Yangchen Pan authored at least 32 papers between 2016 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
Label Alignment Regularization for Distribution Shift.
J. Mach. Learn. Res., 2024

DTR-Bench: An in silico Environment and Benchmark Platform for Reinforcement Learning Based Dynamic Treatment Regime.
CoRR, 2024

Reinforcement Learning in Dynamic Treatment Regimes Needs Critical Reexamination.
CoRR, 2024

An MRP Formulation for Supervised Learning: Generalized Temporal Difference Learning Models.
CoRR, 2024

A Simple Mixture Policy Parameterization for Improving Sample Efficiency of CVaR Optimization.
RLJ, 2024

Position: Reinforcement Learning in Dynamic Treatment Regimes Needs Critical Reexamination.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Improving Adversarial Transferability via Model Alignment.
Proceedings of the Computer Vision - ECCV 2024, 2024

2023
Understanding the robustness difference between stochastic gradient descent and adaptive gradient methods.
Trans. Mach. Learn. Res., 2023

Memory-efficient Reinforcement Learning with Value-based Knowledge Consolidation.
Trans. Mach. Learn. Res., 2023

Conditionally optimistic exploration for cooperative deep multi-agent reinforcement learning.
Proceedings of the Uncertainty in Artificial Intelligence, 2023

An Alternative to Variance: Gini Deviation for Risk-averse Policy Gradient.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

The In-Sample Softmax for Offline Reinforcement Learning.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Greedy Actor-Critic: A New Conditional Cross-Entropy Method for Policy Improvement.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

2022
Label Alignment Regularization for Distribution Shift.
CoRR, 2022

Memory-efficient Reinforcement Learning with Knowledge Consolidation.
CoRR, 2022

Understanding and mitigating the limitations of prioritized experience replay.
Proceedings of the Uncertainty in Artificial Intelligence, 2022

TOPS: Transition-Based Volatility-Reduced Policy Search.
Proceedings of the Autonomous Agents and Multiagent Systems. Best and Visionary Papers, 2022

An Alternate Policy Gradient Estimator for Softmax Policies.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2022

2021
Fuzzy Tiling Activations: A Simple Approach to Learning Sparse Representations Online.
Proceedings of the 9th International Conference on Learning Representations, 2021

2020
Beyond Prioritized Replay: Sampling States in Model-Based RL via Simulated Priorities.
CoRR, 2020

An implicit function learning approach for parametric modal regression.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Frequency-based Search-control in Dyna.
Proceedings of the 8th International Conference on Learning Representations, 2020

Maxmin Q-learning: Controlling the Estimation Bias of Q-learning.
Proceedings of the 8th International Conference on Learning Representations, 2020

2019
Deep Tile Coder: an Efficient Sparse Representation Learning Approach with applications in Reinforcement Learning.
CoRR, 2019

Hill Climbing on Value Estimates for Search-control in Dyna.
Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, 2019

2018
Actor-Expert: A Framework for using Action-Value Methods in Continuous Action Spaces.
CoRR, 2018

Organizing Experience: a Deeper Look at Replay Mechanisms for Sample-Based Planning in Continuous State Domains.
Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, 2018

Reinforcement Learning with Function-Valued Action Spaces for Partial Differential Equation Control.
Proceedings of the 35th International Conference on Machine Learning, 2018

2017
Effective sketching methods for value function approximation.
Proceedings of the Thirty-Third Conference on Uncertainty in Artificial Intelligence, 2017

Adapting Kernel Representations Online Using Submodular Maximization.
Proceedings of the 34th International Conference on Machine Learning, 2017

Accelerated Gradient Temporal Difference Learning.
Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, 2017

2016
Incremental Truncated LSTD.
Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, 2016


  Loading...