Yaodong Yu

Orcid: 0000-0003-0540-8526

According to our database1, Yaodong Yu authored at least 43 papers between 2017 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
Scheduling a multi-agent flow shop with two scenarios and release dates.
Int. J. Prod. Res., January, 2024

Causal Image Modeling for Efficient Visual Understanding.
CoRR, 2024

Accuracy on the wrong line: On the pitfalls of noisy data for out-of-distribution generalisation.
CoRR, 2024

Scaling White-Box Transformers for Vision.
CoRR, 2024

A Study on the Calibration of In-context Learning.
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024

ViP: A Differentially Private Foundation Model for Computer Vision.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Differentially Private Representation Learning via Image Captioning.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

A Global Geometric Analysis of Maximal Coding Rate Reduction.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Masked Completion via Structured Diffusion with White-Box Transformers.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

2023
White-Box Transformers via Sparse Rate Reduction: Compression Is All There Is?
CoRR, 2023

Emergence of Segmentation with Minimalistic White-Box Transformers.
CoRR, 2023

Scaff-PD: Communication Efficient Fair and Robust Federated Learning.
CoRR, 2023

White-Box Transformers via Sparse Rate Reduction.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Federated Conformal Predictors for Distributed Uncertainty Quantification.
Proceedings of the International Conference on Machine Learning, 2023

2022
ReduNet: A White-box Deep Network from the Principle of Maximizing Rate Reduction.
J. Mach. Learn. Res., 2022

CTRL: Closed-Loop Transcription to an LDR via Minimaxing Rate Reduction.
Entropy, 2022

What You See is What You Get: Distributional Generalization for Algorithm Design in Deep Learning.
CoRR, 2022

Robust Calibration with Multi-domain Temperature Scaling.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

TCT: Convexifying Federated Learning using Bootstrapped Neural Tangent Kernels.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

What You See is What You Get: Principled Deep Learning via Distributional Generalization.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Predicting Out-of-Distribution Error with the Projection Norm.
Proceedings of the International Conference on Machine Learning, 2022

Online Nonsubmodular Minimization with Delayed Costs: From Full Information to Bandit Feedback.
Proceedings of the International Conference on Machine Learning, 2022

Conditional Supervised Contrastive Learning for Fair Text Classification.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, 2022

Fast Distributionally Robust Learning with Variance-Reduced Min-Max Optimization.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2022

On the Convergence of Stochastic Extragradient for Bilinear Games using Restarted Iteration Averaging.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2022

2021
The Effect of Model Size on Worst-Group Generalization.
CoRR, 2021

Closed-Loop Data Transcription to an LDR via Minimaxing Rate Reduction.
CoRR, 2021

On the Convergence of Stochastic Extragradient for Bilinear Games with Restarted Iteration Averaging.
CoRR, 2021

Understanding Generalization in Adversarial Training via the Bias-Variance Decomposition.
CoRR, 2021

Adversarial Robustness of Stabilized Neural ODE Might be from Obfuscated Gradients.
Proceedings of the Mathematical and Scientific Machine Learning, 2021

2020
Deep Networks from the Principle of Rate Reduction.
CoRR, 2020

Adversarial Robustness of Stabilized NeuralODEs Might be from Obfuscated Gradients.
CoRR, 2020

Learning Diverse and Discriminative Representations via the Principle of Maximal Coding Rate Reduction.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Boundary thickness and robustness in learning models.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Rethinking Bias-Variance Trade-off for Generalization of Neural Networks.
Proceedings of the 37th International Conference on Machine Learning, 2020

2019
Theoretically Principled Trade-off between Robustness and Accuracy.
Proceedings of the 36th International Conference on Machine Learning, 2019

Learning One-hidden-layer ReLU Networks via Gradient Descent.
Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics, 2019

2018
Third-order Smoothness Helps: Faster Stochastic Optimization Algorithms for Finding Local Minima.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

A Primal-Dual Analysis of Global Optimality in Nonconvex Low-Rank Matrix Recovery.
Proceedings of the 35th International Conference on Machine Learning, 2018

Data Poisoning Attacks on Multi-Task Relationship Learning.
Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, 2018

2017
Third-order Smoothness Helps: Even Faster Stochastic Optimization Algorithms for Finding Local Minima.
CoRR, 2017

Saving Gradient and Negative Curvature Computations: Finding Local Minima More Efficiently.
CoRR, 2017

Communication-Efficient Distributed Primal-Dual Algorithm for Saddle Point Problem.
Proceedings of the Thirty-Third Conference on Uncertainty in Artificial Intelligence, 2017


  Loading...