Niladri S. Chatterji

According to our database1, Niladri S. Chatterji authored at least 25 papers between 2017 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
Proving Test Set Contamination in Black-Box Language Models.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

2023
Holistic Evaluation of Language Models.
Trans. Mach. Learn. Res., 2023

Undersampling is a Minimax Optimal Robustness Intervention in Nonparametric Classification.
Trans. Mach. Learn. Res., 2023

Random Feature Amplification: Feature Learning and Generalization in Neural Networks.
J. Mach. Learn. Res., 2023

Deep linear networks can benignly overfit when shallow ones do.
J. Mach. Learn. Res., 2023

2022
The Interplay Between Implicit Bias and Benign Overfitting in Two-Layer Linear Networks.
J. Mach. Learn. Res., 2022

Foolish Crowds Support Benign Overfitting.
J. Mach. Learn. Res., 2022

Is Importance Weighting Incompatible with Interpolating Classifiers?
Proceedings of the Tenth International Conference on Learning Representations, 2022

Benign Overfitting without Linearity: Neural Network Classifiers Trained by Gradient Descent for Noisy Linear Data.
Proceedings of the Conference on Learning Theory, 2-5 July 2022, London, UK., 2022

2021
Why do Gradient Methods Work in Optimization and Sampling?
PhD thesis, 2021

When Does Gradient Descent with Logistic Loss Find Interpolating Two-Layer Networks?
J. Mach. Learn. Res., 2021

Finite-sample Analysis of Interpolating Linear Classifiers in the Overparameterized Regime.
J. Mach. Learn. Res., 2021

On the Opportunities and Risks of Foundation Models.
CoRR, 2021

On the Theory of Reinforcement Learning with Once-per-Episode Feedback.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

When does gradient descent with logistic loss interpolate using deep networks with smoothed ReLU activations?
Proceedings of the Conference on Learning Theory, 2021

2020
Oracle lower bounds for stochastic gradient sampling algorithms.
CoRR, 2020

The intriguing role of module criticality in the generalization of deep networks.
Proceedings of the 8th International Conference on Learning Representations, 2020

OSOM: A simultaneously optimal algorithm for multi-armed and linear contextual bandits.
Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, 2020

Langevin Monte Carlo without smoothness.
Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, 2020

2019
Is There an Analog of Nesterov Acceleration for MCMC?
CoRR, 2019

Online learning with kernel losses.
Proceedings of the 36th International Conference on Machine Learning, 2019

2018
Sharp Convergence Rates for Langevin Dynamics in the Nonconvex Setting.
CoRR, 2018

On the Theory of Variance Reduction for Stochastic Gradient Monte Carlo.
Proceedings of the 35th International Conference on Machine Learning, 2018

Underdamped Langevin MCMC: A non-asymptotic analysis.
Proceedings of the Conference On Learning Theory, 2018

2017
Alternating minimization for dictionary learning with random initialization.
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017


  Loading...