Nathan Lambert

Orcid: 0000-0002-9997-6817

Affiliations:
  • University of California Berkeley, Department of Electrical Engineering and Computer Sciences, CA, USA


According to our database1, Nathan Lambert authored at least 34 papers between 2018 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
A Unified View on Solving Objective Mismatch in Model-Based Reinforcement Learning.
Trans. Mach. Learn. Res., 2024

A Survey on Data Selection for Language Models.
Trans. Mach. Learn. Res., 2024

WildGuard: Open One-Stop Moderation Tools for Safety Risks, Jailbreaks, and Refusals of LLMs.
CoRR, 2024

Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback.
CoRR, 2024

Towards a Framework for Openness in Foundation Models: Proceedings from the Columbia Convening on Openness in Artificial Intelligence.
CoRR, 2024

D2PO: Discriminator-Guided DPO with Response Evaluation Models.
CoRR, 2024

Social Choice for AI Alignment: Dealing with Diverse Human Feedback.
CoRR, 2024

RewardBench: Evaluating Reward Models for Language Modeling.
CoRR, 2024

OLMo: Accelerating the Science of Language Models.
CoRR, 2024

Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research.
CoRR, 2024

2023
Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2.
CoRR, 2023

The Alignment Ceiling: Objective Mismatch in Reinforcement Learning from Human Feedback.
CoRR, 2023

Zephyr: Direct Distillation of LM Alignment.
CoRR, 2023

Entangled Preferences: The History and Risks of Reinforcement Learning and Human Feedback.
CoRR, 2023

Confidence-Building Measures for Artificial Intelligence: Workshop Proceedings.
CoRR, 2023

BLISS: Interplanetary Exploration with Swarms of Low-Cost Spacecraft.
CoRR, 2023

Reward Reports for Reinforcement Learning.
Proceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society, 2023

2022
Measuring Data.
CoRR, 2022

Reward Reports for Reinforcement Learning.
CoRR, 2022

Investigating Compounding Prediction Errors in Learned Dynamics Models.
CoRR, 2022

Choices, Risks, and Reward Reports: Charting Public Policy for Reinforcement Learning Systems.
CoRR, 2022

The Challenges of Exploration for Offline Reinforcement Learning.
CoRR, 2022

2021
Nonholonomic Yaw Control of an Underactuated Flying Robot With Model-Based Reinforcement Learning.
IEEE Robotics Autom. Lett., 2021

Axes for Sociotechnical Inquiry in AI Research.
CoRR, 2021

MBRL-Lib: A Modular Library for Model-based Reinforcement Learning.
CoRR, 2021

BotNet: A Simulator for Studying the Effects of Accurate Communication Models on Multi-Agent and Swarm Control.
Proceedings of the International Symposium on Multi-Robot and Multi-Agent Systems, 2021

Learning Accurate Long-term Dynamics for Model-based Reinforcement Learning.
Proceedings of the 2021 60th IEEE Conference on Decision and Control (CDC), 2021

On the Importance of Hyperparameter Optimization for Model-based Reinforcement Learning.
Proceedings of the 24th International Conference on Artificial Intelligence and Statistics, 2021

2020
Learning for Microrobot Exploration: Model-based Locomotion, Sparse-robust Navigation, and Low-power Deep Classification.
CoRR, 2020

Objective Mismatch in Model-based Reinforcement Learning.
Proceedings of the 2nd Annual Conference on Learning for Dynamics and Control, 2020

AI Development for the Public Interest: From Abstraction Traps to Sociotechnical Risks.
Proceedings of the IEEE International Symposium on Technology and Society, 2020

Learning Generalizable Locomotion Skills with Hierarchical Reinforcement Learning.
Proceedings of the 2020 IEEE International Conference on Robotics and Automation, 2020

2019
Low-Level Control of a Quadrotor With Deep Model-Based Reinforcement Learning.
IEEE Robotics Autom. Lett., 2019

2018
Toward Controlled Flight of the Ionocraft: A Flying Microrobot Using Electrohydrodynamic Thrust With Onboard Sensing and No Moving Parts.
IEEE Robotics Autom. Lett., 2018


  Loading...