Evaluating What Others Say: The Effect of Accuracy Assessment in Shaping Mental Models of AI Systems.
Proc. ACM Hum. Comput. Interact., 2024
Grounding with Structure: Exploring Design Variations of Grounded Human-AI Collaboration in a Natural Language Interface.
Proc. ACM Hum. Comput. Interact., 2024
Black-box Uncertainty Quantification Method for LLM-as-a-Judge.
CoRR, 2024
Aligning Human and LLM Judgments: Insights from EvalAssist on Task-Specific Evaluations and AI-assisted Assessment Strategy Preferences.
CoRR, 2024
Emerging Reliance Behaviors in Human-AI Text Generation: Hallucinations, Data Quality Assessment, and Cognitive Forcing Functions.
CoRR, 2024
Human-Centered Design Recommendations for LLM-as-a-Judge.
CoRR, 2024
EvaluLLM: LLM assisted evaluation of generative outputs.
Proceedings of the Companion Proceedings of the 29th International Conference on Intelligent User Interfaces, 2024
Decision Making Strategies and Team Efficacy in Human-AI Teams.
Proc. ACM Hum. Comput. Interact., April, 2023
Follow the Successful Herd: Towards Explanations for Improved Use and Mental Models of Natural Language Systems.
,
,
,
,
,
,
,
,
,
,
,
,
,
Proceedings of the 28th International Conference on Intelligent User Interfaces, 2023
SME-in-the-loop: Interaction Preferences when Supervising Bots in Human-AI Communities.
Proceedings of the 2023 ACM Designing Interactive Systems Conference, 2023
A Goal-Driven Natural Language Interface for Creating Application Integration Workflows.
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,
Proceedings of the Thirty-Sixth AAAI Conference on Artificial Intelligence, 2022
The measurement of atmospheric water vapor: radiometer comparison and spatial variations.
IEEE Trans. Geosci. Remote. Sens., 1991