Chaowei Xiao

Orcid: 0000-0002-7043-4926

According to our database1, Chaowei Xiao authored at least 132 papers between 2014 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
Voyager: An Open-Ended Embodied Agent with Large Language Models.
Trans. Mach. Learn. Res., 2024

Prismer: A Vision-Language Model with Multi-Task Experts.
Trans. Mach. Learn. Res., 2024

Time to Think the Security of WiFi-Based Behavior Recognition Systems.
IEEE Trans. Dependable Secur. Comput., 2024

InjecGuard: Benchmarking and Mitigating Over-defense in Prompt Injection Guardrail Models.
CoRR, 2024

FATH: Authentication-based Test-time Defense against Indirect Prompt Injection Attacks.
CoRR, 2024

SudoLM: Learning Access Control of Parametric Knowledge with Authorization Alignment.
CoRR, 2024

RePD: Defending Jailbreak Attack through a Retrieval-based Prompt Decomposition Process.
CoRR, 2024

LeanAgent: Lifelong Learning for Formal Theorem Proving.
CoRR, 2024

AutoDAN-Turbo: A Lifelong Agent for Strategy Self-Exploration to Jailbreak LLMs.
CoRR, 2024

System-Level Defense against Indirect Prompt Injection Attacks: An Information Flow Control Perspective.
CoRR, 2024

HaloScope: Harnessing Unlabeled LLM Generations for Hallucination Detection.
CoRR, 2024

EIA: Environmental Injection Attack on Generalist Web Agents for Privacy Leakage.
CoRR, 2024

IDNet: A Novel Dataset for Identity Document Analysis and Fraud Detection.
CoRR, 2024

Can Editing LLMs Inject Harm?
CoRR, 2024

AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases.
CoRR, 2024

Consistency Purification: Effective and Efficient Diffusion Purification towards Certified Robustness.
CoRR, 2024

UniGen: A Unified Framework for Textual Dataset Generation Using Large Language Models.
CoRR, 2024

MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding.
CoRR, 2024

Visual-RolePlay: Universal Jailbreak Attack on MultiModal Large Language Models via Role-playing Image Characte.
CoRR, 2024

AI Risk Management Should Incorporate Both Safety and Security.
CoRR, 2024

Safeguarding Vision-Language Models Against Patched Visual Prompt Injectors.
CoRR, 2024

JailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against Jailbreak Attacks.
CoRR, 2024

Automatic and Universal Prompt Injection Attacks against Large Language Models.
CoRR, 2024

A New Era in LLM Security: Exploring Security Concerns in Real-World LLM-based Systems.
CoRR, 2024

WIPI: A New Web Threat for LLM-Driven Web Agents.
CoRR, 2024

Mitigating Fine-tuning Jailbreak Attack with Backdoor Enhanced Alignment.
CoRR, 2024

T-Stitch: Accelerating Sampling in Pre-Trained Diffusion Models with Trajectory Stitching.
CoRR, 2024

A Trembling House of Cards? Mapping Adversarial Attacks against Language Agents.
CoRR, 2024

Preference Poisoning Attacks on Reward Model Learning.
CoRR, 2024

A Learning-based Declarative Privacy-Preserving Framework for Federated Data Management.
CoRR, 2024

TrustLLM: Trustworthiness in Large Language Models.
CoRR, 2024

Differentially Private Video Activity Recognition.
Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2024

Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models.
Proceedings of the 33rd USENIX Security Symposium, 2024


Cognitive Overload: Jailbreaking Large Language Models with Overloaded Logical Thinking.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2024, 2024

Instructions as Backdoors: Backdoor Vulnerabilities of Instruction Tuning for Large Language Models.
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024

Instructional Fingerprinting of Large Language Models.
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024

From Shortcuts to Triggers: Backdoor Defense with Denoised PoE.
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024

ChatGPT as an Attack Tool: Stealthy Textual Backdoor Attack via Blackbox Generative Model Trigger.
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024

Reinforcement Learning with Human Feedback for Realistic Traffic Simulation.
Proceedings of the IEEE International Conference on Robotics and Automation, 2024


CALICO: Self-Supervised Camera-LiDAR Contrastive Pre-training for BEV Perception.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

Conversational Drug Editing Using Retrieval and Domain Feedback.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

Leveraging Hierarchical Feature Sharing for Efficient Dataset Condensation.
Proceedings of the Computer Vision - ECCV 2024, 2024

AdaShield : Safeguarding Multimodal Large Language Models from Structure-Based Attack via Adaptive Shield Prompting.
Proceedings of the Computer Vision - ECCV 2024, 2024

Dolphins: Multimodal Language Model for Driving.
Proceedings of the Computer Vision - ECCV 2024, 2024

RealGen: Retrieval Augmented Generation for Controllable Traffic Scenarios.
Proceedings of the Computer Vision - ECCV 2024, 2024

Perada: Parameter-Efficient Federated Learning Personalization with Generalization Guarantees.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024

Mitigating Backdoor Threats to Large Language Models: Advancement and Challenges.
Proceedings of the 60th Annual Allerton Conference on Communication, 2024

RLHFPoison: Reward Poisoning Attack for Reinforcement Learning with Human Feedback in Large Language Models.
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2024

2023
Multi-modal molecule structure-text model for text-based retrieval and editing.
Nat. Mac. Intell., December, 2023

GenSLMs: Genome-scale language models reveal SARS-CoV-2 evolutionary dynamics.
Int. J. High Perform. Comput. Appl., November, 2023

Behavior Privacy Preserving in RF Sensing.
IEEE Trans. Dependable Secur. Comput., 2023

Taxonomy of Machine Learning Safety: A Survey and Primer.
ACM Comput. Surv., 2023

Exploring the Limits of ChatGPT in Software Security Applications.
CoRR, 2023

DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions.
CoRR, 2023

Test-time Backdoor Mitigation for Black-Box Large Language Models with Defensive Demonstrations.
CoRR, 2023

On the Exploitability of Reinforcement Learning with Human Feedback for Large Language Models.
CoRR, 2023

DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery through Sophisticated AI System Technologies.
CoRR, 2023

CSI: Enhancing the Robustness of 3D Point Cloud Recognition against Corruption.
CoRR, 2023

ChatGPT-powered Conversational Drug Editing Using Retrieval and Domain Feedback.
CoRR, 2023

Adversarial Demonstration Attacks on Large Language Models.
CoRR, 2023

From Shortcuts to Triggers: Backdoor Defense with Denoised PoE.
CoRR, 2023

Mole Recruitment: Poisoning of Image Classifiers via Selective Batch Sampling.
CoRR, 2023

Prismer: A Vision-Language Model with An Ensemble of Experts.
CoRR, 2023

PerAda: Parameter-Efficient and Generalizable Federated Learning Personalization with Guarantees.
CoRR, 2023

A Text-guided Protein Design Framework.
CoRR, 2023

DiffSmooth: Certifiably Robust Learning via Diffusion Models and Local Smoothing.
Proceedings of the 32nd USENIX Security Symposium, 2023

SMACK: Semantically Meaningful Adversarial Audio Attack.
Proceedings of the 32nd USENIX Security Symposium, 2023

On the Exploitability of Instruction Tuning.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

CodeIPPrompt: Intellectual Property Infringement Assessment of Code Language Models.
Proceedings of the International Conference on Machine Learning, 2023

A Critical Revisit of Adversarial Robustness in 3D Point Cloud Recognition with Diffusion-Driven Purification.
Proceedings of the International Conference on Machine Learning, 2023

DensePure: Understanding Diffusion Models for Adversarial Robustness.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Defending against Adversarial Audio via Diffusion Model.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Retrieval-based Controllable Molecule Generation.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Re-ViLM: Retrieval-Augmented Visual Language Model for Zero and Few-Shot Image Captioning.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, 2023

Shall We Pretrain Autoregressive Language Models with Retrieval? A Comprehensive Study.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

HiCL: Hierarchical Contrastive Learning of Unsupervised Sentence Embeddings.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, 2023

Detecting Backdoors During the Inference Stage Based on Corruption Robustness Consistency.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023

VoxFormer: Sparse Voxel Transformer for Camera-Based 3D Semantic Scene Completion.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023

Semantic Adversarial Attacks via Diffusion Models.
Proceedings of the 34th British Machine Vision Conference 2023, 2023

Defending against Insertion-based Textual Backdoor Attacks via Attribution.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, 2023

2022
DensePure: Understanding Diffusion Models towards Adversarial Robustness.
CoRR, 2022

PointDP: Diffusion-driven Purification against Adversarial Attacks on 3D Point Cloud Recognition.
CoRR, 2022

Benchmarking Robustness of 3D Point Cloud Recognition Against Common Corruptions.
CoRR, 2022

Exploring the Limits of Domain-Adaptive Training for Detoxifying Large-Scale Language Models.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Test-Time Prompt Tuning for Zero-Shot Generalization in Vision-Language Models.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Physical-World Attack towards WiFi-based Behavior Recognition.
Proceedings of the IEEE INFOCOM 2022, 2022

Understanding The Robustness in Vision Transformers.
Proceedings of the International Conference on Machine Learning, 2022

Diffusion Models for Adversarial Purification.
Proceedings of the International Conference on Machine Learning, 2022

RelViT: Concept-guided Vision Transformer for Visual Relational Reasoning.
Proceedings of the Tenth International Conference on Learning Representations, 2022

SecretGen: Privacy Recovery on Pre-trained Models via Distribution Discrimination.
Proceedings of the Computer Vision - ECCV 2022, 2022

AdvDO: Realistic Adversarial Attacks for Trajectory Prediction.
Proceedings of the Computer Vision - ECCV 2022, 2022

Robust Trajectory Prediction against Adversarial Attacks.
Proceedings of the Conference on Robot Learning, 2022

Characterizing Attacks on Deep Reinforcement Learning.
Proceedings of the 21st International Conference on Autonomous Agents and Multiagent Systems, 2022

2021
Auditing AI models for Verified Deployment under Semantic Specifications.
CoRR, 2021

Practical Machine Learning Safety: A Survey and Primer.
CoRR, 2021

Invisible for both Camera and LiDAR: Security of Multi-Sensor Fusion based Perception in Autonomous Driving Under Physical-World Attacks.
Proceedings of the 42nd IEEE Symposium on Security and Privacy, 2021

Long-Short Transformer: Efficient Transformers for Language and Vision.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

AugMax: Adversarial Composition of Random Augmentations for Robust Training.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Adversarially Robust 3D Point Cloud Recognition Using Self-Supervisions.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

ADVM'21: 1st International Workshop on Adversarial Learning for Multimedia.
Proceedings of the MM '21: ACM Multimedia Conference, Virtual Event, China, October 20, 2021

A Behavior Privacy Preserving Method towards RF Sensing.
Proceedings of the 29th IEEE/ACM International Symposium on Quality of Service, 2021

Can Shape Structure Features Improve Model Robustness under Diverse Adversarial Settings?
Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, 2021

Application-driven Privacy-preserving Data Publishing with Correlated Attributes.
Proceedings of the EWSN '21: Proceedings of the 2021 International Conference on Embedded Wireless Systems and Networks, 2021

2020
Machine Learning in Adversarial Environments.
PhD thesis, 2020

Robust Deep Reinforcement Learning against Adversarial Perturbations on Observations.
CoRR, 2020

Robust Deep Reinforcement Learning against Adversarial Perturbations on State Observations.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Towards Stable and Efficient Training of Verifiably Robust Neural Networks.
Proceedings of the 8th International Conference on Learning Representations, 2020

SemanticAdv: Generating Adversarial Examples via Attribute-Conditioned Image Editing.
Proceedings of the Computer Vision - ECCV 2020, 2020

2019
Characterizing Attacks on Deep Reinforcement Learning.
CoRR, 2019

Adversarial Objects Against LiDAR-Based Autonomous Driving Systems.
CoRR, 2019

SemanticAdv: Generating Adversarial Examples via Attribute-conditional Image Editing.
CoRR, 2019

Towards Stable and Efficient Training of Verifiably Robust Neural Networks.
CoRR, 2019

Improving Robustness of ML Classifiers against Realizable Evasion Attacks Using Conserved Features.
Proceedings of the 28th USENIX Security Symposium, 2019

Performing Co-membership Attacks Against Deep Generative Models.
Proceedings of the 2019 IEEE International Conference on Data Mining, 2019

AdvIT: Adversarial Frames Identifier Based on Temporal Consistency in Videos.
Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, 2019

MeshAdv: Adversarial Meshes for Visual Recognition.
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019

Adversarial Sensor Attack on LiDAR-based Perception in Autonomous Driving.
Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, 2019

2018
Automatic Radio Map Adaptation for Indoor Localization Using Smartphones.
IEEE Trans. Mob. Comput., 2018

Protecting Sensitive Attributes via Generative Adversarial Networks.
CoRR, 2018

Data Poisoning Attack against Unsupervised Node Embedding Methods.
CoRR, 2018

Realistic Adversarial Examples in 3D Meshes.
CoRR, 2018

From Patching Delays to Infection Symptoms: Using Risk Profiles for an Early Discovery of Vulnerabilities Exploited in the Wild.
Proceedings of the 27th USENIX Security Symposium, 2018

Generating Adversarial Examples with Adversarial Networks.
Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, 2018

Spatially Transformed Adversarial Examples.
Proceedings of the 6th International Conference on Learning Representations, 2018

Characterizing Adversarial Examples Based on Spatial Consistency Information for Semantic Segmentation.
Proceedings of the Computer Vision - ECCV 2018, 2018

Robust Physical-World Attacks on Deep Learning Visual Classification.
Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, 2018

2017
Patch Me If You Can: A Study on the Effects of Individual User Behavior on the End-Host Vulnerability State.
Proceedings of the Passive and Active Measurement - 18th International Conference, 2017

2015
Static power of mobile devices: Self-updating radio maps for wireless indoor localization.
Proceedings of the 2015 IEEE Conference on Computer Communications, 2015

2014
Tagoram: real-time tracking of mobile RFID tags to high precision using COTS devices.
Proceedings of the 20th Annual International Conference on Mobile Computing and Networking, 2014


  Loading...