Yang Zhang

Orcid: 0000-0003-3612-7348

Affiliations:
  • CISPA, Helmholtz Center for Information Security, Saarbrücken, Germany
  • Saarland University, Saarland Informatics Campus, Saarbrücken, Germany


According to our database1, Yang Zhang authored at least 153 papers between 2013 and 2025.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
A Comprehensive Study of Privacy Risks in Curriculum Learning.
Proc. Priv. Enhancing Technol., 2025

2024
VeriTrain: Validating MLaaS Training Efforts via Anomaly Detection.
IEEE Trans. Dependable Secur. Comput., 2024

Link Stealing Attacks Against Inductive Graph Neural Networks.
Proc. Priv. Enhancing Technol., 2024

<i>ModSCAN</i>: Measuring Stereotypical Bias in Large Vision-Language Models from Vision and Language Modalities.
CoRR, 2024

Generated Data with Fake Privacy: Hidden Dangers of Fine-tuning Large Language Models on Generated Data.
CoRR, 2024

Understanding Data Importance in Machine Learning Attacks: Does Valuable Data Pose Greater Harm?
CoRR, 2024

Membership Inference Attack Against Masked Image Modeling.
CoRR, 2024

Vera Verto: Multimodal Hijacking Attack.
CoRR, 2024

Breaking Agents: Compromising Autonomous LLM Agents Through Malfunction Amplification.
CoRR, 2024

Towards Understanding Unsafe Video Generation.
CoRR, 2024

ICLGuard: Controlling In-Context Learning Behavior for Applicability Authorization.
CoRR, 2024

SOS! Soft Prompt Attack Against Open-Source Large Language Models.
CoRR, 2024

Voice Jailbreak Attacks Against GPT-4o.
CoRR, 2024

UnsafeBench: Benchmarking Image Safety Classifiers on Real-World and AI-Generated Images.
CoRR, 2024

Efficient Data-Free Model Stealing with Label Diversity.
CoRR, 2024

VGMShield: Mitigating Misuse of Video Generative Models.
CoRR, 2024

Prompt Stealing Attacks Against Large Language Models.
CoRR, 2024

Rapid Adoption, Hidden Risks: The Dual Impact of Large Language Model Customization.
CoRR, 2024

Comprehensive Assessment of Jailbreak Attacks Against LLMs.
CoRR, 2024

Conversation Reconstruction Attack Against GPT Models.
CoRR, 2024

Generated Distributions Are All You Need for Membership Inference Attacks Against Generative Models.
Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2024

Instruction Backdoor Attacks Against Customized LLMs.
Proceedings of the 33rd USENIX Security Symposium, 2024

SecurityNet: Assessing Machine Learning Vulnerabilities on Public Models.
Proceedings of the 33rd USENIX Security Symposium, 2024

Quantifying Privacy Risks of Prompts in Visual Prompt Learning.
Proceedings of the 33rd USENIX Security Symposium, 2024

Prompt Stealing Attacks Against Text-to-Image Generation Models.
Proceedings of the 33rd USENIX Security Symposium, 2024

You Only Prompt Once: On the Capabilities of Prompt Learning on Large Language Models to Tackle Toxic Content.
Proceedings of the IEEE Symposium on Security and Privacy, 2024

Test-Time Poisoning Attacks Against Test-Time Adaptation Models.
Proceedings of the IEEE Symposium on Security and Privacy, 2024

Composite Backdoor Attacks Against Large Language Models.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2024, 2024

Games and Beyond: Analyzing the Bullet Chats of Esports Livestreaming.
Proceedings of the Eighteenth International AAAI Conference on Web and Social Media, 2024

Detection and Attribution of Models Trained on Generated Data.
Proceedings of the IEEE International Conference on Acoustics, 2024

The Death and Life of Great Prompts: Analyzing the Evolution of LLM Prompts from the Structural Perspective.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

ModSCAN: Measuring Stereotypical Bias in Large Vision-Language Models from Vision and Language Modalities.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Reconstruct Your Previous Conversations! Comprehensively Investigating Privacy Leakage Risks in Conversations with GPT Models.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Inside the Black Box: Detecting Data Leakage in Pre-Trained Language Encoders.
Proceedings of the ECAI 2024 - 27th European Conference on Artificial Intelligence, 19-24 October 2024, Santiago de Compostela, Spain, 2024

Image-Perfect Imperfections: Safety, Bias, and Authenticity in the Shadow of Text-To-Image Model Evolution.
Proceedings of the 2024 on ACM SIGSAC Conference on Computer and Communications Security, 2024

Membership Inference Attacks Against In-Context Learning.
Proceedings of the 2024 on ACM SIGSAC Conference on Computer and Communications Security, 2024

"Do Anything Now": Characterizing and Evaluating In-The-Wild Jailbreak Prompts on Large Language Models.
Proceedings of the 2024 on ACM SIGSAC Conference on Computer and Communications Security, 2024

ZeroFake: Zero-Shot Detection of Fake Images Generated and Edited by Text-to-Image Generation Models.
Proceedings of the 2024 on ACM SIGSAC Conference on Computer and Communications Security, 2024

SeqMIA: Sequential-Metric Based Membership Inference Attack.
Proceedings of the 2024 on ACM SIGSAC Conference on Computer and Communications Security, 2024

LAMPS '24: ACM CCS Workshop on Large AI Systems and Models with Privacy and Safety Analysis.
Proceedings of the 2024 on ACM SIGSAC Conference on Computer and Communications Security, 2024

MGTBench: Benchmarking Machine-Generated Text Detection.
Proceedings of the 2024 on ACM SIGSAC Conference on Computer and Communications Security, 2024

FAKEPCD: Fake Point Cloud Detection via Source Attribution.
Proceedings of the 19th ACM Asia Conference on Computer and Communications Security, 2024

2023
Comprehensive Assessment of Toxicity in ChatGPT.
CoRR, 2023

On the Proactive Generation of Unsafe Images From Text-To-Image Models Using Benign Prompts.
CoRR, 2023

Last One Standing: A Comparative Analysis of Security and Privacy of Soft Prompt Tuning, LoRA, and In-Context Learning.
CoRR, 2023

Prompt Backdoors in Visual Prompt Learning.
CoRR, 2023

Robustness Over Time: Understanding Adversarial Examples' Effectiveness on Longitudinal Versions of Large Language Models.
CoRR, 2023

White-box Membership Inference Attacks against Diffusion Models.
CoRR, 2023

You Are How You Walk: Quantifying Privacy Risks in Step Count Data.
CoRR, 2023

Mondrian: Prompt Abstraction Attack Against Large Language Models for Cheaper API Pricing.
CoRR, 2023

Generative Watermarking Against Unauthorized Subject-Driven Image Synthesis.
CoRR, 2023

Vision-language models boost food composition compilation.
CoRR, 2023

Watermarking Diffusion Model.
CoRR, 2023

In ChatGPT We Trust? Measuring and Characterizing the Reliability of ChatGPT.
CoRR, 2023

From Visual Prompt Learning to Zero-Shot Transfer: Mapping Is All You Need.
CoRR, 2023

A Plot is Worth a Thousand Words: Model Information Stealing Attacks via Scientific Plots.
Proceedings of the 32nd USENIX Security Symposium, 2023

PrivTrace: Differentially Private Trajectory Synthesis by Adaptive Markov Models.
Proceedings of the 32nd USENIX Security Symposium, 2023

Two-in-One: A Model Hijacking Attack Against Text Generation Models.
Proceedings of the 32nd USENIX Security Symposium, 2023

UnGANable: Defending Against GAN-based Face Manipulation.
Proceedings of the 32nd USENIX Security Symposium, 2023

FACE-AUDITOR: Data Auditing in Facial Recognition Systems.
Proceedings of the 32nd USENIX Security Symposium, 2023

On the Evolution of (Hateful) Memes by Means of Multimodal Contrastive Learning.
Proceedings of the 44th IEEE Symposium on Security and Privacy, 2023

Backdoor Attacks Against Dataset Distillation.
Proceedings of the 30th Annual Network and Distributed System Security Symposium, 2023

Generated Graph Detection.
Proceedings of the International Conference on Machine Learning, 2023

Data Poisoning Attacks Against Multimodal Encoders.
Proceedings of the International Conference on Machine Learning, 2023

Is Adversarial Training Really a Silver Bullet for Mitigating Data Poisoning?
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Can't Steal? Cont-Steal! Contrastive Stealing Attacks Against Image Encoders.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023

DE-FAKE: Detection and Attribution of Fake Images Generated by Text-to-Image Generation Models.
Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, 2023

Unsafe Diffusion: On the Generation of Unsafe Images and Hateful Memes From Text-To-Image Models.
Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, 2023

Differentially Private Resource Allocation.
Proceedings of the Annual Computer Security Applications Conference, 2023

NOTABLE: Transferable Backdoor Attacks Against Prompt-based NLP Models.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

Pseudo Label-Guided Model Inversion Attack via Conditional Generative Adversarial Network.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023

2022
FairSR: Fairness-aware Sequential Recommendation through Multi-Task Learning with Preference Graph Embeddings.
ACM Trans. Intell. Syst. Technol., 2022

Fine-Tuning Is All You Need to Mitigate Backdoor Attacks.
CoRR, 2022

DE-FAKE: Detection and Attribution of Fake Images Generated by Text-to-Image Diffusion Models.
CoRR, 2022

Backdoor Attacks in the Supply Chain of Masked Image Modeling.
CoRR, 2022

Membership Inference Attacks Against Text-to-image Generation Models.
CoRR, 2022

PrivTrace: Differentially Private Trajectory Synthesis by Adaptive Markov Model.
CoRR, 2022

Membership-Doctor: Comprehensive Assessment of Membership Inference Against Machine Learning Models.
CoRR, 2022

ML-Doctor: Holistic Risk Assessment of Inference Attacks Against Machine Learning Models.
Proceedings of the 31st USENIX Security Symposium, 2022

Inference Attacks Against Graph Neural Networks.
Proceedings of the 31st USENIX Security Symposium, 2022

Teacher Model Fingerprinting Attacks Against Transfer Learning.
Proceedings of the 31st USENIX Security Symposium, 2022

Model Stealing Attacks Against Inductive Graph Neural Networks.
Proceedings of the 43rd IEEE Symposium on Security and Privacy, 2022

Amplifying Membership Exposure via Data Poisoning.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Property Inference Attacks Against GANs.
Proceedings of the 29th Annual Network and Distributed System Security Symposium, 2022

Get a Model! Model Hijacking Attack Against Machine Learning Models.
Proceedings of the 29th Annual Network and Distributed System Security Symposium, 2022

On Xing Tian and the Perseverance of Anti-China Sentiment Online.
Proceedings of the Sixteenth International AAAI Conference on Web and Social Media, 2022

Dynamic Backdoor Attacks Against Machine Learning Models.
Proceedings of the 7th IEEE European Symposium on Security and Privacy, 2022

Semi-Leak: Membership Inference Attacks Against Semi-supervised Learning.
Proceedings of the Computer Vision - ECCV 2022, 2022

Why So Toxic?: Measuring and Triggering Toxic Behavior in Open-Domain Chatbots.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

Finding MNEMON: Reviving Memories of Node Embeddings.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

Membership Inference Attacks by Exploiting Loss Trajectory.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

Auditing Membership Leakages of Multi-Exit Networks.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

On the Privacy Risks of Cell-Based NAS Architectures.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

SSLGuard: A Watermarking Scheme for Self-supervised Learning Pre-trained Encoders.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

Graph Unlearning.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

2021
DatingSec: Detecting Malicious Accounts in Dating Apps Using a Content-Based Attention Network.
IEEE Trans. Dependable Secur. Comput., 2021

On exploring feature representation learning of items to forecast their rise and fall in social media.
J. Intell. Inf. Syst., 2021

Node-Level Membership Inference Attacks Against Graph Neural Networks.
CoRR, 2021

"Go eat a bat, Chang!": On the Emergence of Sinophobic Behavior on Web Communities in the Face of COVID-19.
Proceedings of the WWW '21: The Web Conference 2021, 2021

Stealing Links from Graph Neural Networks.
Proceedings of the 30th USENIX Security Symposium, 2021

PrivSyn: Differentially Private Data Synthesis.
Proceedings of the 30th USENIX Security Symposium, 2021

MLCapsule: Guarded Offline Deployment of Machine Learning as a Service.
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2021

Membership Inference Attacks Against Recommender Systems.
Proceedings of the CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, Republic of Korea, November 15, 2021

Membership Leakage in Label-Only Exposures.
Proceedings of the CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, Republic of Korea, November 15, 2021

Quantifying and Mitigating Privacy Risks of Contrastive Learning.
Proceedings of the CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, Republic of Korea, November 15, 2021

When Machine Unlearning Jeopardizes Privacy.
Proceedings of the CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, Republic of Korea, November 15, 2021

BadNL: Backdoor Attacks against NLP Models with Semantic-preserving Improvements.
Proceedings of the ACSAC '21: Annual Computer Security Applications Conference, Virtual Event, USA, December 6, 2021

2020
Don't Trigger Me! A Triggerless Backdoor Attack Against Deep Neural Networks.
CoRR, 2020

BAAAN: Backdoor Attacks Against Autoencoder and GAN-Based Machine Learning Models.
CoRR, 2020

Privacy Analysis of Deep Learning in the Wild: Membership Inference Attacks against Transfer Learning.
CoRR, 2020

Label-Leaks: Membership Inference Attack with Label.
CoRR, 2020

BadNL: Backdoor Attacks Against NLP Models.
CoRR, 2020

"Go eat a bat, Chang!": An Early Look on the Emergence of Sinophobic Behavior on Web Communities in the Face of COVID-19.
CoRR, 2020

Everything About You: A Multimodal Approach towards Friendship Inference in Online Social Networks.
CoRR, 2020

NeuLP: An End-to-End Deep-Learning Model for Link Prediction.
Proceedings of the Web Information Systems Engineering - WISE 2020, 2020

Updates-Leak: Data Set Inference and Reconstruction Attacks in Online Learning.
Proceedings of the 29th USENIX Security Symposium, 2020

Towards Plausible Graph Anonymization.
Proceedings of the 27th Annual Network and Distributed System Security Symposium, 2020

Membership Inference Against DNA Methylation Databases.
Proceedings of the IEEE European Symposium on Security and Privacy, 2020

LEAF: A Faster Secure Search Algorithm via Localization, Extraction, and Reconstruction.
Proceedings of the CCS '20: 2020 ACM SIGSAC Conference on Computer and Communications Security, 2020

GAN-Leaks: A Taxonomy of Membership Inference Attacks against Generative Models.
Proceedings of the CCS '20: 2020 ACM SIGSAC Conference on Computer and Communications Security, 2020

2019
An active learning-based approach for location-aware acquaintance inference.
Knowl. Inf. Syst., 2019

GAN-Leaks: A Taxonomy of Membership Inference Attacks against GANs.
CoRR, 2019

Language in Our Time: An Empirical Analysis of Hashtags.
Proceedings of the World Wide Web Conference, 2019

A Graph-Based Approach to Explore Relationship Between Hashtags and Images.
Proceedings of the Web Information Systems Engineering - WISE 2019, 2019

ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models.
Proceedings of the 26th Annual Network and Distributed System Security Symposium, 2019

MBeacon: Privacy-Preserving Beacons for DNA Methylation Data.
Proceedings of the 26th Annual Network and Distributed System Security Symposium, 2019

Fairwalk: Towards Fair Graph Embedding.
Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, 2019

MemGuard: Defending against Black-Box Membership Inference Attacks via Adversarial Examples.
Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, 2019

Fast and Scalable VMM Live Upgrade in Large Cloud Infrastructure.
Proceedings of the Twenty-Fourth International Conference on Architectural Support for Programming Languages and Operating Systems, 2019

How to prove your model belongs to you: a blind-watermark based framework to protect intellectual property of DNN.
Proceedings of the 35th Annual Computer Security Applications Conference, 2019

2018
ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models.
CoRR, 2018

Tagvisor: A Privacy Advisor for Sharing Hashtags.
Proceedings of the 2018 World Wide Web Conference on World Wide Web, 2018

You Are Where You App: An Assessment on Location Privacy of Social Applications.
Proceedings of the 29th IEEE International Symposium on Software Reliability Engineering, 2018

Dissecting Privacy Risks in Biomedical Data.
Proceedings of the 2018 IEEE European Symposium on Security and Privacy, 2018

2017
CTRL+Z: Recovering Anonymized Social Graphs.
CoRR, 2017

Does #like4like indeed provoke more likes?
Proceedings of the International Conference on Web Intelligence, 2017

DeepCity: A Feature Learning Framework for Mining Location Check-Ins.
Proceedings of the Eleventh International Conference on Web and Social Media, 2017

Quantifying Location Sociality.
Proceedings of the 28th ACM Conference on Hypertext and Social Media, 2017

Semantic Annotation for Places in LBSN through Graph Embedding.
Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, 2017

walk2friends: Inferring Social Links from Mobility Profiles.
Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, 2017

2016
On Impact of Weather on Human Mobility in Cities.
Proceedings of the Web Information Systems Engineering - WISE 2016, 2016

An Empirical Study on User Access Control in Online Social Networks.
Proceedings of the 21st ACM on Symposium on Access Control Models and Technologies, 2016

Modeling City Locations as Complex Networks: An initial study.
Proceedings of the Complex Networks & Their Applications V - Proceedings of the 5th International Workshop on Complex Networks and their Applications (COMPLEX NETWORKS 2016), Milan, Italy, November 30, 2016

2015
A new access control scheme for Facebook-style social networks.
Comput. Secur., 2015

Exploring Communities for Effective Location Prediction.
Proceedings of the 24th International Conference on World Wide Web Companion, 2015

A Logical Approach to Restricting Access in Online Social Networks.
Proceedings of the 20th ACM Symposium on Access Control Models and Technologies, 2015

Community-Driven Social Influence Analysis and Applications.
Proceedings of the Engineering the Web in the Big Data Era - 15th International Conference, 2015

Location Prediction: Communities Speak Louder than Friends.
Proceedings of the 2015 ACM on Conference on Online Social Networks, 2015

Cryptographic Protocols for Enforcing Relationship-Based Access Control Policies.
Proceedings of the 39th IEEE Annual Computer Software and Applications Conference, 2015

Inferring Friendship from Check-in Data of Location-Based Social Networks.
Proceedings of the 2015 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, 2015

Distance and Friendship: A Distance-Based Model for Link Prediction in Social Networks.
Proceedings of the Web Technologies and Applications - 17th Asia-PacificWeb Conference, 2015

Event Prediction with Community Leaders.
Proceedings of the 10th International Conference on Availability, Reliability and Security, 2015

2013
Twisted edwards-form elliptic curve cryptography for 8-bit AVR-based sensor nodes.
Proceedings of the first ACM workshop on Asia public-key cryptography, 2013


  Loading...