Lukas Struppek

Orcid: 0000-0003-0626-3672

According to our database1, Lukas Struppek authored at least 27 papers between 2021 and 2024.

Collaborative distances:
  • Dijkstra number2 of five.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
Does CLIP Know My Face?
J. Artif. Intell. Res., 2024

CollaFuse: Collaborative Diffusion Models.
CoRR, 2024

Finding NeMo: Localizing Neurons Responsible For Memorization in Diffusion Models.
CoRR, 2024

Exploring the Adversarial Capabilities of Large Language Models.
CoRR, 2024

Exploiting Cultural Biases via Homoglyphs inText-to-Image Synthesis (Abstract Reprint).
Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence, 2024

Be Careful What You Smooth For: Label Smoothing Can Be a Privacy Shield but Also a Catalyst for Model Inversion Attacks.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

Collafuse: Navigating Limited Resources and Privacy in Collaborative Generative AI.
Proceedings of the 32nd European Conference on Information Systems, 2024

Defending Our Privacy with Backdoors.
Proceedings of the ECAI 2024 - 27th European Conference on Artificial Intelligence, 19-24 October 2024, Santiago de Compostela, Spain, 2024

2023
Exploiting Cultural Biases via Homoglyphs in Text-to-Image Synthesis.
J. Artif. Intell. Res., 2023

Leveraging Diffusion-Based Image Variations for Robust Training on Poisoned Data.
CoRR, 2023

Balancing Transparency and Risk: The Security and Privacy Risks of Open-Source Machine Learning Models.
CoRR, 2023

Image Classifiers Leak Sensitive Attributes About Their Classes.
CoRR, 2023

Fair Diffusion: Instructing Text-to-Image Generation Models on Fairness.
CoRR, 2023

SEGA: Instructing Diffusion using Semantic Dimensions.
CoRR, 2023

Balancing Transparency and Risk: An Overview of the Security and Privacy Risks of Open-Source Machine Learning Models.
Proceedings of the Bridging the Gap Between AI and Reality, 2023

SEGA: Instructing Text-to-Image Models using Semantic Guidance.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Sparsely-gated Mixture-of-Expert Layers for CNN Interpretability.
Proceedings of the International Joint Conference on Neural Networks, 2023

Rickrolling the Artist: Injecting Backdoors into Text Encoders for Text-to-Image Synthesis.
Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023

2022
Rickrolling the Artist: Injecting Invisible Backdoors into Text-Guided Image Generation Models.
CoRR, 2022

The Biased Artist: Exploiting Cultural Biases via Homoglyphs in Text-Guided Image Generation Models.
CoRR, 2022

CLIPping Privacy: Identity Inference Attacks on Multi-Modal Machine Learning Models.
CoRR, 2022

Transformer-Boosted Anomaly Detection with Fuzzy Hashes.
CoRR, 2022

Balancing Expert Utilization in Mixture-of-Experts Layers Embedded in CNNs.
CoRR, 2022

To Trust or Not To Trust Prediction Scores for Membership Inference Attacks.
Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, 2022

Plug & Play Attacks: Towards Robust and Flexible Model Inversion Attacks.
Proceedings of the International Conference on Machine Learning, 2022

Learning to Break Deep Perceptual Hashing: The Use Case NeuralHash.
Proceedings of the FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul, Republic of Korea, June 21, 2022

2021
Do Not Trust Prediction Scores for Membership Inference Attacks.
CoRR, 2021


  Loading...