Florian Tramèr

Orcid: 0000-0001-8703-8762

Affiliations:
  • ETH Zurich, Switzerland


According to our database1, Florian Tramèr authored at least 95 papers between 2015 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Persistent Pre-Training Poisoning of LLMs.
CoRR, 2024

Gradient-based Jailbreak Images for Multimodal Fusion Models.
CoRR, 2024

Membership Inference Attacks Cannot Prove that a Model Was Trained On Your Data.
CoRR, 2024

Adversarial Search Engine Optimization for Large Language Models.
CoRR, 2024

Blind Baselines Beat Membership Inference Attacks for Foundation Models.
CoRR, 2024

AgentDojo: A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.
CoRR, 2024

Adversarial Perturbations Cannot Reliably Protect Artists From Generative AI.
CoRR, 2024

Dataset and Lessons Learned from the 2024 SaTML LLM Capture-the-Flag Competition.
CoRR, 2024

Evaluations of Machine Learning Privacy Defenses are Misleading.
CoRR, 2024

Competition Report: Finding Universal Jailbreak Backdoors in Aligned LLMs.
CoRR, 2024

Foundational Challenges in Assuring Alignment and Safety of Large Language Models.
CoRR, 2024

JailbreakBench: An Open Robustness Benchmark for Jailbreaking Large Language Models.
CoRR, 2024

Query-Based Adversarial Prompt Generation.
CoRR, 2024

Privacy Side Channels in Machine Learning Systems.
Proceedings of the 33rd USENIX Security Symposium, 2024

Poisoning Web-Scale Training Datasets is Practical.
Proceedings of the IEEE Symposium on Security and Privacy, 2024

Evaluating Superhuman Models with Consistency Checks.
Proceedings of the IEEE Conference on Secure and Trustworthy Machine Learning, 2024

Evading Black-box Classifiers Without Breaking Eggs.
Proceedings of the IEEE Conference on Secure and Trustworthy Machine Learning, 2024

Position: Considerations for Differentially Private Learning with Large-Scale Public Pretraining.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Extracting Training Data From Document-Based VQA Models.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Privacy Backdoors: Stealing Data with Corrupted Pretrained Models.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Stealing part of a production language model.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Universal Jailbreak Backdoors from Poisoned Human Feedback.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

2023
Scalable Extraction of Training Data from (Production) Language Models.
CoRR, 2023

Backdoor Attacks for In-Context Learning with Language Models.
CoRR, 2023

Are aligned neural networks adversarially aligned?
CoRR, 2023

Randomness in ML Defenses Helps Persistent Attackers and Hinders Evaluators.
CoRR, 2023

Tight Auditing of Differentially Private Machine Learning.
Proceedings of the 32nd USENIX Security Symposium, 2023

Extracting Training Data from Diffusion Models.
Proceedings of the 32nd USENIX Security Symposium, 2023

SNAP: Efficient Extraction of Private Properties with Poisoning.
Proceedings of the 44th IEEE Symposium on Security and Privacy, 2023

Counterfactual Memorization in Neural Language Models.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Students Parrot Their Teachers: Membership Inference on Model Distillation.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Are aligned neural networks adversarially aligned?
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Preventing Generation of Verbatim Memorization in Language Models Gives a False Sense of Privacy.
Proceedings of the 16th International Natural Language Generation Conference, 2023

Preprocessors Matter! Realistic Decision-Based Attacks on Machine Learning Systems.
Proceedings of the International Conference on Machine Learning, 2023

Measuring Forgetting of Memorized Training Examples.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

(Certified!!) Adversarial Robustness for Free!
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Quantifying Memorization Across Neural Language Models.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

AISec '23: 16th ACM Workshop on Artificial Intelligence and Security.
Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, 2023

2022
Considerations for Differentially Private Learning with Large-Scale Public Pretraining.
CoRR, 2022

Preventing Verbatim Memorization in Language Models Gives a False Sense of Privacy.
CoRR, 2022

Red-Teaming the Stable Diffusion Safety Filter.
CoRR, 2022

(Certified!!) Adversarial Robustness for Free!
CoRR, 2022

Debugging Differential Privacy: A Case Study for Privacy Auditing.
CoRR, 2022

Membership Inference Attacks From First Principles.
Proceedings of the 43rd IEEE Symposium on Security and Privacy, 2022

Increasing Confidence in Adversarial Robustness Evaluations.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

The Privacy Onion Effect: Memorization is Relative.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Detecting Adversarial Examples Is (Nearly) As Hard As Classifying Them.
Proceedings of the International Conference on Machine Learning, 2022

Data Poisoning Won't Save You From Facial Recognition.
Proceedings of the Tenth International Conference on Learning Representations, 2022

Large Language Models Can Be Strong Differentially Private Learners.
Proceedings of the Tenth International Conference on Learning Representations, 2022

What Does it Mean for a Language Model to Preserve Privacy?
Proceedings of the FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul, Republic of Korea, June 21, 2022

Truth Serum: Poisoning Machine Learning Models to Reveal Their Secrets.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

AISec '22: 15th ACM Workshop on Artificial Intelligence and Security.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

2021
Measuring and enhancing the security of machine learning.
PhD thesis, 2021

Advances and Open Problems in Federated Learning.
Found. Trends Mach. Learn., 2021

NeuraCrypt is not private.
CoRR, 2021

Data Poisoning Won't Save You From Facial Recognition.
CoRR, 2021

Extracting Training Data from Large Language Models.
Proceedings of the 30th USENIX Security Symposium, 2021

Is Private Learning Possible with Instance Encoding?
Proceedings of the 42nd IEEE Symposium on Security and Privacy, 2021

Antipodes of Label Differential Privacy: PATE and ALIBI.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

SquirRL: Automating Attack Analysis on Blockchain Incentive Mechanisms with Deep Reinforcement Learning.
Proceedings of the 28th Annual Network and Distributed System Security Symposium, 2021

Label-Only Membership Inference Attacks.
Proceedings of the 38th International Conference on Machine Learning, 2021

Differentially Private Learning Needs Better Features (or Much More Data).
Proceedings of the 9th International Conference on Learning Representations, 2021

Fourth International Workshop on Dependable and Secure Machine Learning - DSML 2021.
Proceedings of the 51st Annual IEEE/IFIP International Conference on Dependable Systems and Networks Workshops, 2021

2020
Remote Side-Channel Attacks on Anonymous Transactions.
IACR Cryptol. ePrint Arch., 2020

An Attack on InstaHide: Is Private Learning Possible with Instance Encoding?
CoRR, 2020

SentiNet: Detecting Localized Universal Attacks Against Deep Learning Systems.
Proceedings of the 2020 IEEE Security and Privacy Workshops, 2020

On Adaptive Attacks to Adversarial Example Defenses.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Fundamental Tradeoffs between Invariance and Sensitivity to Adversarial Perturbations.
Proceedings of the 37th International Conference on Machine Learning, 2020

Third International Workshop on Dependable and Secure Machine Learning - DSML 2020.
Proceedings of the 50th Annual IEEE/IFIP International Conference on Dependable Systems and Networks Workshops, 2020

2019
The Hydra Framework for Principled, Automated Bug Bounties.
IEEE Secur. Priv., 2019

Advances and Open Problems in Federated Learning.
CoRR, 2019

SquirRL: Automating Attack Discovery on Blockchain Incentive Mechanisms with Deep Reinforcement Learning.
CoRR, 2019

Exploiting Excessive Invariance caused by Norm-Bounded Adversarial Robustness.
CoRR, 2019

Adversarial Training and Robustness for Multiple Perturbations.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Slalom: Fast, Verifiable and Private Execution of Neural Networks in Trusted Hardware.
Proceedings of the 7th International Conference on Learning Representations, 2019

AdVersarial: Perceptual Ad Blocking meets Adversarial Machine Learning.
Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, 2019

2018
SentiNet: Detecting Physical Attacks Against Deep Learning Systems.
CoRR, 2018

Ad-versarial: Defeating Perceptual Ad-Blocking.
CoRR, 2018

Physical Adversarial Examples for Object Detectors.
Proceedings of the 12th USENIX Workshop on Offensive Technologies, 2018

Ensemble Adversarial Training: Attacks and Defenses.
Proceedings of the 6th International Conference on Learning Representations, 2018

2017
PrivateRide: A Privacy-Enhanced Ride-Hailing Service.
Proc. Priv. Enhancing Technol., 2017

Addressing Beacon re-identification attacks: quantification and mitigation of privacy risks.
J. Am. Medical Informatics Assoc., 2017

Enter the Hydra: Towards Principled Bug Bounties and Exploit-Resistant Smart Contracts.
IACR Cryptol. ePrint Arch., 2017

Note on Attacking Object Detectors with Adversarial Stickers.
CoRR, 2017

The Space of Transferable Adversarial Examples.
CoRR, 2017

Ensemble Adversarial Training: Attacks and Defenses.
CoRR, 2017

FairTest: Discovering Unwarranted Associations in Data-Driven Applications.
Proceedings of the 2017 IEEE European Symposium on Security and Privacy, 2017

2016
Sealed-Glass Proofs: Using Transparent Enclaves to Prove and Sell Knowledge.
IACR Cryptol. ePrint Arch., 2016

Formal Abstractions for Attested Execution Secure Processors.
IACR Cryptol. ePrint Arch., 2016

On solving L P N using B K W and variants - Implementation and analysis.
Cryptogr. Commun., 2016

Stealing Machine Learning Models via Prediction APIs.
Proceedings of the 25th USENIX Security Symposium, 2016

2015
Better Algorithms for LWE and LWR.
IACR Cryptol. ePrint Arch., 2015

On Solving Lpn using BKW and Variants.
IACR Cryptol. ePrint Arch., 2015

Discovering Unwarranted Associations in Data-Driven Applications with the FairTest Testing Toolkit.
CoRR, 2015

Differential Privacy with Bounded Priors: Reconciling Utility and Privacy in Genome-Wide Association Studies.
Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security, 2015


  Loading...