Niloofar Mireshghallah

Orcid: 0000-0003-4090-9756

According to our database1, Niloofar Mireshghallah authored at least 52 papers between 2019 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Differentially Private Learning Needs Better Model Initialization and Self-Distillation.
CoRR, 2024

AI as Humanity's Salieri: Quantifying Linguistic Creativity of Language Models via Systematic Attribution of Machine Text against Web Text.
CoRR, 2024

HAICOSYSTEM: An Ecosystem for Sandboxing Safety Risks in Human-AI Interactions.
CoRR, 2024

Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild.
CoRR, 2024

WildTeaming at Scale: From In-the-Wild Jailbreaks to (Adversarially) Safer Language Models.
CoRR, 2024

Breaking News: Case Studies of Generative AI's Use in Journalism.
CoRR, 2024

Alpaca against Vicuna: Using LLMs to Uncover Memorization of LLMs.
CoRR, 2024

Do Membership Inference Attacks Work on Large Language Models?
CoRR, 2024

A Roadmap to Pluralistic Alignment.
CoRR, 2024

LatticeGen: Hiding Generated Text in a Lattice for Privacy-Aware Large Language Model Generation on Cloud.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2024, 2024

Position: A Roadmap to Pluralistic Alignment.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Privacy-Preserving In-Context Learning with Differentially Private Few-Shot Generation.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

CopyBench: Measuring Literal and Non-Literal Reproduction of Copyright-Protected Text in Language Model Generation.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Smaller Language Models are Better Zero-shot Machine-Generated Text Detectors.
Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics, 2024

2023
Auditing and Mitigating Safety Risks in Large Language Models
PhD thesis, 2023

Report of the 1st Workshop on Generative AI and Law.
CoRR, 2023

Misusing Tools in Large Language Models With Visual Adversarial Examples.
CoRR, 2023

LatticeGen: A Cooperative Framework which Hides Generated Text in a Lattice for Privacy-Aware Generation on Cloud.
CoRR, 2023

Are Chatbots Ready for Privacy-Sensitive Applications? An Investigation into Input Regurgitation and Prompt-Induced Sanitization.
CoRR, 2023

Smaller Language Models are Better Black-box Machine-Generated Text Detectors.
CoRR, 2023

Simple Temporal Adaptation to Changing Label Sets: Hashtag Prediction via Dense KNN.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

A Block Metropolis-Hastings Sampler for Controllable Energy-based Text Generation.
Proceedings of the 27th Conference on Computational Natural Language Learning, 2023

Privacy-Preserving Domain Adaptation of Semantic Parsers.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

Membership Inference Attacks against Language Models via Neighbourhood Comparison.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, 2023

2022
Non-Parametric Temporal Adaptation for Social Media Topic Classification.
CoRR, 2022

Memorization in NLP Fine-tuning Methods.
CoRR, 2022

Mix and Match: Learning-free Controllable Text Generation using Energy Language Models.
CoRR, 2022

Differentially Private Model Compression.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

UserIdentifier: Implicit User Representations for Simple and Effective Personalized Sentiment Analysis.
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2022

What Does it Mean for a Language Model to Preserve Privacy?
Proceedings of the FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul, Republic of Korea, June 21, 2022

An Empirical Analysis of Memorization in Fine-tuned Autoregressive Language Models.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

Quantifying Privacy Risks of Masked Language Models Using Membership Inference Attacks.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

Mix and Match: Learning-free Controllable Text Generationusing Energy Language Models.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022

2021
Efficient Hyperparameter Optimization for Differentially Private Deep Learning.
CoRR, 2021

Benchmarking Differential Privacy and Federated Learning for BERT Models.
CoRR, 2021

When Differential Privacy Meets Interpretability: A Case Study.
CoRR, 2021

DP-SGD vs PATE: Which Has Less Disparate Impact on Model Accuracy?
CoRR, 2021

Privacy Regularization: Joint Privacy-Utility Optimization in Language Models.
CoRR, 2021

Not All Features Are Equal: Discovering Essential Features for Preserving Prediction Privacy.
Proceedings of the WWW '21: The Web Conference 2021, 2021

Privacy Regularization: Joint Privacy-Utility Optimization in LanguageModels.
Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021

U-Noise: Learnable Noise Masks for Interpretable Image Segmentation.
Proceedings of the 2021 IEEE International Conference on Image Processing, 2021

Style Pooling: Automatic Text Style Obfuscation for Improved Classification Fairness.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021

2020
ReLeQ : A Reinforcement Learning Approach for Automatic Deep Quantization of Neural Networks.
IEEE Micro, 2020

Privacy in Deep Learning: A Survey.
CoRR, 2020

A Principled Approach to Learning Stochastic Representations for Privacy in Deep Neural Inference.
CoRR, 2020

Gradient-Based Deep Quantization of Neural Networks through Sinusoidal Adaptive Regularization.
CoRR, 2020

Divide and Conquer: Leveraging Intermediate Feature Representations for Quantized Training of Neural Networks.
Proceedings of the 37th International Conference on Machine Learning, 2020

Neither Private Nor Fair: Impact of Data Imbalance on Utility and Fairness in Differential Privacy.
Proceedings of the PPMLP'20: Proceedings of the 2020 Workshop on Privacy-Preserving Machine Learning in Practice, 2020

Shredder: Learning Noise Distributions to Protect Inference Privacy.
Proceedings of the ASPLOS '20: Architectural Support for Programming Languages and Operating Systems, 2020

2019
Energy-Efficient Permanent Fault Tolerance in Hard Real-Time Systems.
IEEE Trans. Computers, 2019

Shredder: Learning Noise to Protect Privacy with Partial DNN Inference on the Edge.
CoRR, 2019


  Loading...