Goldstein, Tom
161 publications
NeurIPS
2025
A Technical Report on “Erasing the Invisible”: The 2024 NeurIPS Competition on Stress Testing Image Watermarks
AAAI
2025
Can Watermarking Large Language Models Prevent Copyrighted Text Generation and Hide Training Data?
NeurIPS
2025
FineGRAIN: Evaluating Failure Modes of Text-to-Image Models with Vision Language Model Judges
ICMLW
2024
Can Watermarking Large Language Models Prevent Copyrighted Text Generation and Hide Training Data?
NeurIPSW
2024
Can Watermarking Large Language Models Prevent Copyrighted Text Generation and Hide Training Data?
NeurIPS
2024
Easy2Hard-Bench: Standardized Difficulty Labels for Profiling LLM Performance and Generalization
NeurIPS
2023
Battle of the Backbones: A Large-Scale Comparison of Pretrained Models Across Computer Vision Tasks
ICLR
2023
Decepticons: Corrupted Transformers Breach Privacy in Federated Learning for Language Models
NeurIPS
2023
Hard Prompts Made Easy: Gradient-Based Discrete Optimization for Prompt Tuning and Discovery
NeurIPSW
2022
DP-InstaHide: Data Augmentations Provably Enhance Guarantees Against Dataset Manipulations
NeurIPSW
2022
Decepticons: Corrupted Transformers Breach Privacy in Federated Learning for Language Models
NeurIPS
2022
End-to-End Algorithm Synthesis with Recurrent Networks: Extrapolation Without Overthinking
NeurIPSW
2022
Panning for Gold in Federated Learning: Targeted Text Extraction Under Arbitrarily Large-Scale Aggregation
ICLR
2022
Robbing the Fed: Directly Obtaining Private Data in Federated Learning with Modified Models
NeurIPSW
2022
SAINT: Improved Neural Networks for Tabular Data via Row Attention and Contrastive Pre-Training
NeurIPS
2022
Sleeper Agent: Scalable Hidden Trigger Backdoors for Neural Networks Trained from Scratch
NeurIPS
2021
Can You Learn an Algorithm? Generalizing from Easy to Hard Problems with Recurrent Networks
NeurIPS
2021
Gradient-Free Adversarial Training Against Image Corruption for Learning-Based Steering
ICML
2021
Just How Toxic Is Data Poisoning? a Unified Benchmark for Backdoor and Data Poisoning Attacks
ICLR
2021
LowKey: Leveraging Adversarial Attacks to Protect Social Media Users from Facial Recognition
NeurIPS
2021
VQ-GNN: A Universal Framework to Scale up Graph Neural Networks Using Vector Quantization