Maksym Andriushchenko
9
Papers
905
Total Citations
Papers (9)
Formal Guarantees on the Robustness of a Classifier against Adversarial Manipulation
NeurIPS 2017arXiv
530
citations
Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks
ICLR 2025
375
citations
Why ReLU Networks Yield High-Confidence Predictions Far Away From the Training Data and How to Mitigate the Problem
CVPR 2019
0
citations
Square Attack: a query-efficient black-box adversarial attack via random search
ECCV 2020
0
citations
Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning
ICML 2024
0
citations
Provably robust boosted decision stumps and trees against adversarial attacks
NeurIPS 2019
0
citations
Understanding and Improving Fast Adversarial Training
NeurIPS 2020
0
citations
Transferable Adversarial Robustness for Categorical Data via Universal Robust Embeddings
NeurIPS 2023
0
citations
Sharpness-Aware Minimization Leads to Low-Rank Features
NeurIPS 2023
0
citations