Maksym Andriushchenko
8
Papers
375
Total Citations
Papers (8)
Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks
ICLR 2025arXiv
375
citations
OS-Harm: A Benchmark for Measuring Safety of Computer Use Agents
NeurIPS 2025arXiv
0
citations
Is In-Context Learning Sufficient for Instruction Following in LLMs?
ICLR 2025arXiv
0
citations
Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning
ICML 2024
0
citations
Square Attack: a query-efficient black-box adversarial attack via random search
ECCV 2020
0
citations
Understanding and Improving Fast Adversarial Training
NeurIPS 2020arXiv
0
citations
Transferable Adversarial Robustness for Categorical Data via Universal Robust Embeddings
NeurIPS 2023arXiv
0
citations
Sharpness-Aware Minimization Leads to Low-Rank Features
NeurIPS 2023arXiv
0
citations