Activation Approximations Can Incur Safety Vulnerabilities Even in Aligned LLMs: Comprehensive Analysis and Defense Paper • 2502.00840 • Published 12 days ago • 3
HarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusal Paper • 2402.04249 • Published Feb 6, 2024 • 4
Star Attention: Efficient LLM Inference over Long Sequences Paper • 2411.17116 • Published Nov 26, 2024 • 50
LookAhead: Preventing DeFi Attacks via Unveiling Adversarial Contracts Paper • 2401.07261 • Published Jan 14, 2024