Turning Logic Against Itself : Probing Model Defenses Through Contrastive Questions Paper • 2501.01872 • Published 22 days ago • 2
Navigating the Cultural Kaleidoscope: A Hitchhiker's Guide to Sensitivity in Large Language Models Paper • 2410.12880 • Published Oct 15, 2024 • 3
How (un)ethical are instruction-centric responses of LLMs? Unveiling the vulnerabilities of safety guardrails to harmful queries Paper • 2402.15302 • Published Feb 23, 2024 • 4
Safety Arithmetic: A Framework for Test-time Safety Alignment of Language Models by Steering Parameters and Activations Paper • 2406.11801 • Published Jun 17, 2024 • 16
Breaking Boundaries: Investigating the Effects of Model Editing on Cross-linguistic Performance Paper • 2406.11139 • Published Jun 17, 2024 • 13
Breaking Boundaries: Investigating the Effects of Model Editing on Cross-linguistic Performance Paper • 2406.11139 • Published Jun 17, 2024 • 13
Breaking Boundaries: Investigating the Effects of Model Editing on Cross-linguistic Performance Paper • 2406.11139 • Published Jun 17, 2024 • 13
SafeInfer: Context Adaptive Decoding Time Safety Alignment for Large Language Models Paper • 2406.12274 • Published Jun 18, 2024 • 15
Safety Arithmetic: A Framework for Test-time Safety Alignment of Language Models by Steering Parameters and Activations Paper • 2406.11801 • Published Jun 17, 2024 • 16
Safety Arithmetic: A Framework for Test-time Safety Alignment of Language Models by Steering Parameters and Activations Paper • 2406.11801 • Published Jun 17, 2024 • 16
SafeInfer: Context Adaptive Decoding Time Safety Alignment for Large Language Models Paper • 2406.12274 • Published Jun 18, 2024 • 15
SafeInfer: Context Adaptive Decoding Time Safety Alignment for Large Language Models Paper • 2406.12274 • Published Jun 18, 2024 • 15