Search Results for author: Agastya Seth

Found 1 papers, 0 papers with code

The Art of Defending: A Systematic Evaluation and Analysis of LLM Defense Strategies on Safety and Over-Defensiveness

no code implementations30 Dec 2023 Neeraj Varshney, Pavel Dolin, Agastya Seth, Chitta Baral

As Large Language Models (LLMs) play an increasingly pivotal role in natural language processing applications, their safety concerns become critical areas of NLP research.

Cannot find the paper you are looking for? You can Submit a new open access paper.