IIIT-Delhi Institutional Repository

Robustness of foundational models for hate speech detection: analyzing pretraining dynamics and finetuning settings

Show simple item record

dc.contributor.author Khan, Mohammad Aflah
dc.contributor.author Akhtar, Md. Shad (Advisor)
dc.date.accessioned 2024-05-09T12:52:50Z
dc.date.available 2024-05-09T12:52:50Z
dc.date.issued 2023-11-29
dc.identifier.uri http://repository.iiitd.edu.in/xmlui/handle/123456789/1414
dc.description.abstract Despite the widespread adoption, there is a lack of research into how various critical aspects of LLMs affect their performance in hate speech detection. Through five research questions, our findings and recommendations lay the groundwork for empirically investigating different aspects of LLMs’ use in hate speech detection. We deep dive into comparing different pretrained models, evaluating their seed robustness, finetuning settings, and the impact of pretraining data collection time. Our analysis reveals early peaks for downstream tasks during pretraining, the limited benefit of employing a more recent pretraining corpus, and the significance of specific layers during finetuning. We further call into question the use of domain-specific models and highlight the lack of dynamic datasets for benchmarking hate speech detection. en_US
dc.language.iso en_US en_US
dc.publisher IIIT-Delhi en_US
dc.subject Hate Speech en_US
dc.subject Social Media en_US
dc.subject Implicit Hate Speech en_US
dc.subject Explicit Hate Speech en_US
dc.subject Dataset Balancing en_US
dc.subject Data Augmentation en_US
dc.subject Training Dynamics en_US
dc.title Robustness of foundational models for hate speech detection: analyzing pretraining dynamics and finetuning settings en_US
dc.type Other en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search Repository


Advanced Search

Browse

My Account