< Explain other AI papers

Towards Cross-Lingual Audio Abuse Detection in Low-Resource Settings with Few-Shot Learning

Aditya Narayan Sankaran, Reza Farahbaksh, Noel Crespi

2024-12-03

Towards Cross-Lingual Audio Abuse Detection in Low-Resource Settings with Few-Shot Learning

Summary

This paper explores how to detect abusive language in audio recordings of low-resource languages using a method called Few-Shot Learning (FSL), which requires only a small amount of training data.

What's the problem?

Detecting abusive content in audio is challenging, especially in languages that don't have a lot of available data or resources. Most existing systems struggle to identify abusive language accurately because they are trained on limited examples, making it hard for them to recognize different ways people might express abuse in various languages.

What's the solution?

The researchers propose a new approach that uses pre-trained audio models like Wav2Vec and Whisper to help detect abusive language across multiple languages. They employ a technique called Model-Agnostic Meta-Learning (MAML) within the Few-Shot Learning framework, allowing the system to learn from just a few examples of abusive speech. By training the model on a dataset called ADIMA, they show that it can effectively generalize and identify abusive language even when there is little data available for each specific language.

Why it matters?

This research is significant because it helps improve the ability to detect harmful speech in languages that are often overlooked. By using advanced machine learning techniques, the study aims to make online spaces safer for speakers of low-resource languages, ensuring that everyone can communicate without facing abuse or harassment.

Abstract

Online abusive content detection, particularly in low-resource settings and within the audio modality, remains underexplored. We investigate the potential of pre-trained audio representations for detecting abusive language in low-resource languages, in this case, in Indian languages using Few Shot Learning (FSL). Leveraging powerful representations from models such as Wav2Vec and Whisper, we explore cross-lingual abuse detection using the ADIMA dataset with FSL. Our approach integrates these representations within the Model-Agnostic Meta-Learning (MAML) framework to classify abusive language in 10 languages. We experiment with various shot sizes (50-200) evaluating the impact of limited data on performance. Additionally, a feature visualization study was conducted to better understand model behaviour. This study highlights the generalization ability of pre-trained models in low-resource scenarios and offers valuable insights into detecting abusive language in multilingual contexts.