Ayan Datta


2024

pdf bib
Weighted Layer Averaging RoBERTa for Black-Box Machine-Generated Text Detection
Ayan Datta | Aryan Chandramania | Radhika Mamidi
Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024)

We propose a novel approach for machine-generated text detection using a RoBERTa model with weighted layer averaging and AdaLoRA for parameter-efficient fine-tuning. Our method incorporates information from all model layers, capturing diverse linguistic cues beyond those accessible from the final layer alone. To mitigate potential overfitting and improve generalizability, we leverage AdaLoRA, which injects trainable low-rank matrices into each Transformer layer, significantly reducing the number of trainable parameters. Furthermore, we employ data mixing to ensure our model encounters text from various domains and generators during training, enhancing its ability to generalize to unseen data. This work highlights the potential of combining layer-wise information with parameter-efficient fine-tuning and data mixing for effective machine-generated text detection.