Mentor
Brenda Liliana Aguinaga Serrano, Universidad de Guadalajara, Mexico
Publication Date
2024
Document Type
Paper
DOI
10.25776/a0vf-jj23
Pages
1-4 pp.
Abstract
With advancements in AI-driven natural language generation, distinguishing between AI-generated and human-written text has become imperative for ensuring content authenticity across industries. This study explores the effectiveness of Bidirectional Encoder Representations from Transformers (BERT) in addressing this classification challenge. Utilizing a diverse dataset and robust preprocessing techniques, BERT achieved a peak F1-score of 0.94364, outperforming traditional models such as Logistic Regression and Support Vector Machines. The results underscore the potential of transformer-based models in addressing real-world con- tent verification problems. Future enhancements include fine-tuning and expanding datasets for greater generalizability.
Repository Citation
Katlariwala, Soham Biren, "BERT-Based Detection of AI-Generated Text for Content Verification" (2024). 2024 REYES Proceedings. 14.
https://digitalcommons.odu.edu/reyes-2024/14
Included in
Artificial Intelligence and Robotics Commons, Computer Engineering Commons, Programming Languages and Compilers Commons