Mentor

Brenda Liliana Aguinaga Serrano, Universidad de Guadalajara, Mexico

Publication Date

2024

Document Type

Paper

DOI

10.25776/a0vf-jj23

Pages

1-4 pp.

Abstract

With advancements in AI-driven natural language generation, distinguishing between AI-generated and human-written text has become imperative for ensuring content authenticity across industries. This study explores the effectiveness of Bidirectional Encoder Representations from Transformers (BERT) in addressing this classification challenge. Utilizing a diverse dataset and robust preprocessing techniques, BERT achieved a peak F1-score of 0.94364, outperforming traditional models such as Logistic Regression and Support Vector Machines. The results underscore the potential of transformer-based models in addressing real-world con- tent verification problems. Future enhancements include fine-tuning and expanding datasets for greater generalizability.

Share

COinS