Location

https://www.kennesaw.edu/ccse/events/computing-showcase/sp26-cday-program.php

Document Type

Event

Start Date

22-4-2026 4:00 PM

Description

This project evaluates automated SMS spam classification by comparing traditional machine learning against modern transformer architectures. We built a Bidirectional LSTM (BiLSTM) baseline using TF-IDF feature extraction and NearMiss-1 undersampling to handle severe class imbalances. We then compared this against a fine-tuned Hugging Face Sentence-BERT model. Preliminary results show Sentence-BERT significantly outperformed the BiLSTM baseline (99.01% vs. 95.65% accuracy). These findings demonstrate that transformer-based embeddings offer a highly accurate, scalable solution for spam mitigation without relying on aggressive data undersampling.

Share

COinS
 
Apr 22nd, 4:00 PM

GC-172-139 Detection of SMS Spam using Transformer BERT Model

https://www.kennesaw.edu/ccse/events/computing-showcase/sp26-cday-program.php

This project evaluates automated SMS spam classification by comparing traditional machine learning against modern transformer architectures. We built a Bidirectional LSTM (BiLSTM) baseline using TF-IDF feature extraction and NearMiss-1 undersampling to handle severe class imbalances. We then compared this against a fine-tuned Hugging Face Sentence-BERT model. Preliminary results show Sentence-BERT significantly outperformed the BiLSTM baseline (99.01% vs. 95.65% accuracy). These findings demonstrate that transformer-based embeddings offer a highly accurate, scalable solution for spam mitigation without relying on aggressive data undersampling.