Unmasking the Imposters: How Censorship and Domain Adaptation Affect the Detection of Machine-Generated Tweets
Bryan E. Tuck, Rakesh M. Verma
TL;DR
This work examines how censorship and domain adaptation shape the detectability of machine-generated tweets. By fine-tuning nine model variants (censored and uncensored) from four LLM families on in-domain Twitter data and generating emotion-preserving synthetic tweets, the authors create nine paired datasets for robust detector evaluation using BERTweet, DeBERTaV3, and a soft-ensemble approach, supplemented by stylometric features. The results show that uncensored models achieve greater lexical richness and human-like structure, but at the cost of higher toxicity and substantially weakened detection performance, while censored models remain more detectable but less linguistically diverse. The findings underscore the need for advanced, domain-aware moderation and detectors that can adapt to rapidly evolving generative capabilities, with careful attention to ethical implications and platform-specific dynamics.
Abstract
The rapid development of large language models (LLMs) has significantly improved the generation of fluent and convincing text, raising concerns about their potential misuse on social media platforms. We present a comprehensive methodology for creating nine Twitter datasets to examine the generative capabilities of four prominent LLMs: Llama 3, Mistral, Qwen2, and GPT4o. These datasets encompass four censored and five uncensored model configurations, including 7B and 8B parameter base-instruction models of the three open-source LLMs. Additionally, we perform a data quality analysis to assess the characteristics of textual outputs from human, "censored," and "uncensored" models, employing semantic meaning, lexical richness, structural patterns, content characteristics, and detector performance metrics to identify differences and similarities. Our evaluation demonstrates that "uncensored" models significantly undermine the effectiveness of automated detection methods. This study addresses a critical gap by exploring smaller open-source models and the ramifications of "uncensoring," providing valuable insights into how domain adaptation and content moderation strategies influence both the detectability and structural characteristics of machine-generated text.
