Paperpal recently launched its own AI detector, promising a major step forward for scholarly work. Trained on over 100,000 academic texts – human and AI-generated (ChatGPT, Claude, Gemini) – it claims superior accuracy, a ]40% reduction in false positives through cross-verification, and a three-tier verdict: human, mixed or AI. Privacy is prioritised: texts are processed temporarily, multiply encrypted, never stored or reused for training.
The tool receives ongoing updates to match new AI models.To test it, I submitted one of my own pre-2022 academic papers – unquestionably human. Result: 50% flagged as AI patterns. My formal, structured style apparently seemed too polished for the algorithm, resembling modern LLM output.This exposes a core weakness shared by most detectors: they track statistical language patterns, not true authorship or ideas. Clear, concise academic writing – exactly what is taught – often gets penalised. Non-native speakers striving for clarity face the same risk.
Many academics report similar false positives when checking their own old work. Confidence in these tools is fading fast.On X and elsewhere, criticism is growing. One widely shared post captured the mood: brilliant students’ futures can be damaged on what amounts to a coin toss.
Paperpal stands out for its transparent interface, academic focus, responsible data handling and nuanced scoring rather than binary judgements.Still, like all current detectors, it analyses surface patterns, not intellectual substance. Universities and journals should therefore never base serious decisions solely on an AI score. Human review, conventional plagiarism checks and – where needed – oral defences must remain essential.
Clear disclosure policies for AI use would bring welcome fairness.Paperpal is among the better tools available, yet the fundamental limit persists. Until detectors can truly grasp authorship, they should inform discussion, not deliver verdicts.




























