Have you ever looked at the impressive results that LLMs get on benchmarks or difficult exams and wondered if these results are everything they seem? Some of these are due to something called data leakage, where assessments leak into the training data and the models memorise the answers.
If you'd like to learn more about how this classic pitfall of machine learning calls many of the results we see on LLM performance into question, check out my latest blog post.
https://t-redactyl.io/posts/2025-12-30-data-leakage-llm-measurement/
