Large language models (LLMs) that power chatbots like ChatGPT may be getting better at answering benchmark questions that measure mathematical reasoning. But this may actually be a bad thing. This is when data resembling benchmark questions leaks into training data.
ORIGINAL LINK: https://www.tomsguide.com/ai/ai-models-are-getting-better-at-grade-school-math-but-a-new-study-suggests-they-may-be-cheating