mvivirito0/Pixabay

Artificial intelligence chatbot ChatGPT-4 appeared to think on its feet and make human-like errors when researchers challenged it with a 2,400-year-old mathematical problem from Plato’s teachings.

Cambridge University education researchers expected the AI system to simply recall the famous solution to Socrates’ “doubling the square” puzzle from its training data. Instead, ChatGPT seemed to develop its own approach and made distinctly student-like mistakes.

The experiment recreated Plato’s dialogue where Socrates teaches an uneducated boy to double a square’s area. Rather than doubling each side length, the correct solution involves creating a new square using the original’s diagonal.

Dr Nadav Marco from Hebrew University and Professor Andreas Stylianides from Cambridge University deliberately introduced errors and variants to test whether ChatGPT would retrieve pre-existing knowledge or generate novel solutions.

“When we face a new problem, our instinct is often to try things out based on our past experience,” said Marco. “In our experiment, ChatGPT seemed to do something similar. Like a learner or scholar, it appeared to come up with its own hypotheses and solutions.”

ChatGPT initially chose algebraic methods unknown in Plato’s era, resisting attempts to guide it towards geometric reasoning. Only when researchers expressed disappointment at its approach did the system provide the classical geometric solution.

The AI demonstrated complete knowledge of Plato’s work when asked directly about it, suggesting the improvised responses weren’t due to missing information.

“If it had only been recalling from memory, it would almost certainly have referenced the classical solution of building a new square on the original square’s diagonal straight away,” said Stylianides. “Instead, it seemed to take its own approach.”

When tested with variations involving rectangles and triangles, ChatGPT made a false claim that geometric solutions were impossible for rectangles, despite such methods existing.

The researchers propose that AI systems may have a “zone of proximal development” similar to human learners, where they cannot solve problems immediately but can develop solutions through guidance and prompting.

“Unlike proofs found in reputable textbooks, students cannot assume that Chat GPT’s proofs are valid,” said Stylianides, highlighting the need for critical evaluation skills in mathematics education.

Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like

Meta launches ad-free subscriptions after ICO forces compliance changes

Meta will offer UK users paid subscriptions to use Facebook and Instagram…

World nears quarter million crypto millionaires in historic wealth boom

Global cryptocurrency millionaires have reached 241,700 individuals, marking a 40 per cent…

Wong warns AI nuclear weapons threaten future of humanity at UN

Australia’s Foreign Minister Penny Wong has warned that artificial intelligence’s potential use…

Mistral targets enterprise data as public AI training resources dry up

Europe’s leading artificial intelligence startup Mistral AI is turning to proprietary enterprise…

Legal scholar warns AI could devalue humanity without urgent regulatory action

Artificial intelligence systems pose worldwide threats to human dignity by potentially reducing…

MIT accelerator shows AI enhances startup building without replacing core principles

Entrepreneurs participating in MIT’s flagship summer programme are integrating artificial intelligence tools…

AI creates living viruses for first time as scientists make artificial “life”

Stanford University researchers have achieved a scientific milestone by creating the world’s…

Engineers create smarter artificial intelligence for power grids and autonomous vehicles

Researchers have developed an artificial intelligence system that manages complex networks where…