![](https://magicitmiracles.com/wp-content/uploads/2025/01/3806840-0-83684600-1737484143-cio_middle_east_egypt_sphinx_and_pyramids_at_sunset_in_giza_by_waleed_hammoudeh_gettyimages-509806156_2400x1600-100799879-orig.jpg)
Today’s AI models do a poor job of providing accurate information about world history, according to a new report from the Austrian research institute Complexity Science Hub (CSH).
In an experiment, OpenAI’s GPT-4, Meta’s Llama, and Google’s Gemini were asked to answer yes or no to historical questions — and only 46% of the answers were correct. GPT-4, for example, answered “yes” to the question of whether Ancient Egypt had a standing army, likely because the AI model chose to extrapolate data from other empires such as Persia.
“If you are told A and B 100 times and C one time, and then asked a question about C, you might just remember A and B and try to extrapolate from that,” researcher Maria del Rio-Chanona told Techcrunch.