Ancient stone ruins emerging from deep darkness, illuminated by precise digital scanning light in amber and blue tones, inscriptions and architectural details revealed in high contrast against obsidian shadow

History has always been an incomplete document. Not because the past didn't leave records, but because most of those records were destroyed, buried, fragmented, or burned. For two thousand years, hundreds of papyrus scrolls survived the eruption of Mount Vesuvius only to become unreadable, carbonized into fragile cylinders that crumbled when touched. For decades, archaeologists mapped the Maya world one painstaking excavation at a time, unaware that entire metropolitan networks lay concealed beneath the jungle canopy. For generations, classicists worked with inscriptions so damaged they could only guess at the missing words.

The constraints were not intellectual. They were physical. The past existed. We simply had no instrument capable of reading it.

That instrument now exists. It is not a telescope or a drill. It is machine learning. In the last four years, it has produced a series of breakthroughs that have changed, in permanent ways, what it means to study the ancient world.

Key Points

  • In 2024, three students used X-ray scanning and machine learning to read, for the first time, text inside a carbonized Herculaneum scroll that had been sealed since 79 AD. The content: a treatise on music and sensory pleasure by Philodemus.
  • LiDAR surveys combined with AI image analysis have identified over 60,000 previously unknown Maya structures in the Guatemalan lowlands alone, revealing an urban network denser than pre-industrial Europe.
  • DeepMind's Ithaca model, trained on 78,000 ancient Greek inscriptions, restores damaged or missing text with 62% accuracy, outperforming human experts working without AI at 43%.
  • Two independent computational models published in Nature (2025–2026) have validated a construction timeline of 20 to 27 years for the Great Pyramid of Giza using finite element analysis and logistics simulation.
  • The common mechanism: AI does not guess. It identifies statistically consistent patterns across a body of evidence too large for any human scholar to hold simultaneously.

The Scrolls Nobody Could Open

In August of 79 AD, the eruption of Mount Vesuvius buried the city of Herculaneum under fifteen meters of volcanic material. In the Villa dei Papiri, a private library of roughly 1,800 papyrus scrolls was carbonized in the heat. The scrolls survived. They are the only intact classical library to come down to us from antiquity. For nearly two thousand years, however, they could not be read. Every attempt to physically unroll them ended in destruction.

The Vesuvius Challenge, launched in 2023, offered a prize of $700,000 to any team that could decode at least four passages from sealed scrolls using non-invasive methods alone. The mechanism: high-resolution X-ray tomography produces a three-dimensional scan of the scroll interior; machine learning models then detect the minute differences in density caused by ancient ink on carbonized papyrus and render the text visible.

In February 2024, three university students (one from the University of Nebraska, one from the University of Kentucky, one from ETH Zürich) claimed the prize. Working independently and then collaborating, they trained a model that identified and assembled readable Greek text from inside a scroll that had not been opened since the first century AD.

The text belonged to Philodemus, an Epicurean philosopher. The subject: music, sensory perception, and the experience of pleasure. The scroll was a philosophical argument about how sound produces feeling. It had been waiting, carbonized and sealed, for 1,946 years.

THE VESUVIUS CHALLENGE

Over 800 papyrus scrolls from Herculaneum remain unread. The winning model identified text by detecting variations in surface texture at the micron level, differences invisible to the human eye but consistent enough for a trained neural network to distinguish ink from ash. The University of Oxford confirmed the findings in a peer-reviewed verification.

60,000 Structures Hidden in Plain Sight

LiDAR (Light Detection and Ranging) fires millions of laser pulses per second from an aircraft, measures the return time of each, and builds a three-dimensional map of the terrain below. In dense jungle, where conventional aerial photography sees only canopy, LiDAR penetrates the tree cover and reveals what lies beneath.

When archaeologists from Cambridge University and partner institutions applied AI-assisted LiDAR analysis to the Maya lowlands of northern Guatemala in 2024, they expected to refine existing maps. What they found overturned a century of assumptions about Maya civilization.

More than 60,000 structures (plazas, pyramids, causeways, reservoirs, agricultural terraces) were identified beneath the jungle in a single survey. Not scattered settlements. A continuous urban network. Elevated roads connecting cities. Sophisticated water management systems spanning hundreds of kilometers. Population density estimates for the Mirador Basin alone now approach those of pre-industrial northern Europe.

The paradigm shift is not incremental. For generations, the working model was that the Maya built impressive ceremonial centers surrounded by dispersed agricultural communities. The LiDAR data replaces that model entirely. What existed was an interconnected civilization of substantial urban scale: one that archaeology had simply never had the tools to see.

THE SCALE

The 2024 LiDAR survey identified more Maya structures in a single campaign than the entire twentieth century of ground-based archaeology combined. The technology does not require excavation. It reads the landscape as a data surface and allows AI models to distinguish man-made geometries from natural formations at a resolution of centimeters.

Restoring Words That Survived Only in Fragments

Ancient Greek inscriptions were carved into stone across the Mediterranean world for more than a thousand years. Thousands survive. Most are damaged: edges chipped, surfaces worn, sections missing. The standard scholarly practice is to work with parallel inscriptions, known formulaic phrases, and contextual knowledge to propose restorations of the missing text. It is slow, contested, and dependent on the depth of a single scholar's expertise.

In 2022, DeepMind published Ithaca in Nature: a deep learning model trained on 78,000 ancient Greek inscriptions from the Packard Humanities Institute. Given a fragment of damaged text, Ithaca returns three outputs: the most probable missing words, ranked by confidence; an estimated date of composition within a range; and a probable region of origin.

The accuracy numbers are specific. Ithaca alone achieves 62% on text restoration tasks where the ground truth is known. Human experts working alone achieve 43%. Human experts working with Ithaca achieve 72%.

"The model doesn't just fill in the blanks. It tells you which blanks are more likely than others, and why."

Ithaca research team, DeepMind / Nature, 2022

The methodological point matters. Ithaca does not replace the epigrapher. It amplifies them. No single scholar can hold 78,000 inscriptions in working memory simultaneously, track every grammatical pattern, every regional spelling convention, every formulaic variation by period. The model can. It compresses a library into a lookup function and makes that lookup function available in real time to the human working with a damaged stone.

The Pyramids, Recomputed

How the Great Pyramid of Giza was built has been a subject of serious engineering debate for more than two centuries. The structure is composed of approximately 2.3 million stone blocks, most weighing between 2.5 and 15 tonnes, assembled to a height of 146 meters with a precision that modern engineers find difficult to replicate. The workforce, the logistics, the ramp systems, the timeline: all have been proposed, modeled, disputed, and revised.

In 2025 and 2026, two independent research teams published competing models in npj Heritage Science, a Nature peer-reviewed journal. The first proposed a pulley-and-counterweight system on sliding ramps. The second, published in March 2026, used Finite Element Analysis to test an integrated edge-ramp model: a continuous spiral ramp built into the outer casing of the pyramid itself, combined with internal logistics corridors.

Both models converge on a construction timeline of 20 to 27 years. Both use computational simulation not to assert how the pyramid was built but to eliminate configurations that are structurally or logistically impossible given the material constraints of the site: the weight of the blocks, the slope limits for human-powered transport, the quarry locations, the flood cycle of the Nile.

The AI component is not dramatic. It is methodological. Finite Element Analysis requires solving millions of equations simultaneously to model how stress distributes through a structure under different loading conditions. Logistics simulation requires testing thousands of transport configurations against hard physical constraints. Neither is possible at the required scale without computational power. The software does what archaeology and engineering could not do by hand: it tests every plausible answer until only the structurally consistent ones survive.

What These Experiments Actually Prove About AI

The cases are different in subject and method. They share one mechanism. In each, AI is doing something specific: it is identifying patterns across a body of evidence too large, too damaged, or too spatially distributed for human analysis to cover systematically.

The Vesuvius scrolls were not unreadable because no one was clever enough. They were unreadable because the signal (ink traces at the micron level inside carbonized papyrus) was beneath the threshold of human visual processing. The neural network operates at a different threshold.

The Maya cities were not undiscovered because archaeologists were negligent. They were invisible because the jungle absorbed every ground-level signal. LiDAR eliminated the jungle as a variable. AI reduced 60,000 LiDAR returns to a structured map in the time that a ground survey would have covered a single site.

The distinction matters because it defines the limit. AI expands what is measurable. It cannot expand what was never recorded. The Vesuvius collection represents one library, belonging to one wealthy Roman family, in one city. The Maya structures visible to LiDAR are those made of stone or densely packed earth. The Greek inscriptions in Ithaca's training data are those that survived, were excavated, and were digitized. The AI is as powerful as the data it trains on, and no more. The past it reveals is the past that left physical traces. The rest remains silent, as it always has.

What the Past Owes the Present

There is a question underneath all of this that computational archaeology does not answer, and is not designed to answer. When a machine restores a word to a damaged inscription, it is drawing on patterns learned from thousands of other inscriptions: the ones that were preserved, the ones that were found, the ones belonging to cultures wealthy enough and prominent enough to leave stone records across the Mediterranean. The model reflects the survivorship bias of history itself.

This is not a criticism of the technology. It is a description of its limits, limits that human scholarship shares and that AI makes newly visible precisely because the AI makes the selection criteria explicit. A model trained on 78,000 inscriptions is not neutral. It is a model trained on the inscriptions that survived. The absent records (the oral traditions, the wooden tablets, the perishable materials of cultures that did not build in stone) are absent from the model because they are absent from the archive.

What AI has given archaeology, history, and classics is not omniscience. It is scale and sensitivity: the ability to read signals that were always there but below the threshold of human perception, and to do so across a volume of evidence that no individual scholar could process in a lifetime.

For the first time, questions that were ontologically closed (what did that scroll say? how long did it take to build that pyramid? what word belongs in that gap in the stone?) now have answers that can be tested, revised, and refined. The same architecture that predicts the next word in a sentence is now predicting the next word in an inscription carved two thousand years ago.

The past was always there. The instrument for reading it has only just arrived.