Home /
Usecases /
The New York Times automates historic photo content identification and detail extraction with Google computer vision
Jan. 12, 2019
Summary:
The New York Times is digitizing their repository of several million historic photos using Google Cloud Services. Machine learning is used to identify and extract bits of text in the photos automatically, adding important content and context to the files without human intervention.
Problem:
"For over 100 years, The Times has archived approximately five to seven million of its old photos in hundreds of file cabinets three stories below street level near their Times Square offices in a location called the “morgue.” Many of the photos have been stored in folders and not seen in years. Although a card catalog provides an overview of the archive’s contents, there are many details in the photos that are not captured in an indexed form."