19.02.2019 Views

DM1901

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

CASE STUDY: NEW YORK TIMES Dm<br />

The Times to protect one of their most<br />

unique assets migrating from steel filing<br />

cabinets to a cloud-based platform where<br />

journalists can bring visual storytelling to a<br />

whole new level."<br />

Simply storing high-resolution images is<br />

not enough to create a system that photo<br />

editors can easily use. A working asset<br />

management system must allow the users<br />

to be able to browse and search for photos<br />

easily. The Times built a processing pipeline<br />

that stores and processes the photos and<br />

will use cloud technology to process and<br />

recognise text, handwriting and other<br />

details that can be found in the images.<br />

MACHINE LEARNING ADDS INSIGHTS<br />

Storing the images is only one half of<br />

the story. To make an archive like The<br />

Times' morgue even more accessible and<br />

useful, it's beneficial to leverage<br />

additional GCP features. In the case of<br />

The Times, one of the bigger challenges<br />

in scanning their photo archive has been<br />

adding data regarding the contents of<br />

the images. Google's Cloud Vision API<br />

helps fill that gap.<br />

The photo (above) of the old Penn<br />

Station from The Times gives an example.<br />

Without additional context, it's not clear<br />

from the front of the photo what it<br />

contains, but the back of the photo<br />

contains a wealth of useful information,<br />

and the Cloud Vision API can help<br />

process, store, and read it. When we<br />

submit the back of the image to the API<br />

with no additional processing, the Cloud<br />

Vision API detects the text and recognises<br />

data such as dates, descriptions ('The<br />

scene in Pennsylvania Station yesterday<br />

afternoon') and context ('Pub NYT Sun<br />

5/2/93', 'RECEIVED DEC 25 1942'). Of<br />

course, the digital text transcription isn't<br />

perfect, but it's faster and more cost<br />

effective than alternatives for processing<br />

millions of images.<br />

This is only the start of what's possible for<br />

companies with physical archives. They can<br />

use the Vision API to identify objects,<br />

places and images. For example, if we pass<br />

the black and white photo above through<br />

the Cloud Vision API with Logo Detection,<br />

it recognises Pennsylvania Station.<br />

Furthermore, AutoML can be used to<br />

better identify images in collections using a<br />

corpus of already captioned images.<br />

The Cloud Natural Language API could<br />

also be used to add additional semantic<br />

information to recognised text. For<br />

example, passing the text "THE WAY IT<br />

WAS - Crowded Penn Station in 1942, an<br />

era when only the brave flew - to<br />

Washington, Miami and assorted way<br />

stations." through the Cloud Natural<br />

Language API, it correctly identifies "Penn<br />

Station," "Washington," and "Miami" as<br />

locations, and classifies the entire sentence<br />

into the category "travel" and the<br />

subcategory "bus & rail."<br />

Going forward the NYT newsroom will<br />

use the digitised archives to inspire stories<br />

for Past Tense, a body of coverage<br />

dedicated to revisiting history. The first<br />

package from The Times newsroom to<br />

utilise the digitised archives will focus on<br />

how The Times covered California in the<br />

20th century, examining how California's<br />

free-spiritedness and culture of recreation<br />

and innovation appeared to Times<br />

journalists 3,000 miles away.<br />

More info: cloud.google.com<br />

www.document-manager.com<br />

January/February 2019<br />

@DMMagAndAwards<br />

27

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!