Human Generated Data

Title

Photo Album

Date

c. 1857 - c. 1874

People
Classification

Photographs

Credit Line

Harvard Art Museums/Arthur M. Sackler Museum, Transfer from Widener Library, Harvard University, 1978.484.49

Human Generated Data

Title

Photo Album

People
Date

c. 1857 - c. 1874

Classification

Photographs

Machine Generated Data

Tags

Amazon
created on 2022-01-09

Person 99.7
Human 99.7
Painting 92.6
Art 92.6
Animal 91.5
Mammal 91.5
Horse 91.5
Person 91.3
Horse 89.6
Outdoors 76.6
Theme Park 75.9
Amusement Park 75.9
Nature 70.3
Clothing 68
Apparel 68
People 66.8
Horse 66
Person 62.1

Imagga
created on 2022-01-09

sand 35.1
tree 30.2
beach 27.2
landscape 26.8
sky 23.6
travel 21.1
sun 20.9
sea 19.6
trees 18.7
vacation 17.2
sunset 17.1
summer 16.7
ocean 16.6
snow 16.1
resort 15
park 14.9
tourism 14.9
water 14
forest 13.9
holiday 13.6
cattle 13.4
ox 13.1
outdoors 12.8
horse 12.7
field 12.6
rural 12.3
tropical 11.9
winter 11.9
grass 11.9
coast 11.7
country 11.4
shore 11.4
tourist 11.2
earth 10.7
soil 10.7
desert 10.4
sunrise 10.3
weather 10.3
ranch 10.1
island 10.1
outdoor 9.9
farm 9.8
swimsuit 9.8
scenic 9.7
light 9.4
old 9.1
scenery 9
people 8.9
camel 8.9
man 8.7
bovine 8.7
walking 8.5
clouds 8.5
palm 8.4
lake 8.2
peaceful 8.2
countryside 8.2
garment 8.1
male 7.9
wild 7.8
rock 7.8
sunny 7.8
sport 7.7
bikini 7.7
paradise 7.5
dark 7.5
city 7.5
silhouette 7.5
environment 7.4
morning 7.2
mountain 7.1
season 7

Google
created on 2022-01-09

Microsoft
created on 2022-01-09

outdoor 99.8
tree 99.4
horse 97.6
old 93.4
people 72.9
animal 65.5
group 65.5
cattle 57.8
mammal 54.1
drawn 31.8

Color Analysis

Face analysis

Amazon

Google

AWS Rekognition

Age 23-31
Gender Male, 99.7%
Sad 49.9%
Calm 28.5%
Happy 7.8%
Confused 3.6%
Disgusted 3.5%
Angry 3.1%
Fear 2.3%
Surprised 1.3%

AWS Rekognition

Age 23-33
Gender Male, 97.7%
Sad 42.8%
Calm 15.9%
Fear 14.4%
Happy 9.5%
Disgusted 6.8%
Angry 4.4%
Surprised 4.3%
Confused 1.9%

AWS Rekognition

Age 22-30
Gender Male, 98.7%
Calm 36.6%
Sad 33.3%
Disgusted 17.8%
Fear 6%
Angry 2.4%
Happy 2.3%
Confused 1%
Surprised 0.6%

AWS Rekognition

Age 26-36
Gender Male, 99.1%
Disgusted 88.6%
Calm 5.6%
Happy 1.5%
Angry 1.2%
Sad 0.9%
Surprised 0.9%
Fear 0.8%
Confused 0.4%

AWS Rekognition

Age 20-28
Gender Male, 97.5%
Sad 98.4%
Fear 1.3%
Calm 0.1%
Angry 0.1%
Disgusted 0%
Confused 0%
Happy 0%
Surprised 0%

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very unlikely

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very likely

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Likely

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very unlikely

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very unlikely

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very unlikely

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very unlikely

Feature analysis

Amazon

Person 99.7%
Painting 92.6%
Horse 91.5%

Captions

Microsoft

a group of people in an old photo of a horse 96.6%
a vintage photo of a group of people standing next to a horse 95.4%
a group of people standing in an old photo of a horse 95.3%

Text analysis

Amazon

Ming
Co.
Nanking
Road.
42 Nanking Road.
&
42
Sue Yuen Ming & Co.
58.
Sue
Yuen

Google

52 Sse Yen Ming & Co. 42. Nankng Road.
52
Road.
Sse
Yen
Ming
&
Co.
42.
Nankng