Human Generated Data

Title

Untitled (wedding guests standing near cake table)

Date

1953

People

Artist: Joseph Janney Steinmetz, American 1905 - 1985

Classification

Photographs

Credit Line

Harvard Art Museums/Fogg Museum, Transfer from the Carpenter Center for the Visual Arts, American Professional Photographers Collection, 4.2002.8747

Copyright

© Estate of Joseph Janney Steinmetz

Human Generated Data

Title

Untitled (wedding guests standing near cake table)

People

Artist: Joseph Janney Steinmetz, American 1905 - 1985

Date

1953

Classification

Photographs

Credit Line

Harvard Art Museums/Fogg Museum, Transfer from the Carpenter Center for the Visual Arts, American Professional Photographers Collection, 4.2002.8747

Copyright

© Estate of Joseph Janney Steinmetz

Machine Generated Data

Tags

Amazon
created on 2022-01-09

Person 97.3
Human 97.3
Person 97.3
Person 90
Person 87.2
Person 87
Person 86.4
Person 84.8
Person 84.6
Person 82.2
Person 80.3
Person 78.8
Poster 77.2
Advertisement 77.2
Urban 72.1
Person 71.6
Person 70.8
Meal 69.7
Food 69.7
People 68.5
Outdoors 68.2
Person 61.6
Nature 60.6
Person 60.5
Person 57.9
Clothing 57.8
Apparel 57.8
Photography 57.8
Photo 57.8
Face 56.2
Person 49

Clarifai
created on 2023-10-25

people 99.8
many 99.1
group 98.5
group together 97.7
man 96.9
war 96.7
military 94.5
adult 93
several 90
soldier 88.2
vehicle 87.8
interaction 87.5
leader 87.5
woman 83.7
outfit 83.3
art 82.5
child 82.5
wear 82.4
administration 82.4
veil 80.4

Imagga
created on 2022-01-09

musical instrument 31.5
accordion 17.3
man 16.1
wind instrument 15.5
person 15.4
old 14.6
keyboard instrument 14
adult 12.4
people 12.3
black 11.4
shop 11.2
holiday 10.7
clothing 10.7
bass 10.5
shopping cart 10.5
style 10.4
wheeled vehicle 10.3
grunge 10.2
business 9.7
urban 9.6
male 9.3
shopping 9.2
studio 9.1
stringed instrument 9.1
music 9
equipment 8.9
happy 8.8
play 8.6
men 8.6
store 8.5
house 8.3
device 8.3
one 8.2
dirty 8.1
handcart 8.1
standing 7.8
cart 7.8
portrait 7.8
outdoors 7.6
power 7.5
fashion 7.5
buy 7.5
city 7.5
vintage 7.4
guitar 7.4
basket 7.3
seller 7.3
art 7.2
hair 7.1
women 7.1
market 7.1
child 7.1
day 7.1
bowed stringed instrument 7

Google
created on 2022-01-09

Microsoft
created on 2022-01-09

outdoor 98.5
black and white 93.6
person 93.5
people 80
text 73.4
group 72.1
man 63.1
clothing 62.3
monochrome 51.3
crowd 25.1

Color Analysis

Face analysis

Amazon

AWS Rekognition

Age 16-24
Gender Female, 55.6%
Sad 49.3%
Calm 34.8%
Fear 5.5%
Surprised 4.4%
Confused 3.6%
Disgusted 1.4%
Angry 0.5%
Happy 0.4%

AWS Rekognition

Age 21-29
Gender Female, 83.1%
Calm 82.4%
Happy 16.4%
Sad 0.6%
Surprised 0.2%
Disgusted 0.1%
Confused 0.1%
Angry 0.1%
Fear 0%

AWS Rekognition

Age 16-24
Gender Female, 61.8%
Calm 63.9%
Angry 28.1%
Sad 4.5%
Happy 1.4%
Surprised 0.7%
Confused 0.6%
Fear 0.4%
Disgusted 0.4%

Feature analysis

Amazon

Person 97.3%
Poster 77.2%

Text analysis

Amazon

5
38594
٢8
YТ3-X

Google

38S94
YT37A°2-
38S94 58 YT37A°2- AO
58
AO