Human Generated Data

Title

Untitled (two photographs: bride and groom posing in front of groomsmen and male guests in carpeted room; guests sitting around tables during wedding reception)

Date

1963, printed later

People

Artist: Martin Schweig, American 20th century

Classification

Photographs

Credit Line

Harvard Art Museums/Fogg Museum, Transfer from the Carpenter Center for the Visual Arts, American Professional Photographers Collection, 4.2002.12373

Human Generated Data

Title

Untitled (two photographs: bride and groom posing in front of groomsmen and male guests in carpeted room; guests sitting around tables during wedding reception)

People

Artist: Martin Schweig, American 20th century

Date

1963, printed later

Classification

Photographs

Credit Line

Harvard Art Museums/Fogg Museum, Transfer from the Carpenter Center for the Visual Arts, American Professional Photographers Collection, 4.2002.12373

Machine Generated Data

Tags

Amazon
created on 2019-11-16

Apparel 100
Clothing 100
Human 99
Person 99
Person 98.7
Person 98.5
Person 98.3
Person 98.3
Person 97.9
Person 97.7
Interior Design 97
Indoors 97
Person 96.4
Person 96
Person 94.9
Person 94.9
Person 93.6
Fashion 93.4
Gown 93.4
Robe 92.4
Person 91.5
Person 89
Person 88.1
Female 87.8
Room 87.5
Advertisement 87.3
Collage 87.3
Poster 87.3
Wedding 81.6
Crowd 80.4
Audience 80.4
Dress 76.1
Wedding Gown 75
Woman 74.5
Person 72.7
Evening Dress 71.8
Overcoat 69.7
Coat 69.7
Suit 69.7
Bride 68.3
Head 62.5
Person 62.4
Person 60.4
Stage 60
Face 59.1
Portrait 59
Photo 59
Photography 59
Theater 56.6
Long Sleeve 55.7
Sleeve 55.7

Clarifai
created on 2019-11-16

people 99.9
room 99.5
furniture 99.1
adult 98.5
group 98.5
indoors 98.5
many 97.7
man 97.6
one 97.3
woman 94.8
monochrome 93
wear 92.2
mirror 90.9
chair 90.6
child 90
group together 88.7
desk 88.2
sit 88
family 87.2
exhibition 86.7

Imagga
created on 2019-11-16

background 20.2
film 19.3
screen 19.2
people 16.7
art 14.9
man 14.9
display 14.2
person 14
black 13.9
x-ray film 13.5
paper 13.4
business 13.4
case 13.1
office 13
old 12.5
design 12.4
grunge 11.9
computer 11.6
photographic paper 10.4
graphic 10.2
symbol 10.1
vintage 10
holding 9.9
technology 9.6
sketch 9.5
money 9.4
male 9.2
adult 9.1
texture 9
sign 9
currency 9
financial 8.9
style 8.9
pattern 8.9
drawing 8.9
blank 8.6
finance 8.4
equipment 8.4
message 8.2
retro 8.2
bank 8.2
board 8.1
electronic device 8.1
antique 8
women 7.9
monitor 7.7
professional 7.7
modern 7.7
card 7.6
communication 7.5
house 7.5
dark 7.5
human 7.5
economy 7.4
student 7.2
window 7.2
home 7.2
portrait 7.1
market 7.1
working 7.1
work 7.1

Google
created on 2019-11-16

Microsoft
created on 2019-11-16

text 97.8
wedding dress 89.5
gallery 86.9
person 82.7
black and white 81
room 78.7
dress 70.4
bride 68.7
clothing 67.9
woman 58
old 44.4

Color Analysis

Face analysis

Amazon

AWS Rekognition

Age 28-44
Gender Male, 54.9%
Happy 45.6%
Sad 45.1%
Angry 47.6%
Calm 46.1%
Fear 45.2%
Surprised 48.7%
Disgusted 45.9%
Confused 45.8%

AWS Rekognition

Age 39-57
Gender Female, 54.6%
Disgusted 45%
Fear 45%
Calm 45.4%
Surprised 45.1%
Angry 45.1%
Confused 45%
Happy 54.3%
Sad 45%

AWS Rekognition

Age 38-56
Gender Male, 50.2%
Surprised 49.5%
Calm 49.6%
Happy 49.6%
Angry 49.6%
Fear 49.8%
Sad 49.6%
Confused 49.5%
Disgusted 49.8%

AWS Rekognition

Age 36-54
Gender Male, 50.5%
Calm 49.6%
Sad 49.6%
Happy 49.7%
Angry 49.7%
Surprised 49.5%
Confused 49.9%
Disgusted 49.5%
Fear 49.5%

AWS Rekognition

Age 34-50
Gender Female, 50%
Disgusted 49.5%
Sad 49.6%
Calm 49.7%
Fear 49.6%
Happy 49.9%
Confused 49.5%
Angry 49.6%
Surprised 49.6%

AWS Rekognition

Age 13-25
Gender Male, 53.8%
Fear 45.2%
Angry 45.1%
Calm 53.8%
Happy 45%
Sad 45.6%
Disgusted 45.1%
Confused 45.1%
Surprised 45.1%

AWS Rekognition

Age 22-34
Gender Male, 50.3%
Calm 49.5%
Angry 49.5%
Fear 49.6%
Happy 49.5%
Sad 50.4%
Confused 49.5%
Disgusted 49.5%
Surprised 49.5%

AWS Rekognition

Age 15-27
Gender Male, 50.1%
Calm 49.5%
Fear 49.8%
Disgusted 49.5%
Surprised 49.5%
Happy 49.5%
Angry 49.6%
Confused 49.5%
Sad 50%

AWS Rekognition

Age 51-69
Gender Male, 54.6%
Angry 45.1%
Fear 45.1%
Confused 45.2%
Happy 53.9%
Calm 45%
Disgusted 45%
Surprised 45.1%
Sad 45.7%

AWS Rekognition

Age 47-65
Gender Male, 50.5%
Surprised 49.6%
Sad 49.6%
Disgusted 49.5%
Calm 50.2%
Fear 49.5%
Angry 49.6%
Happy 49.5%
Confused 49.5%

AWS Rekognition

Age 23-37
Gender Male, 50.5%
Angry 49.5%
Disgusted 49.5%
Happy 50.3%
Calm 49.5%
Sad 49.5%
Surprised 49.5%
Fear 49.5%
Confused 49.5%

AWS Rekognition

Age 38-56
Gender Male, 50.2%
Angry 49.7%
Happy 49.5%
Calm 49.5%
Confused 49.5%
Disgusted 49.6%
Fear 49.6%
Sad 50%
Surprised 49.5%

AWS Rekognition

Age 33-49
Gender Male, 50.4%
Confused 49.5%
Calm 49.6%
Sad 49.5%
Surprised 50.4%
Happy 49.5%
Disgusted 49.5%
Fear 49.5%
Angry 49.5%

AWS Rekognition

Age 48-66
Gender Male, 50.4%
Surprised 49.5%
Angry 49.7%
Sad 49.8%
Fear 49.5%
Disgusted 49.5%
Happy 49.5%
Calm 49.9%
Confused 49.5%

AWS Rekognition

Age 28-44
Gender Male, 50.4%
Surprised 49.5%
Disgusted 49.5%
Sad 49.5%
Angry 50.4%
Calm 49.5%
Happy 49.5%
Fear 49.5%
Confused 49.5%

AWS Rekognition

Age 17-29
Gender Female, 50.2%
Angry 45.3%
Confused 45.5%
Disgusted 46.6%
Happy 45.3%
Sad 50%
Fear 46.4%
Calm 45.6%
Surprised 45.2%

Feature analysis

Amazon

Person 99%

Categories

Imagga

interior objects 79.7%
food drinks 9.7%
paintings art 9.7%

Text analysis

Amazon

2055J

Google

2 e S S J M T
2
e
S
J
M
T