Human Generated Data

Title

Untitled (three women seated on flowered couch)

Date

1942

People

Artist: Joseph Janney Steinmetz, American 1905 - 1985

Classification

Photographs

Credit Line

Harvard Art Museums/Fogg Museum, Transfer from the Carpenter Center for the Visual Arts, American Professional Photographers Collection, 4.2002.10542

Copyright

© Estate of Joseph Janney Steinmetz

Human Generated Data

Title

Untitled (three women seated on flowered couch)

People

Artist: Joseph Janney Steinmetz, American 1905 - 1985

Date

1942

Classification

Photographs

Credit Line

Harvard Art Museums/Fogg Museum, Transfer from the Carpenter Center for the Visual Arts, American Professional Photographers Collection, 4.2002.10542

Copyright

© Estate of Joseph Janney Steinmetz

Machine Generated Data

Tags

Amazon
created on 2022-01-09

Human 99.8
Person 99.8
Person 98.3
Person 96.1
Person 93
Art 86.3
Leisure Activities 79.4
Sitting 79
Musical Instrument 71.6
Guitar 70
Musician 69.6
Furniture 65.8
Female 64.9
Drawing 63.8
Photography 63.2
Photo 63.2
Face 62.2
Portrait 62.2
Girl 58.5
Painting 58.2
Canvas 56.9
Guitarist 56.9
Performer 56.9

Clarifai
created on 2023-10-25

people 99.9
adult 99.5
woman 99.2
group 98.4
man 98.2
furniture 97.9
sit 96
child 92
art 91
administration 90.6
elderly 89.8
family 89.7
room 89.1
two 89.1
wear 86.4
music 85.7
three 85.5
war 84.6
painting 83.8
four 83

Imagga
created on 2022-01-09

man 34.9
male 29.1
person 27.7
people 27.3
adult 23.8
television 22.2
room 21.4
happy 20
office 19.6
smiling 18.1
portrait 16.8
sitting 16.3
senior 15
smile 15
computer 14.8
business 14.6
looking 14.4
men 13.7
group 13.7
home 13.6
laptop 13.4
couple 13.1
indoor 12.8
old 12.5
telecommunication system 12.4
businessman 12.4
fun 12
dark 11.7
working 11.5
light 11.4
mature 11.2
desk 11.1
lifestyle 10.8
black 10.8
indoors 10.5
love 10.3
casual 10.2
chair 9.8
modern 9.8
together 9.6
career 9.5
pretty 9.1
holding 9.1
fashion 9
classroom 9
one 9
lady 8.9
handsome 8.9
job 8.8
women 8.7
work 8.6
respirator 8.6
happiness 8.6
world 8.5
iron lung 8.3
team 8.1
sexy 8
family 8
case 7.9
patient 7.9
nightlife 7.8
education 7.8
teacher 7.7
attractive 7.7
crowd 7.7
husband 7.6
adults 7.6
device 7.5
human 7.5
board 7.2
music 7.2
hair 7.1
face 7.1

Google
created on 2022-01-09

Microsoft
created on 2022-01-09

text 99.9
person 92.2
clothing 83.5
black and white 75.1
human face 73

Color Analysis

Face analysis

Amazon

Google

AWS Rekognition

Age 27-37
Gender Male, 91.4%
Surprised 24.1%
Calm 18.6%
Angry 17.3%
Sad 13.5%
Fear 12.5%
Happy 6.1%
Confused 4.8%
Disgusted 3%

AWS Rekognition

Age 22-30
Gender Male, 98.3%
Calm 86%
Sad 10%
Surprised 1.1%
Confused 0.8%
Happy 0.6%
Disgusted 0.5%
Angry 0.5%
Fear 0.3%

AWS Rekognition

Age 45-51
Gender Male, 61.9%
Happy 58.6%
Calm 34.3%
Surprised 3%
Sad 2%
Confused 0.8%
Disgusted 0.6%
Angry 0.4%
Fear 0.3%

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very unlikely

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very unlikely

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very unlikely

Feature analysis

Amazon

Person 99.8%
Guitar 70%

Text analysis

Amazon

20291.
DC
in

Google

20291.
2024
20291. 2024 20291.