Human Generated Data

Title

Untitled (couple seated on paino stool)

Date

1938

People

Artist: Joseph Janney Steinmetz, American 1905 - 1985

Classification

Photographs

Credit Line

Harvard Art Museums/Fogg Museum, Transfer from the Carpenter Center for the Visual Arts, American Professional Photographers Collection, 4.2002.11657

Copyright

© Estate of Joseph Janney Steinmetz

Human Generated Data

Title

Untitled (couple seated on paino stool)

People

Artist: Joseph Janney Steinmetz, American 1905 - 1985

Date

1938

Classification

Photographs

Credit Line

Harvard Art Museums/Fogg Museum, Transfer from the Carpenter Center for the Visual Arts, American Professional Photographers Collection, 4.2002.11657

Copyright

© Estate of Joseph Janney Steinmetz

Machine Generated Data

Tags

Amazon
created on 2022-01-15

Person 97.4
Human 97.4
Person 97.2
Helmet 80.7
Clothing 80.7
Apparel 80.7
Indoors 74.5
Room 73.6
Portrait 64.7
Face 64.7
Photography 64.7
Photo 64.7
Furniture 58.9
Monitor 57.1
Electronics 57.1
Display 57.1
Screen 57.1
Text 56.5
Leisure Activities 56.5
Oven 55.1
Appliance 55.1

Clarifai
created on 2023-10-25

people 99.9
monochrome 99.7
adult 99
wear 97.1
man 96.6
one 93.8
administration 93.7
room 92.7
two 91.8
furniture 90.3
group 89.8
chair 88.3
woman 87
scientist 85.7
indoors 84.1
group together 83.7
three 82.9
outfit 80.9
sit 80.7
outerwear 80.6

Imagga
created on 2022-01-15

man 36.3
person 31.3
people 31.2
male 29.8
adult 25.1
men 19.7
working 18.6
indoors 18.4
worker 17
patient 16.1
lifestyle 15.9
professional 15.5
work 14.9
room 14.8
kitchen 14.3
musical instrument 14.3
portrait 14.2
crutch 14
happy 13.8
smiling 13.7
equipment 13.2
office 12.4
job 12.4
home 12
business 11.5
medical 11.5
casual 11
occupation 11
staff 10.8
holding 10.7
computer 10.4
surgeon 10.3
device 10
shop 9.9
instrument 9.7
stick 9.6
black 9.6
case 9.4
happiness 9.4
back 9.2
hospital 9.1
chair 9
health 9
music 9
fun 9
one 9
newspaper 8.9
handsome 8.9
women 8.7
standing 8.7
tool 8.5
senior 8.4
musician 8.4
hand 8.4
camera 8.3
human 8.2
team 8.1
clothing 8.1
hairdresser 8
percussion instrument 8
interior 8
businessman 7.9
life 7.8
sixties 7.8
60s 7.8
education 7.8
modern 7.7
sick person 7.7
profession 7.7
active 7.7
two 7.6
doctor 7.5
leisure 7.5
cheerful 7.3
industrial 7.3
group 7.3
student 7.2
looking 7.2
face 7.1
medicine 7

Google
created on 2022-01-15

Microsoft
created on 2022-01-15

text 99.7
black and white 93.4
indoor 92.6
person 87.8
clothing 86.7
drawing 50.8
cluttered 10

Color Analysis

Face analysis

Amazon

Google

AWS Rekognition

Age 29-39
Gender Female, 89.9%
Calm 98%
Happy 1.1%
Sad 0.5%
Angry 0.1%
Disgusted 0.1%
Surprised 0.1%
Fear 0.1%
Confused 0.1%

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very unlikely

Google Vision

Surprise Very unlikely
Anger Very unlikely
Sorrow Very unlikely
Joy Very unlikely
Headwear Very unlikely
Blurred Very unlikely

Feature analysis

Amazon

Person 97.4%
Helmet 80.7%

Categories

Text analysis

Amazon

8
350
EARS
GO
7795
NO 350
NO
MJ17
779.5
EARS ROLBERK
MJ17 YEER AFOA
AFOA
YEER
For
ROLBERK

Google

7795 EARS ROEHRY GK NO350. 1195 1195 MIR YT33A2 AA
7795
EARS
GK
1195
MIR
YT33A2
AA
ROEHRY
NO350.