Human Generated Data

Title

Untitled (people in department store)

Date

1935

People

Artist: Hamblin Studio, American active 1930s

Classification

Photographs

Credit Line

Harvard Art Museums/Fogg Museum, Transfer from the Carpenter Center for the Visual Arts, American Professional Photographers Collection, 4.2002.21916

Human Generated Data

Title

Untitled (people in department store)

People

Artist: Hamblin Studio, American active 1930s

Date

1935

Classification

Photographs

Machine Generated Data

Tags

Amazon
created on 2022-03-11

Lighting 95.8
Human 89.5
Person 89.5
Building 87.3
Clinic 85.9
Person 78.5
Airplane 74.6
Aircraft 74.6
Transportation 74.6
Vehicle 74.6
Person 73.8
Workshop 73.7
Factory 72.6
Person 67.8
Person 64
Person 62.8
Coat 60.4
Clothing 60.4
Apparel 60.4
Person 58.4
Market 57.3
Restaurant 57
Cafeteria 57
Person 47.9

Imagga
created on 2022-03-11

building 40.1
structure 31.5
greenhouse 24.2
sketch 22.9
architecture 22.8
construction 22.2
business 21.3
drawing 20.6
urban 20.1
metal 19.3
industry 18.8
steel 17.7
industrial 17.2
canvas tent 17.1
interior 16.8
modern 16.1
city 15.8
technology 15.6
sky 15.3
roof 15.3
work 14.9
glass 14.8
inside 14.7
reflection 13.1
station 12.7
representation 12.6
equipment 12.5
perspective 12.2
travel 12
power 11.8
airport 11.7
transportation 11.7
window 11.3
tower 10.7
factory 10.6
transport 10
water 10
line 9.9
design 9.6
man 9.4
light 9.4
office 9.1
wall 9
futuristic 9
stall 9
hall 8.8
engineering 8.6
iron 8.4
new 8.1
center 7.9
gate 7.9
manufacturing 7.8
people 7.8
mall 7.8
public 7.8
men 7.7
tube 7.7
built 7.7
build 7.6
energy 7.6
chairlift 7.5
electric 7.5
device 7.4
deck 7.4
cable 7.2
worker 7.1
job 7.1

Google
created on 2022-03-11

Microsoft
created on 2022-03-11

indoor 92.2
black and white 91.3
text 89.7

Face analysis

Amazon

AWS Rekognition

Age 20-28
Gender Female, 85.4%
Fear 67.9%
Surprised 18%
Calm 4.5%
Sad 4.3%
Happy 1.8%
Confused 1.7%
Angry 1.1%
Disgusted 0.8%

AWS Rekognition

Age 12-20
Gender Male, 81.7%
Calm 83.6%
Sad 6.7%
Disgusted 3.7%
Happy 2.7%
Angry 2.2%
Confused 0.5%
Fear 0.4%
Surprised 0.2%

Feature analysis

Amazon

Person 89.5%
Airplane 74.6%

Captions

Microsoft

a group of people standing in front of a building 52.6%
a group of people standing next to a building 50.9%
a group of people in a room 50.8%

Text analysis

Amazon

19
15
BVEE
145
YTEE