Human Generated Data

Title

Covered long-necked jar (hu)

Date

25-220 CE

People
Classification

Vessels

Human Generated Data

Title

Covered long-necked jar (hu)

People
Date

25-220 CE

Classification

Vessels

Machine Generated Data

Tags

Amazon

Jar 99.2
Pottery 99.2
Vase 99.2
Porcelain 85.2
Art 85.2
Ketchup 72.2
Food 72.2

Clarifai

pottery 100
no person 99.8
clay 99.7
earthenware 99.4
one 99.3
vase 99
container 98.8
still life 97.5
art 97.5
ceramic 96.9
ancient 96.8
antique 96.5
watercraft 96.2
sculpture 95.9
tableware 95.4
urn 95.3
handmade 94.2
pot 92.7
porcelain 92.6
retro 92.2

Imagga

jar 100
vase 100
vessel 83
container 74.8
earthenware 45.1
ceramic ware 30.1
glass 25.5
utensil 22.6
pot 22.3
antique 20.8
old 20.2
craft 20
object 19.8
brown 19.1
bottle 18.6
jug 18.6
decorative 18.4
pottery 17.8
traditional 17.5
ancient 17.3
clay 16.6
decoration 15.9
drink 15.9
classical 13.4
pitcher 13.2
wooden 13.2
liquid 13
culture 12.8
vintage 12.4
art 12.4
retro 12.3
close 12
ceramics 11.7
ceramic 11.6
table 11.2
wood 10.8
past 10.6
old fashioned 10.4
china 10.1
water 10
terracotta 9.9
style 9.6
ornament 9.5
ornate 9.1
beverage 9
black 9
transparent 9
history 8.9
closeup 8.8
porcelain 8.6
handle 8.6
nobody 8.5
food 8.5
alcohol 8.3
kitchen 8
capstan 7.9
cup 7.8
power 7.5
pattern 7.5
single 7.4
aged 7.2
color 7.2

Google

earthenware 97.4
Vase 96
Ceramic 92.2
Pottery 90.7
Artifact 88.5
Urn 70.8
Serveware 56.2
Art 50.2

Microsoft

vase 99.7
sitting 99.5
wall 99.4
earthenware 96.6
pottery 95.8
indoor 92.4
ceramic 88.9
ceramic ware 83.6
urn 81.9
orange 81.8
plant 77.4
jug 76.5
artifact 76
bowl 73
background 67.3
flowerpot 65.9
red 65.5
pitcher 63.6
teapot 63.2
still life photography 58.2
lid 52.5
still life 51.1
old 49.7
jar 29.5
porcelain 23.8
stoneware 22.7

Feature analysis

Amazon

Ketchup 72.2%

Captions

Microsoft

a vase sitting on a table 70.6%
a vase sitting on top of a table 66.7%
a brown vase on a table 66.6%