Automatic hierarchical classification of silhouettes of 3D objects

Yoram Gdalyahu*, Daphna Weinshall

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

6 Scopus citations

Abstract

The organization of image databases can rely upon different aspects of image similarity. Here we extract silhouettes from images of three dimensional objects, and rely upon curve similarity for image classification. Our scheme avoids the embedding of images in a vector space. Instead, we propose a curve dissimilarity measure which relies upon a novel curve matching syntactic algorithm, and use it to represent the database as a complete graph, with nodes representing the images and dissimilarity values assigning weights to the edges. A robust clustering algorithm, which is based on a physical ferromagnet model, is used to find the hierarchical structure underlying the collection of images. We tested our scheme with a database of 90 real images of 6 objects, some of them very different, others rather similar. We get a perfect hierarchical classification of these images into 6 classes of objects belonging to 3 different families.

Original languageEnglish
Title of host publicationProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
Pages787-793
Number of pages7
DOIs
StatePublished - 1998
EventProceedings of the 1998 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Santa Barbara, CA, USA
Duration: 23 Jun 199825 Jun 1998

Publication series

NameProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
ISSN (Print)1063-6919

Conference

ConferenceProceedings of the 1998 IEEE Computer Society Conference on Computer Vision and Pattern Recognition
CitySanta Barbara, CA, USA
Period23/06/9825/06/98

Fingerprint

Dive into the research topics of 'Automatic hierarchical classification of silhouettes of 3D objects'. Together they form a unique fingerprint.

Cite this