View-based 3-D Object Retrieval -  Qionghai Dai,  Yue Gao

View-based 3-D Object Retrieval (eBook)

eBook Download: PDF | EPUB
2014 | 1. Auflage
154 Seiten
Elsevier Science (Verlag)
978-0-12-802623-6 (ISBN)
31,95 € inkl. MwSt
Systemvoraussetzungen
31,09 € inkl. MwSt
Systemvoraussetzungen
  • Download sofort lieferbar
  • Zahlungsarten anzeigen

Content-based 3-D object retrieval has attracted extensive attention recently and has applications in a variety of fields, such as, computer-aided design, tele-medicine,mobile multimedia, virtual reality, and entertainment. The development of efficient and effective content-based 3-D object retrieval techniques has enabled the use of fast 3-D reconstruction and model design. Recent technical progress, such as the development of camera technologies, has made it possible to capture the views of 3-D objects. As a result, view-based 3-D object retrieval has become an essential but challenging research topic.

View-based 3-D Object Retrieval introduces and discusses the fundamental challenges in view-based 3-D object retrieval, proposes a collection of selected state-of-the-art methods for accomplishing this task developed by the authors, and summarizes recent achievements in view-based 3-D object retrieval. Part I presents an Introduction to View-based 3-D Object Retrieval, Part II discusses View Extraction, Selection, and Representation, Part III provides a deep dive into View-Based 3-D Object Comparison, and Part IV looks at future research and developments including Big Data application and geographical location-based applications.


  • Systematically introduces view-based 3-D object retrieval, including problem definitions and settings, methodologies, and benchmark testing beds
  • Discusses several key challenges in view-based 3-D object retrieval, and introduces the state-of-the-art solutions
  • Presents the progression from general image retrieval techniques to view-based 3-D object retrieval
  • Introduces future research efforts in the areas of Big Data, feature extraction, and geographical location-based applications


Yue Gao is with the Department of Automation, Tsinghua University. His recent research focuses on the areas of neuroimaging, multimedia and remote sensing. He is a senior member of IEEE.
Content-based 3-D object retrieval has attracted extensive attention recently and has applications in a variety of fields, such as, computer-aided design, tele-medicine,mobile multimedia, virtual reality, and entertainment. The development of efficient and effective content-based 3-D object retrieval techniques has enabled the use of fast 3-D reconstruction and model design. Recent technical progress, such as the development of camera technologies, has made it possible to capture the views of 3-D objects. As a result, view-based 3-D object retrieval has become an essential but challenging research topic. View-based 3-D Object Retrieval introduces and discusses the fundamental challenges in view-based 3-D object retrieval, proposes a collection of selected state-of-the-art methods for accomplishing this task developed by the authors, and summarizes recent achievements in view-based 3-D object retrieval. Part I presents an Introduction to View-based 3-D Object Retrieval, Part II discusses View Extraction, Selection, and Representation, Part III provides a deep dive into View-Based 3-D Object Comparison, and Part IV looks at future research and developments including Big Data application and geographical location-based applications. Systematically introduces view-based 3-D object retrieval, including problem definitions and settings, methodologies, and benchmark testing beds Discusses several key challenges in view-based 3-D object retrieval, and introduces the state-of-the-art solutions Presents the progression from general image retrieval techniques to view-based 3-D object retrieval Introduces future research efforts in the areas of Big Data, feature extraction, and geographical location-based applications

Front Cover 1
View-Based 3-D Object Retrieval 4
Copyright 5
Contents 6
Acknowledgments 8
Preface 10
Part I: The Start 12
Chapter 1: Introduction 14
1.1 The Definition of 3DOR 14
1.2 Model-Based 3DOR Versus View-Based 3DOR 15
1.3 The Challenges of V3DOR 16
1.4 Summary of Our Work 17
1.4.1 View Extraction 18
1.4.2 Representative View Selection 18
1.4.3 Learning the Weights for Multiple Views 19
1.4.4 Distance Measures for Object Matching 20
1.4.5 Learning the Relevance Among 3-D Objects 21
1.5 Structure of This Book 23
1.6 Summary 25
References 25
Chapter 2: The Benchmark and Evaluation 28
2.1 Introduction 28
2.2 The Standard Benchmarks 28
2.3 The Shape Retrieval Contest 36
2.4 Evaluation Criteria in 3DOR 43
2.5 Summary 46
References 46
Part II View Extraction, Selection, and Representation 50
Chapter 3: View Extraction 52
3.1 Introduction 52
3.2 Dense Sampling Viewpoints 52
3.3 Predefined Camera Array 53
3.4 Generated View 56
3.5 Summary 61
References 61
Chapter 4: View Selection 62
4.1 Introduction 62
4.2 Unsupervised View Selection 63
4.3 Interactive View Selection 65
4.3.1 Multiview 3-D Object Matching 66
4.3.2 View Clustering 68
4.3.3 Initial Query View Selection 68
4.3.4 Interactive View Selection with User Relevance Feedback 70
4.3.5 Learning a Distance Metric 71
4.3.6 Multiple Query Views Linear Combination 73
4.3.7 The Computational Cost 75
4.4 Summary 75
References 76
Chapter 5: View Representation 78
5.1 Introduction 78
5.2 Shape Feature Extraction 79
5.2.1 Zernike Moments 79
5.2.2 Fourier Descriptor 80
5.3 The Bag-of-Visual-Features Method 81
5.3.1 The Bag-of-Visual-Words 81
5.3.2 The Bag-of-Region-Words 82
5.4 Learning the Weights for Multiple Views 85
5.4.1 K-Partite Graph Reinforcement 86
5.4.2 Weight Learning for Multiple Views Usingthe k-Partite Graph 90
5.5 Summary 92
References 92
Part III View-Based 3-D Object Comparison 96
Chapter 6: Multiple-View Distance Metric 98
6.1 Introduction 98
6.2 Fundamental Many-to-Many Distance Measures 99
6.3 Bipartite Graph Matching 101
6.3.1 View Selection and Weighting 102
6.3.2 Bipartite Graph Construction 104
6.3.3 Bipartite Graph Matching 104
6.4 Statistical Matching 104
6.4.1 Adaptive View Clustering 105
6.4.2 CCFV 105
6.4.2.1 View Clustering and Query Model Training 106
6.4.2.2 Positive and Negative Matching Models 107
6.4.2.3 Calculation of the Similarity Between Q and O S(Q,O) 108
6.4.2.4 Analysis of Computational Cost 110
6.4.3 Markov Chain 111
6.4.4 Gaussian Mixture Model Formulation 112
6.4.4.1 Conventional GMM Training 112
6.4.4.2 Generative Adaptation of GMM 114
6.4.4.3 Discriminative Adaptation of GMM 115
6.4.4.4 Learning the Weights for Multiple GMMs 117
6.5 Summary 119
References 119
Chapter 7: Learning-Based 3-D Object Retrieval 122
7.1 Introduction 122
7.2 Learning Optimal Distance Metrics 123
7.2.1 Hausdorff Distance Learning 123
7.2.2 Learning Bipartite Graph Optimal Matching 128
7.3 3-D Object Relevance Estimation via Hypergraph Learning 132
7.3.1 Hypergraph and Its Applications 133
7.3.2 Learning on Single Hypergraph 135
7.3.3 Learning on Multiple Hypergraphs 139
7.3.4 Learning the Weights for Multiple Hypergraphs 141
7.3.5 Learning the Weights for Edges 143
7.4 Summary 145
References 145
Part IV Conclusions and Future Work 148
Chapter 8: Conclusions and Future Work 150
8.1 Summary of This Book 150
8.2 Future Work 151
8.2.1 The Issue of Big Data 151
8.2.2 Feature Extraction 151
8.2.3 Multiple-View Matching 152
8.2.4 Multimodal Data 152
8.2.5 Geographical Location-Based Applications 152
References 153

Chapter 2

The Benchmark and Evaluation


Abstract


In all research areas, standard benchmarks are important for justifying the performance of different methods. With the progress of recent 3-D technology, many 3-D benchmarks have been released. In this chapter, we introduce the widely employed 3-D benchmarks and the 3DOR contest, that is, the Shape Retrieval Contest (SHREC). In the last part of this chapter, we discuss widely used 3DOR retrieval performance evaluation methods.

Keywords

3-D Benchmark

Data example

Data distribution

Shape retrieval contest

Evaluation criteria

2.1 Introduction


Given the array of available applications, 3-D shape representation and matching and retrieval algorithms have been extensively investigated. These methods require standard benchmarks to evaluate their performance. Standard benchmarks are important for justifying the effectiveness of different method, and many 3-D object benchmarks have been released in recent years.

In this chapter, we review several popular 3-D benchmarks, including the Princeton Shape Benchmark (PSB) [1], the Engineering Shape Benchmark (ESB) [2], the ITI database [3], the Amsterdam Library of Object Images (ALOI) [4], the Eidgenössische Technische Hochschule (ETH) [5], and the National Taiwan University (NTU) [6] 3-D data sets. We also introduce the recent 3DOR contest, SHREC, on each year, respectively.

Evaluation methods are important to judge the performance of 3DOR. In the last part of this chapter, we introduce several widely used 3DOR retrieval performance evaluation criteria.

2.2 The Standard Benchmarks


In this section, we introduce six standard benchmarks.

 Princeton shape benchmark
PSB was collected by Princeton University. PSB contains 1814 polygonal 3-D models, which are collected from the World Wide Web from 293 different Web domains. All these models are manually classified based on functions and forms and multiple semantic labels are annotated for each 3-D model. These labels belong to a hierarchical structure, which can reflect the semantic in different levels. For example, a 3-D model with an annotation of “aircraft” can be further divided into subclasses, such as “winged_vehicle aircraft,” “balloon_vehicle aircraft,” and “helicopter aircraft.” PSB has been split into a training set and a testing set. The training set contains 907 models, and the testing set contains the other 907 models. The distribution of all the models in PSB is shown in Figures 2.1 and 2.2. Example 3-D models from PSB are shown in Figure 2.3.

Figure 2.1 The distribution of different model categories in the PSB training set.
Figure 2.2 The distribution of different model categories in the PSB testing set.
Figure 2.3 Example 3-D models in PSB.

 Engineering shape benchmark
ESB focuses on engineering shape representation. Computer-aided design product design is an iterative process, which is highly dependent on previous model designs. ESB aims to evaluate the shape representation to discriminate different shape forms for product design. ESB is composed of 867 3-D CAD models. For each 3-D model, there are CAD files, including two neutral formats: “.stl” and “.obj”, and one thumbnail image. These models have been classified into three higher-level categories: Flat-thin wall components, Rectangular-cubic prism, and Solids of revolution. As introduced in [2], these three higher-level categories are defined as follows:

 Flat-thin wall components are the parts with thin-walled sections and shell-like components.

 Rectangular-cubic prism are the parts whose envelopes are largely rectangular or cubic prism.

 Solids of revolution are the parts whose envelopes are largely solids of revolution.

These higher-level classes are further divided into 45 shape categories. The distribution of all the models in ESB is shown in Figure 2.4. Example 3-D models from ESB are shown in Figure 2.5.

Figure 2.4 The distribution of different model categories in the ESB.
Figure 2.5 Example 3-D models in ESB.

 The ITI benchmark
The ITI data set is from the VICTORY 3-D search engineer [7]. The ITI data set contains 544 3-D models from 13 categories, including animals, spheroid objects, conventional airplanes, delta airplanes, helicopters, cars, motorcycles, tubes, couches, chairs, fish, humans, and some other models. Each 3-D model contains 5,080 vertices and 7,061 triangles on average. The distribution of all the models in the ITI data set is shown in Figure 2.6.

Figure 2.6 The distribution of different model categories in the ITI data set.

 Amsterdam Library of Object Images
ALOI is a color image collection containing 1,000 small objects. We note that this data set does not provide the virtual 3-D model data, but only a set of images for each object. To obtain these images, the camera view angle, the illumination angle, and the illumination color for each object are systematically varied. For the camera view angle, the frontal camera is rotated in the plane at 5° to record 72 views of each object. For the illumination angle, there are eight different light conditions employed in the image data collection procedure, and five lights are employed. Each object is recorded with only one light turned on (five kinds in total), with only the left or right lights turned on (two kinds), and with all the lights turned on (one kind). Three cameras are employed to capture the views, which leads to a total of 24 different illumination conditions for recording views. The voltage of the lamps is varied 12 times for the illumination color, which yields 12 different illumination colors. Wide-baseline stereo images for 750 different scenarios are also recorded. In total, there are 110,250 images in the ALOI data set. Example objects from the ALOI benchmark are shown in Figure 2.7.

Figure 2.7 Example objects in the ALOI benchmark.

 The Eidgenössische Technische Hochschule Zürich data set
The ETH benchmark contains 80 objects from eight categories and each category consists of 10 objects. These categories cover the following areas: fruits and vegetables, animals, small human-made objects, and large human-made objects. These eight categories include apple, pear, tomato, cow, dog, horse, cup, and car. In the ETH data set, a group of 41 views are recorded for each object. To capture these views, the object is placed on a table with blue chroma keying background and the views are obtained from directions spaced equally over the upper viewing hemisphere, subdividing an octahedron to the third recursion level. These views are recorded using a Sony DFW-X700 progressive scan digital camera at a resolution of 1024 × 768. For each view, the original color image and an associated high-quality segmentation mask are provided. Example objects from the ETH benchmark are shown in Figure 2.8.

Figure 2.8 Example objects in the ETH benchmark.

 The National Taiwan University data set
NTU contains two parts, the NTU 3-D model benchmark and the NTU 3-D model database. The NTU 3-D model benchmark consists of 1833 3-D models and the NTU 3-D model database consists of 10,911 3-D models. All these 3-D models are downloaded from the Internet in July 2002. These models are in Wavefront file format. For each model, a thumbnail image is provided in the data set. The NTU 3-D model benchmark contains 47 categories with 549 3-D models, while another 1284 models are regarded as miscellaneous. The distribution of all models in the NTU 3-D model benchmark is shown in Figure 2.9, and example objects from the NTU 3-D model benchmark are shown in Figure 2.10.

Figure 2.9 The distribution of different model categories in the NTU benchmark.
Figure 2.10 Example objects in the NTU benchmark.

2.3 The Shape Retrieval Contest


SHREC [8] is an annual contest that started from 2006. SHREC aims to evaluate the effectiveness of 3-D shape retrieval algorithms. Each year, different tracks are organized with various objectives and different testing benchmarks are employed or released. In this section, we briefly review the existing SHREC events, the employed benchmarks, and the challenging tasks.

 SHREC2006
SHREC2006 is the first SHREC and there is only one task: retrieving a ranking list of similar 3-D objects given one...

Erscheint lt. Verlag 4.12.2014
Sprache englisch
Themenwelt Informatik Grafik / Design Digitale Bildverarbeitung
ISBN-10 0-12-802623-5 / 0128026235
ISBN-13 978-0-12-802623-6 / 9780128026236
Haben Sie eine Frage zum Produkt?
PDFPDF (Adobe DRM)
Größe: 8,7 MB

Kopierschutz: Adobe-DRM
Adobe-DRM ist ein Kopierschutz, der das eBook vor Mißbrauch schützen soll. Dabei wird das eBook bereits beim Download auf Ihre persönliche Adobe-ID autorisiert. Lesen können Sie das eBook dann nur auf den Geräten, welche ebenfalls auf Ihre Adobe-ID registriert sind.
Details zum Adobe-DRM

Dateiformat: PDF (Portable Document Format)
Mit einem festen Seiten­layout eignet sich die PDF besonders für Fach­bücher mit Spalten, Tabellen und Abbild­ungen. Eine PDF kann auf fast allen Geräten ange­zeigt werden, ist aber für kleine Displays (Smart­phone, eReader) nur einge­schränkt geeignet.

Systemvoraussetzungen:
PC/Mac: Mit einem PC oder Mac können Sie dieses eBook lesen. Sie benötigen eine Adobe-ID und die Software Adobe Digital Editions (kostenlos). Von der Benutzung der OverDrive Media Console raten wir Ihnen ab. Erfahrungsgemäß treten hier gehäuft Probleme mit dem Adobe DRM auf.
eReader: Dieses eBook kann mit (fast) allen eBook-Readern gelesen werden. Mit dem amazon-Kindle ist es aber nicht kompatibel.
Smartphone/Tablet: Egal ob Apple oder Android, dieses eBook können Sie lesen. Sie benötigen eine Adobe-ID sowie eine kostenlose App.
Geräteliste und zusätzliche Hinweise

Buying eBooks from abroad
For tax law reasons we can sell eBooks just within Germany and Switzerland. Regrettably we cannot fulfill eBook-orders from other countries.

EPUBEPUB (Adobe DRM)

Kopierschutz: Adobe-DRM
Adobe-DRM ist ein Kopierschutz, der das eBook vor Mißbrauch schützen soll. Dabei wird das eBook bereits beim Download auf Ihre persönliche Adobe-ID autorisiert. Lesen können Sie das eBook dann nur auf den Geräten, welche ebenfalls auf Ihre Adobe-ID registriert sind.
Details zum Adobe-DRM

Dateiformat: EPUB (Electronic Publication)
EPUB ist ein offener Standard für eBooks und eignet sich besonders zur Darstellung von Belle­tristik und Sach­büchern. Der Fließ­text wird dynamisch an die Display- und Schrift­größe ange­passt. Auch für mobile Lese­geräte ist EPUB daher gut geeignet.

Systemvoraussetzungen:
PC/Mac: Mit einem PC oder Mac können Sie dieses eBook lesen. Sie benötigen eine Adobe-ID und die Software Adobe Digital Editions (kostenlos). Von der Benutzung der OverDrive Media Console raten wir Ihnen ab. Erfahrungsgemäß treten hier gehäuft Probleme mit dem Adobe DRM auf.
eReader: Dieses eBook kann mit (fast) allen eBook-Readern gelesen werden. Mit dem amazon-Kindle ist es aber nicht kompatibel.
Smartphone/Tablet: Egal ob Apple oder Android, dieses eBook können Sie lesen. Sie benötigen eine Adobe-ID sowie eine kostenlose App.
Geräteliste und zusätzliche Hinweise

Buying eBooks from abroad
For tax law reasons we can sell eBooks just within Germany and Switzerland. Regrettably we cannot fulfill eBook-orders from other countries.

Mehr entdecken
aus dem Bereich
Explore powerful modeling and character creation techniques used for …

von Lukas Kutschera

eBook Download (2024)
Packt Publishing (Verlag)
43,19
Discover the smart way to polish your digital imagery skills by …

von Gary Bradley

eBook Download (2024)
Packt Publishing (Verlag)
45,59
Generate creative images from text prompts and seamlessly integrate …

von Margarida Barreto

eBook Download (2024)
Packt Publishing (Verlag)
32,39