This work focuses on the search of a sample object (car) in video sequences and images based on shape similarity. We form a new description for cars, using relational graphs in order to annotate the images where the object of interest (OOI) is present. Query by text can be performed afterward to ex
Illumination Invariance and Object Model in Content-Based Image and Video Retrieval
✍ Scribed by Ze-Nian Li; Osmar R Zaı̈ane; Zinovi Tauber
- Publisher
- Elsevier Science
- Year
- 1999
- Tongue
- English
- Weight
- 722 KB
- Volume
- 10
- Category
- Article
- ISSN
- 1047-3203
No coin nor oath required. For personal study only.
✦ Synopsis
With huge amounts of multimedia information connected to the global information network (Internet), efficient and effective image retrieval from large image and video repositories has become an imminent research issue. This article presents our research in the C-BIRD (content-based image retrieval in digital-libraries) project. In addition to the use of common features such as color, texture, shape, and their conjuncts, and the combined content-based and description-based techniques, it is shown that (a) color-channel-normalization enables search by illumination invariance, and (b) feature localization and a three-step matching algorithm (color hypothesis, texture support, shape verification) facilitate search by object model in image and video databases.
📜 SIMILAR VOLUMES
We describe a method for computing an image signature, suitable for contentbased retrieval from image databases. The signature is extracted from the Fourier power spectrum by performing a mapping from cartesian to logarithmic-polar coordinates, projecting this mapping onto two 1D signature vectors,
## Abstract In content‐based image retrieval, relevance feedback is an interactive process, which builds a bridge to connect users with the search engine. It leads to much improved retrieval performance by updating the query and the similarity measure according to a user's preference; and recently
Multimedia data types, such as image and video, are structurally more complex than traditional data types. We view an image as a compound object containing many sub-objects. Each sub-object corresponds to image regions that are visually and semantically meaningful (e.g., car, man, etc.). In this pap
We introduce a robust algorithm to recognize objects in 3D space from one 2D video image and to localize the objects in all six degrees of freedom. Point-like attached features are used in the input image and additional edge information provides grouping. In an initial phase, a 3D model of all objec