Efficient parallel texture classification for image retrieval

J. You, H. Shen, H. A. Cohen

Research output: Contribution to conferencePaperpeer-review

4 Citations (Scopus)


This paper proposes an efficient parallel approach to texture classification for image retrieval. The idea behind this method is to pre-extract texture features in terms of texture energy measurement associated with a 'tuned' mask and store them in a multi-scale and multi-orientation texture class database via a two-dimensional linked list for query. Thus each texture class sample in the database can be traced by its texture energy in a two-dimensional row sorted matrix. The parallel searching strategies are introduced for fast identifying the entities closest to the input texture throughout the given texture energy matrix. In contrast to the traditional search methods, our approach incorporates different computation patterns for different cases of available processor numbers and concerns with robust and work-optimal parallel algorithms for row-search and minimum-find based on the accelerated cascading technique and the dynamic processor allocation scheme. Applications of the proposed parallel search and multisearch algorithms to both single image classification and multiple image classification are discussed. The time complexity analysis shows that our proposal will speed up the classification tasks in a simple but dynamic manner. Examples are presented of the texture classification task applied to image retrieval of Brodatz textures, comprising various orientations and scales.

Original languageEnglish
Number of pages8
Publication statusPublished - 1997
Externally publishedYes
EventProceedings of the 1997 Conference on Advances in Parallel and Distributed Computing - Shanghai, China
Duration: 19 Mar 199721 Mar 1997


ConferenceProceedings of the 1997 Conference on Advances in Parallel and Distributed Computing
CityShanghai, China


Dive into the research topics of 'Efficient parallel texture classification for image retrieval'. Together they form a unique fingerprint.

Cite this