Color and grey level object retrieval using a 3D representation of force histogram
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
Image and Vision Computing 21 (2003) 483–495 www.elsevier.com/locate/imavis Color and grey level object retrieval using a 3D representation of force histogram Salvatore Tabbone, Laurent Wendling* LORIA, Campus Scientifique, BP 239, Vandoeuvre-les-Nancy Cedex 54-506, France Received 20 July 2001; received in revised form 14 January 2003; accepted 20 January 2003 Abstract A new method for both grey level and color object retrieval is presented in this paper. Our feature is based on previous works on force histogram notion which is extended here to handle with photometric information. This kind of feature has low computing time and allows keeping fundamental geometric transformations as scale, translation, symmetry and rotation. More precisely objects processed by defining a tridimensional signature which takes into account their photometric variations and their shapes. Experimental results show the promising aspect of our approach. q 2003 Published by Elsevier Science B.V. Keywords: Photometric objects; Object matching; Force histogram; Tridimensional signature 1. Introduction 2. Related woks A new approach for the content-based retrieval problem More and more content-based image retrieval systems is presented in this paper. Our feature is based on force have been developed in the early years [29]. Full overviews histogram introduced by Matsakis and Wendling [18,19] concerning the indexing problem can be found in the which allows keeping fundamental geometric transform- literature [1,11,29]. Some systems allow querying by ations as scale factor, translation, rotation and symmetry. It keywords, assuming that each image has been manually has been shown in previous works that such a feature is a indexed when inserted into the database. Others systems powerful tool to define spatial relations [18,19] between often use one or more features to describe the content of objects and also to describe robust signatures for binary images in order to provide an automatic content-based objects [19,35]. In this paper we show that such a notion can retrieval process. We give here a succinct description of be extended to retrieve both grey level and color objects by some content-based approaches close to our research field. including photometric variations in the force histogram. Generally a similarity measure based on features is used to First, a brief overview of content-based retrieval approaches help the user to find specific images in the database. is given in Section 2. Then, the definition of a force Typically, the indexing scheme makes use of color, texture, histogram and its main properties are recalled in Section 3. edge and shape information. The calculation of F3D-signatures on grey level and color Some approaches focus on interest points [2,4,22,28] to objects is presented in Section 4. The measure of similarity compute different local characteristics of each image. between two F3D-signatures is given in Section 5. Finally, Others use texture information [6,16,17,27]. Typically, some experimental studies using image databases and a Gabor filters [34] or Haar wavelet transform coefficients discussion about the advantages and the limits of our [15] are well suited to compute these kinds of features. Early approach are provided in Section 6. systems were based on color histograms [12,24,30,33] or on * Corresponding author. shape information [3,20,32]. Several content-based image E-mail address: wendling@loria.fr (L. Wendling); tabbone@loria.fr retrieval systems merge different ways (that is encapsulation (S. Tabbone). of features) to describe images [14,23,25,26,31]. These 0262-8856/03/$ - see front matter q 2003 Published by Elsevier Science B.V. doi:10.1016/S0262-8856(03)00016-7
484 S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 Fig. 1. Definition of a F-signature. systems give generally better results than the ones using histogram obtained in this way can be noisy and strongly only one attribute. dependent on the angle threshold step of the histogram. To In this paper we address the issue of indexing images overcome this problem, the use of straight lines following a set based on visual content. We focus on approaches that use of directions u is considered, and the method handles with low-level features. A new approach dealing with grey level segments, in order to decrease the processing time (see Fig. 1). and color objects is proposed in this paper. Our feature is The extracted forces are calculated between segments—i.e. based on the force histogram notion which is extended here the attraction force of one segment with respect to another to handle with photometric information. Such a feature is segment is computed. Let I1 and I2 be two segments carried by isotropic and requires low time complexity. It describes a the same straight line, let lI1 l and lI2 l be their respective 3D-signature keeping nice geometric properties as scale, lengths, and let DuI1 I2 be the distance between I1 and I2 : The rotation and translation. attraction force from I1 to I2 is given by: ðlI2 l ðlI1 lþDuI1 I2 þlI2 l fr ðlI1 l; DuI1 I2 ; lI2 lÞ ¼ wr ðu 2 vÞdv du ð2Þ 3. Signature of binary object 0 DuI þlI2 l 1 I2 3.1. Description of the method For example, the calculation of f0 corresponds to lI1 l £ lI2 l when I precedes J and to lI1 l2 =2; when I1 and I2 are In this section, the scheme for computing the F-signature superimposed; the calculation of f2 when I1 precedes I2 ; is of a binary graphical object, which is a particular histogram of given by forces [18,19,35] is recalled. A histogram of forces can be ðDuI1 I2 þ lI1 lÞðDuI1 I2 þ lI2 lÞ assumed to be the calculation of all the forces exerted between ln : DuI1 I2 ðDuI1 I2 þ lI1 l þ ðDuI1 I2 þ lI2 lÞÞ the pixels of a same object. Let wr be the map from R into Rþ, null on R2 and continuous on Rpþ, such that: Let A be an object of the plane which is entirely described by a ;d [ Rpþ ; wr ðdÞ ¼ 1=d r ð1Þ pencil of parallel straight lines of angle u from the orthogonal frame of the image. Let Duh be a line; the set of segments of A Let a1 and a2 be two points of R , and d be the distance between 2 beared by this straight line corresponds to Au ðhÞ ¼
S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 485 Fig. 3. F3D-signatures of previous butterflies. All the pencils of lines Dhu of the plane which entirely describe † Translation. The information of the object is processed A are then considered: independently of its location in the frame of the image. Let Tu~ be a translation of vector u~ : ðþ1 F AA l R ! Rþ ; u 7! Fðu; Au ðvÞ; Au ðvÞÞdv ð4Þ F AA ðuÞ ¼ F Tu~ ðAÞTu~ ðAÞ ðuÞ ð5Þ 21 In the discrete case, the calculation of F AA ðuÞ provides an † Symmetry. The forces exerted on an object are the same evaluation of the forces exerted by the object on itself in following two opposite directions: the direction u: The calculation of F AA following the set of angles ui ðui [ ½2p; þpÞ defines the FAA 2D signature of the F AA ðuÞ ¼ F AA ðp 2 uÞ ð6Þ object A: † Scale factor. Only the shape of the F-signatures is 3.2. Geometric properties considered. If a scale factor is applied to an object, its histogram will be stretched. In such a case, the forces are The aim of our study is to classify similar objects multiplied by a value that depends or r and on the scale independently of their location, their orientation and their factor: size. So it is important to have a feature able to take into account fundamental geometric properties as translation, fr ðklIl; kDuIJ ; klJlÞ ¼ k22r fr ðlIl; DuIJ ; lJlÞ ð7Þ rotation and scale factor. By definition, in the continuous case, the properties of F allow us to find easily such † Rotation. When a rotation is applied to an object, the geometric transformations [18,19]: histogram is shifted, as the approach is isotropic. Let u be Fig. 4. Visualization of geometric properties.
486 S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 Fig. 5. Transformation of a color image into a matrix composed of F3D-signatures. Fig. 6. F3D-signatures of butterflies.
S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 487 Fig. 7. Nine nearest images of pap28_1 contained in the database. Fig. 8. Nine nearest images of pap14_1 contained in the database. Fig. 9. Similarity ratio between the signatures of pap10_1 and another insect.
488 S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 Fig. 10. Comparison with other approaches—butterfly database—. a direction and r be a rotation of angle u0 applied to an object A; we have: 0 0 F AA ðuÞ ¼ F rðA;u ÞrðA;u Þ ðu þ u0 Þ ð8Þ These properties are also valid in the discrete case, with weak error variations, due to both histogram and sampling steps. 4. F3D-signature 4.1. Handling with photometric objects 4.1.1. Usefulness Fig. 11. Measure of quality for the eight main clusters of butterfly. It is possible to integrate the photometric variations shapes but in different areas of intensities. For example, the during the definition of a F2D-signature. In this case the butterflies presented Fig. 2 have close signatures. grey level difference between two consecutive cuts should A new dimension has been added to improve the be processed. Actually two kinds of schemes may be used to classification and to better take into account the variations handle the level cuts ai using force histogram [18]. The first directly derives from the generic double sum scheme defined by Dubois and Jaulent [8] and has a high processing time. We have then used the simple sum scheme, proposed by Khrisnapuram et al. [13] which decreases the computing time from quadratic to linear, that is: X #g ai A ai F AA ¼ mi F A ð9Þ i¼1 with #g the number of grey levels, mi the difference between ai ai two consecutive grey level cuts ðai 2 ai21 Þ and F A A the signature of the binary object obtained from the level cut ai : Generally, a bidimensional signature is accurate because it takes care of the variations of grey levels in the image. Nevertheless, as only the shape of the histogram is studied, it is difficult to discriminate two objects having similar Fig. 12. Measure of quality for all the database.
S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 489 Fig. 13. F-signatures of four objects (view 08). of photometry following the set grey level cuts. The processing due to the computation of force histograms signatures of the previous butterfly samples are then better between planes. differentiated by adding a third dimension to the signature As in Refs. [9,33] the new plane r; computed from R was (see Fig. 3). defined as follows: R rðR; G; BÞ ¼ 4.1.2. Definitions RþGþB It is easy to show that a F3D-signature verifies the The same processing is carried out for the other configur- fundamental geometric properties described in Section ations gðR; G; BÞ and bðR; G; BÞ: Then the F3D-signatures of 3.2. Let A be a photometric object and let {li }i¼1;n be a each normalized plane were computed to define the new strictly increasing series of grey levels ðl1 , l2 , · · · , gg diagonal of the matrix (see Fig. 5): Frr 3D ; F3D and F3D : bb ln Þ including in A: Let FAA li be the force histogram In order to take into account the link between the couple of calculated on binary objects from a binarization of planes ðRGÞ; ðRBÞ and ðBRÞ; a classical transformation [10] threshold li (if Aðx; yÞ . li then Aðx; yÞ ¼ 1 else was used. The link between the two planes R and G became: Aðx; yÞ ¼ 0). The set of included regions Aln # Aln21 # · · · # Al1 may be interpreted as another representation of ðR 2 GÞ2 lRG ¼ A: So, a F3D-signature is defined by an ordered set of ðR 2 GÞ2 þ ðR 2 BÞ2 þ ðG 2 BÞ2 bidimensional signatures (checking separately the proper- The three news planes (lRG ; lRB and lBG ) describe the link of ties given in Section 3.2): FAA3D ¼
490 S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 Fig. 15. Fourteen nearest images of view 0 of object 53. Fig. 5 shows an example of color object and its associated (cf. Section 3.2) can be obtained using the histogram of F3D-matrix. forces. Each F3D-signatures is normalized by its volume to take into account the scale factor between objects of a 4.3. Object retrieval same cluster. Let A and B be two objects, F AA BB 3D and F3D be their 4.3.1. Distance between signatures associated normalized signature and p be the number of A representation of the object in the same unitary frame directions processed. We set F AA AA 3D ¼
S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 491 Fig. 17. Example of F3D-signatures for progressive views of a same object. 4.1). A criteria of similarity S is maximized following all time to achieve consists results. We can remark that the the histogram shifts to obtain an approximation of the calculation of FS is a generalization of S: rotation factor: ( Xp ) 4.3.2. Complexity of the method u¼0 m ðu; xÞ AB S ¼ max X ð10Þ The timepffifficomplexity ffi for computing a histogram of forces p u¼0 M ðu; xÞ AB x[½0;p is in Oðpm mÞ; where m is the number of pixels of the image P and p the number of digitalization steps of the histogram with: mAB ðu;PxÞ ¼ ni¼1 minðF AA BB li ðuÞ; Fli ðu þ xÞ%pÞÞ and AA BB [18]. This bound is reached for very highly textured images. M ðu; xÞ ¼ i¼1 maxðF AB n l ðuÞ; F i l ðu þ xÞ%pÞÞ: i In our approach, the complexity is in OðpmÞ when using The value of S is not sensitive to object scaling because constant force histograms (r ¼ 0 in Eq. (1)), and also in the matched signatures are normalized. Moreover circular most cases when using other force histograms ðr – 0Þ: shifts are applied to F AA 3D and S is obtained by maximizing For efficiency reasons, Bresenham’s algorithm [5] is used the classical Tanimoto index (min over max). Then S is not in our implementation, as it is a fast method which sensitive to object rotation either. minimizes error in drawing lines on integer grid points. It The same scheme is applied to process color objects. approximates line segments defined by rational coefficients Based on color information, normalization is split into two using only integral points. Bresenham’s algorithm is applied cases: the diagonal and the other cells of the matrix in order only to one line, to store the horizontal and vertical shifts to separate the characteristic signatures of the plane and representing a line scanning the image following the between the planes interaction. The final score FS is given direction u: These shifts are used to define the pencil of by: lines allowing to scan the whole image (with this approach 8 Xp X3 X3 9 and for a given orientation u; any pixel is only processed by 1 < mAij Bij ð u ; xÞ = u¼0 i¼1 j¼1 one line). The p directions are then processed to define the FS ¼ max Xp X3 X3 ð11Þ 2 x[½0;p : M Aij Bij ðu; xÞ ; F-signature. u¼0 i¼1 j¼1 The complexity, using grey or color images, is generally where i; j are the coordinates of the signature in the matrix. OðnpmÞ (with n the number of grey levels considered). To The shifts are performed for all the signatures in the same decrease the processing time no binarization was performed Fig. 18. Comparison with other approaches—coil database—.
492 S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 the results provided are right and both the scale and the rotation factors are taken into account by our approach. The robustness of our approach to a specific occultation is shown in another example presented in Fig. 8: pap14_4 and pap14_5 represents a folded butterfly belonging to the same cluster. Moreover the less classified butterflies in Figs. 7 and 8 are dissimilar in term of photometry but are similar in shapes (from pap19_1 to pap15_3 in Fig. 7 and from pap16_1 to pap68_1 in Fig. 8). Fig. 9 shows the signatures of a butterfly and another insect. The score obtained guarantees that the matched object is really far. We have also used the experimental studies proposed by Derrode [7] to make a comparative study. Fig. 10 presents the nearest butterflies of pap15_2 using, respectively, Hue Fig. 19. Measure of quality r—coil-100, pgm—. (a) One view, (b) two invariants (HI), Fourier-Mellin invariants (FMI) and F3D- views, (c) four views per object. signatures matching (F3D). A value near 0 corresponds to a but an accumulator table has been used to manage with grey good retrieval for HI and FMI and near 1 for our approach. levels. Then all the processings have been carried out in p We can note the presence of the butterfly pap15_6 scans of the image. recognized by only our method which is more robust to the deformation. A classical measure of match quality [10] was applied to all the database to have an idea of the behavior of our 5. Experimental results approach: ! 1 XC X Ck k 5.1. Grey levels objects N 2 r Qi r ¼ ð12Þ N k¼1 i¼1 N 2 1 5.1.1. Butterfly database with C the number of clusters and Ck the number of samples We have used a database provided by Derrode et al. [7]. in cluster k: N is the number of images contained in the k It consists of around a hundred of images of butterflies database and the rank r Qi denotes the position of the match describing several clusters following different scales and (that is set to 1 for a correct match and to N in the worst orientations. Fig. 6 presents four types of butterflies with case) of the query image Qki (of cluster k) in an ordered list their associated F3D-signatures. of N match values. Fig. 7 gives the nine nearest images of butterfly pap28_1 Figs. 11 and 12 present, in term of percentage (Y-axis), and the scores reached. We can see that for each image the application of such a criteria over all the database of Fig. 20. Nine nearest images of view 0 of object 51.
S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 493 Fig. 23. Similar objects. was computed (curve X in Fig. 14) and the total number of views of the same object found among the 71 possible choices (curve Y in Fig. 14). Fig. 14 shows the results Fig. 21. Robustness to view-changes. obtained for each object using one view per object. An increasing sort was applied to the curve X to achieve a butterflies with increasing sizes of signatures (X-axis: grey better visualization of the results. levels £ directions). The behavior of the eight main clusters The minimum reach by the curve X is 4 and the are shown in Fig. 11 and the mean rate over all the database maximum is 71. The high score obtained and the closeness is given Fig. 12. The results provided show the robustness of of the two curves attest of the robustness of our approach to our approach. We can remark that a signature having a size the change of view-points. 8 £ 8 is enough for such an application. Fig. 15 presents the 14 nearest views of an object well recognized ðX ¼ 33Þ and the similarity ratio reached. The first number corresponds to the identifier of the object 5.1.2. Coil-100 database (here 53) and the second to the rotation angle in degree (here The robustness of our approach to a change of views, 0 for the first view). The different views show that the has been tested using the well-known coil-100 database invariance by symmetry is considered by our approach. [21]. Such database contains 100 objects of varying Fig. 16 shows a case of low identification towards the shapes following 72 views of 58 apart. We consider in change of view-points. Only a series of four images has been this section grey level representations of these views. identified. Nevertheless the object 12 is in the same area of Fig. 13 shows a view of four objects and the associated grey levels and has a close shape. F3D-signatures. Fig. 17 presents the signatures of the views of the same The signature of a view of each object of the database 3D objects with weak rotations (with steps of 58). Shapes of has been matched to the 7199 remaining views. The these signatures are close because the associated views are nearest views of a particular view of each object were close. Obviously, when the rotation applied is important, the computed. As an object is represented following 72 signatures are different because the differences between the views, the 71 nearest views, in term of similarity, were shapes of the objects increase. kept. For each object the number or views rightly As previously, we have compared our method to both classified without any view of another objects inserted Fourier-Mellin and Hue invariants. The results provide in Fig. 18 shows that our method is more robust to the change of view-points and has a better behavior using photometric data. Such a result is also underlined by the computation of the predefined measured of quality r with Ck ¼ 72 and N ¼ 7200 (see Eq. (12)). Three curves are proposed (Fig. 19): an object describes by only one view (a), an object is assumed to two opposite views—0 and p—and the maximum of the two final similarities reached is considered (b), from four views—following the cardinal directions—(c). 5.2. Color images In this section, the coil-100 database was used again to Fig. 22. Measures of quality r (coil-100, color). (a) One view, (b) two determine if the color brings an improvement during the views, (c) four views per object. identification step. We have also used to database of around
494 S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 Fig. 24. Nearest image to a query photo (upper-left corner). Fig. 25. Other example from another query photo (upper-left corner). 2500 broad images to study the behavior of our approach in F3D-signatures, following successive variations is currently a real case. under consideration to improve the classification. Fig. 23 shows two views of objects having similar shapes and having low disparities in photometry. The use of color 5.2.1. Color objects information is obviously more discriminant than the one of Coil-100 database has been used to test if the matching of grey levels. Such a result shows that the shape in the F3D-signatures of color objects brings an advance com- signature predominates when the photometric variations pared to grey level objects. So the object previously between two images are low. presented in Fig. 16 having a low identification score was studied again. The new result are given in Fig. 20 (X ¼ 11 5.2.2. Color photos and Y ¼ 23). The similarity ratios are greater than those We have also tested our method with a database presented in Fig. 16. composed of around 2500 broad images. Figs. 24 and 25 The rates of robustness to the change of view-points are present two results reached using our approach. The images also improved (Fig. 21). The curves X and Y are nearer and achieved were close in term of photometry and shapes. the number of classified views is greater. For example the 71 Other results have shown that our method provides views of 27 objects have been full recognized. That is interesting results n several series of broad images. also attested by the computation of r on the color database (Fig. 22 with X-axis: planes £ grey levels £ directions), using one view (a), two views (b) and four views (c). 6. Conclusion It is obvious that full recognized objects have similar shape following any view. Such a result is weak, without We have shown in this paper that an F3D-signature learning steps, using objects having various shapes follow- allows to reach a fast and robust retrieval using both color ing their tridimensional views. For example the views of the and grey level objects. Preliminary tests were performed width and the length of a car are different. The learning of on color photos. Currently, we search to extend our
S. Tabbone, L. Wendling / Image and Vision Computing 21 (2003) 483–495 495 approach to the main regions of an image. The study of [16] B. Manjunath, W. Ma, Texture features for browsing and retrieval of fast methods of ‘rough’ segmentation of the image is image data, IEEE Transactions on PAMI 18 (8) (1996) 837–842. [17] R. Manmatha, S. Ravela, Y. Chitti, On computing local and global under consideration. The main connected regions obtained similarity in images, Proceedings of the SPIE Conference on Human will be used like masks and the F3D-signatures computed Vision and Electronic Imaging III, San Jose, CA 3299 (1998) 540–551. on them will be matched with the database. Such an [18] P. Matsakis, L. Wendling, A new way to represent the relative approach should better retrieve images having main position between areal objects, IEEE Transactions on PAMI 7 (21) objects or landscape images. At last, we plan to (1999) 634–643. [19] P. Matsakis, Structural spatial relations and image understanding, PhD approximate our F3D-signatures by continuous functions thesis (in French), University Paul Sabatier, France, 1998. to decrease their cardinality and to better take into [20] R. Mehrotra, J.E. Gary, Similar-shape retrieval in shape data account their properties. management, IEEE Computer 28 (9) (1995) 57 –62. [21] S.K. Nayar, S.A. Nene, H. Murase, Real-time 100 object recognition system, Proceedings of the ARPA Image Understanding Workshop References (1996). [22] C. Nastar, M. Mitschke, C. Meilhac, N. Boujema, H. Bernard, M. Mautref, Retrieving images by content: the surfimage system, [1] S. Antani, R. Kasturi, R. Jain, A survey on the use of pattern Multimedia Information System, Istanbul, Turkey (1998). recognition methods for abstraction, indexing and retrieval of images [23] W. Niblack, R. Barber, W. Equitz, M. Flickner, E. Glasman, D. and video, Pattern Recognition 35 (2002) 945–965. Petkovic, P. Yanker, C. Faloutsos, G. Taubin, The QBIC project: [2] S.K. Bhattacharjee, A computational approach to image retrieval, PhD querying images by content using color, texture and shape, in: W. Thesis number 2002, EPFL, Lausanne, Switzerland, 1999. Niblack (Ed.), Storage and Retrieval for Image and Video Databases, [3] J. Bigun, S.K. Bhattacharjee, S. Michel, Orientation radiograms for SPIE, San Jose, CA, 1993, pp. 173–181. image retrieval: an alternative to segmentation, IEEE International [24] V.E. Ogle, M. Stonebraker, CHABOT: retrieval from a relational Conference on Pattern Recognition, Vienna, Austria C (1996) database of images, IEEE Computer 28 (9) (1995) 40–48. 346–350. [25] Y. Park, F. Ghoslani, ImageRoadMap: a new content-based image [4] S. Bres, J.M. Jolion, Detection of interest point for image indexation, retrieval system, Proceedings of the Eighth DEXA, Toulouse, France Third International Conference on Visual Information Systems, (1997) 225–239. Amsterdam June (1999) 427–434. [26] A.P. Pentland, R.W. Picard, S. Scaroff, Photobook: tools for content- [5] J.E. Bresenham, Algorithm for computer control of a digital plotter, IBM Systems Journal 4 (1) (1965) 25–30. based manipulation of image-databases, Proceedings of the SPIE [6] Y.Q. Chen, M.S. Nixon, D. Thomas, Statistical geometrical features Conference on Storage and Retrieval of Image and Video Databases for texture classification, Pattern Recognition 28 (4) (1995) 537–552. II, San Jose, CA, USA 2158 (1994) 34 –46. [7] S. Derrode, M. Daoudi, F. Ghorbel, Invariant content-based image [27] R.W. Picard, T.P. Minka, Vision texture annotation, Journal of retrieval using a complete set of Fourier-Mellin descriptors, IEEE Multimedia Systems 3 (1995) 3–14. International Conference on Multimedia Computing and Systems 2 [28] C. Schmid, R. Mohr, Local greyvalue invariants for image retrieval, (1999) 877–881. IEEE Transactions on PAMI 19 (5) (1997) 530–535. [8] D. Dubois, M.C. Jaulent, A general approach to parameter evaluation [29] A.W.M. Smeulders, M. Worring, S. Santini, A. Gupta, R. Jain, in fuzzy digital pictures, Pattern Recognition Letters 6 (1987) Content-based image retrieval at the end of the early years, IEEE 251–259. Transactions on PAMI 22 (12) (2000) 1349–1380. [9] T. Gevers, Color image invariant segmentation and retrieval, PhD [30] J.R. Smith, S.F. Chang, Single color extraction and image query, thesis, University of Amsterdam, The Netherlands, 1996. Proceedings of the IEEE International Conference on Image [10] T. Gevers, A.W.M. Smeulders, Context-based image retrieval by Processing, Washington,DC (1995) 528– 531. viewpoint-invariant color indexing, Image and Vision Computing 17 [31] J.R. Smith, S.F. Chang, Querying by color regions using the (1999) 475–488. visualseek content-based visual query system, in: M.T. Maybury [11] A.K. Jain, R.P.W. Duin, J. Mao, Statistical pattern recognition: a (Ed.), Intelligent Multimedia Information Retrieval, AAAI Press, review, IEEE Transactions on PAMI 22 (1) (2000) 4–37. Menlo Park, 1997, pp. 23–41. [12] T. Kato, Database architecture for current based image retrieval, [32] R.K. Srihari, Use of multimedia input in automated image annotation Image Storage and Retrieval Systems, vol. 1662, SPIE, San Jose, CA, and content-based retrieval, in: W. Niblack, R.C. Jain (Eds.), Storage 1992. and Retrieval for Image and Video Databases, SPIE, San Jose, CA, [13] R. Krishnapuram, J.M. Keller, Y. Ma, Quantitative analysis of 1995, pp. 249–260. properties and spatial relations of fuzzy image regions, IEEE [33] M.J. Swain, D.H. Ballard, Color indexing, International Journal of Transactions on Fuzzy Systems 3 (1) (1993) 222 –233. Computer Vision 7 (1) (1991) 11 –32. [14] C.P. Lam, J.K. Wu, B. Mehtre, STAR—a system for trademark [34] T.P. Weldon, W.E. Higgins, Design of multiple gabor filters for archival and retrieval, Proceedings of the Second Asian Conference texture segmentation, IEEE ICASSP, Atlanta, GA IV (1996) on Computer Vision, Singapore 3 (1995) 214 –217. 2245– 2248. [15] T. Lonnestad, A new set of texture features based on the Haar [35] L. Wendling, S. Tabbone, P. Matsakis, Fast and robust recognition of transform, Proceedings of the 11th IAPR International Conference on orbit and sinus drawings using histograms of forces, Pattern Pattern Recognition, The Hague, Netherlands 3 (1992) 676–679. Recognition Letters 23 (14) (2002) 1687–1693.
You can also read