A Color Segmentation Algorithm to Annotate Fine Art Images


ºtÁ¿ºK­n

¥Øªº¡G

¥Î±m¦â°Ï°ì¤À¹jªº¤èªk¨Ó¿ë»{ÃÀ³Nªºµe¹³¡A¬Ý¬O¤@­Ó¤Hª«ªº¨v¹³¡AÁÙ¬O¤@­Ó¦ÛµM­·´ºªºµe¡C

°Ê¾÷¡G

¦bºô¸ô¤W¥i¥H¬Ý¨ì¦³web museum¡A¤]´N¬Oºô¸ô¤Wªº³Õª«À]¡A¸Ì­±¦³«Ü¦hªºµe¹³¡BÃÀ³N«~¡A¦³³o»ò¦hªºµe¹³­n«ç»ò¾ã²z©O¡H©Î¬O­n¦b¨ä¤¤§ä¨ì¤@´Tµe¡A¦ý¬O§A¤£ª¾¹D§@ªÌ¡A¥uª¾¹Dµe¤¤¦³­Ó¬ü¤k¦b·L¯º¡A¨º»ò¥u»Ý­n¥´¶i³o¼Ëªº·Qªk¡A³o­Ó¨t²Î´N¯àÀ°§A§ä¨ì³o´Tµe¡C¥t¥~¤@­Ó°Ê¾÷¤£¬OºÞ²z¬ü³NÀ]¡A¦Ó¬OºÞ²z¨ä¥LªºªF¦è¡A§Ú­Ì¥i¥H¥Î¦³Ãö¼v¹³ªº¸ê°T¨Ó³B²z¡B¤ÀªR³o­ÓªF¦è¡C³Ì«á§Æ±æ¹F¨ìªº¥\¯à¡A¬O¨Ï¥ÎªÌ¤£»Ý­n¤¤¶¡¥h¿ï¾Ü¡A¤ñ¦p«ö¶i­þ¨Ç¦r¡A¥¦´N·|¥X²{­þ¥»®Ñ¡F¦b¼v¹³¤è­±¡A§Ú­Ì§Æ±æ§A¯à°÷µ¹¥¦¤@­Ó½d¨Ò¡A¥¦´N¥i¥Hª½±µÀ°§A§ä¨ì¡A§A¤£»Ý­n¥h«ü¾É¹q¸£¥h§ä¨ì¥¦¡C

«e¤Hªº¬ã¨s¡G

¦b¸ê®Æ³B²z¤è­±¡A¦³¤H´£¥X¤@­Ómodel¡A¥D­n¬O·Q­n¿ë»{¦³¨v¹³¦bªºimage¡A¹ï©óportrait image¡A¤w¸g¦³¨â­Ó¤ñ¸û¦¨¥\ªº¬ã¨s¡C²Ä¤@­Ó¥L­Ì¬O¥ÎRGB¡A§Ú­Ìª¾¹D±m¦âspace¬ORGB¡A¦ý¬O§Ú­Ì¥i¥H¥ÎRGB¤T­Óspace´«¦¨¥ô¦ó¤@­Ócolor space¡C¨º»ò³o­Ó¤H¥L¬O§â¤jªºRGB space´«¦¨¤pªºrgbªºcolor space¡C³o­Ó¤H¬OMiyake¦b1990¦~¡A¦o¯à°÷§â§Ú­Ì¥Ö½§ªºÃC¦â§ä¥X¨Ó¡A¦]¬°portrait³Ì¦hªº´N¤Hªº¥Ö½§ªºÃC¦â¡C¦oµo²{¥Ö½§ªºÃC¦â¦brg³o¨â«×ªÅ¶¡ªºspace¤W­±¥i¥H§Î¦¨¤@­Ó¾ò¶ê§Î­±¡C²Ä¤G­Óµo²{¬O¦³¤H§âRGB´«¦¨YES¡AYES¬O¥t¥~¤@­Ócolor space¡C³o­Ó¤H¬OSaber¦b1996¦~´£¥Xªº¡A¥Lµo²{¦³¤T­Ó±Ú¸sªºÃC¦â¡A´N¬O¥Ö½§¡B¤ÑªÅ¡B©M¯ó¦aªºÃC¦â¦b"ES"³o­Ó¥­­±¤W·|§Î¦¨2-DªºGaussian functionªº¤À¥¬¡C

OUTLINE¡G

­º¥ý§Ú­Ì¿ï¾ÜTESªºcolor space¡A¤]´N¬ORGB¦A¸g¹L¤@­ÓmatrixªºÂà´«¤§«á¦¨YES¡C³o¨Ç«Y¼Æ«ç»ò¨Óªº©O¡H°ò¥»¤W¬O¸gÅ窺¡A³o¨ÇEngineers test«Ü¦h¤£¦P·Ó©úµ{«×ªº¹Ï¤ù¡AµM«á¨Ó¿ï¥X³o­Ómatrixªº«Y¼Æ¡A¦o­Ì¨M©w³o­Ó¥sYES¡CYES³£¦³¥¦ªº·N¸q¦b¡AY:¨Æ¹ê¤W´N¬O§â¤@­Ópixel¤¤RGBªº¦¨¤À©Ò¦³¦³Ãö«G«×ªº¦¨¤À§â¥¦®³¥X¨Ó¡A©Ò¥H¥sliminance¡CE©MS°ò¥»¤W´N§â¥¦¦â±mªº¦¨¤À¡A¦³ÃC®Æªº³¡¤À¨ú¥X¨Ó¡C¬°¤°»ò­n¿ï¾Ü³o­Ócolor space¡A¥¦ªº¦n³B¬O¡G¤@¡B¯à°÷§âluminanceª½±µ®³¥X¨Ó¡AÄ´¦p¦P¼Ë¤@¥ó¶Â¦âªºÅ¨­m¦b¤£¦Pªº«G«×¡A¥¦ªºÃC¦â´N·|ÅܤơC©Ò¥H«Ü­«­nªº¬O«Ü®e©ö¤zÂZ§Ú­Ìªº¤èªk´N¬O¥¦ªº«G«×¡A©Ò¥H³o­Ócolor space¥¦ªº²Ä¤@­Ó¦n³B´N¬O¯à§â«G«×¨ú¥X¨Ó¡C²Ä¤G­Ó¦n³B¬O¥i¥H«D±`¦³®Ä²v¡C³q±`¤@­Ó¹Ï¤ù³£¬O500x500©Î¬O500x1000 ¡A©Ò¥H§Ú­Ì­n«D±`¦³®Ä²v¨Ó°µcolor spaceªºÂà´«¡C¦]¬°¦b³o­Óresearch·í¤¤¡A¦]¬°§Ú­Ì¥u¬O»Ý­nª¾¹D³o­ÓpixelªºÃC®Æ¡A©Ò¥H§Ú­Ì¥u­nª¾¹DE¡BS¡A¤£»Ý­nY¡C©Ò¥H¥i¥Hµo²{¡AEµ¥©ó¬OR-G¦A°£¥H¤G¡A©Ò¥H«D±`ªº§Ö³t¡F¦ÓS¬O§âR©MG¥[°_¨Ó¡A°£¥H¤G¡A´î±¼B¡A¦A°£¥H¤G¡C©Ò¥H°ò¥»¤W¥¦ªº¹Bºâ«D±`²³æ¡Bª½±µ¡C²Ä¤T­Ó¦n³B¬O¥¦¨S¦³signularities¡CSigularities¥D­n¬Oµo¥Í¦bÂà´«®É¬Ononlinearªºmatrix¡A´N¬O­è­è¤¤¶¡¨º­Ómatrix¬Ononlinear¡A´N·|¦³singularitiesªº°ÝÃD¥X²{¡F¦ý¬O¦]¬°³o¬O¤@­Ólinearªºspace¡A©Ò¥H¨S¦³³o­Ó°ÝÃD¡C³o´N¬O§Ú­Ì¬°¤°»ò¿ï¾ÜYESªº­ì¦]¡C¿ï¾Ü¤F³o­Ódata¤§«á¡A´N­n§ämean©Mcovarianceªº­È¡C¦]¬°¦b³o­Óresearch¬O­n°µ¤H¹³¸ò¤á¥~­·´º¡A©Ò¥H¿ï¾Ü¨âºØºØÃþ¡A¤@­Ó¬O¥Ö½§ªºÃC¦â¡A¥t¤@­Ó¬O¤ÑªÅªºÃC¦â¡C¦pªG¤@­Ó2-DªºGaussian function±q¥¿¤W¤è¬Ý¹L¥h¡A¥¦¬O¤@­Ó¾ò¶ê§Î¡C§Ú­Ì¥i¥H¦bºô¸ô¤W§ä¨ì¤@¨ÇÃÀ³Nªº¤H¹³¡A­º¥ý§Ú­Ì­n¨ú¤@¨Çtraining data¡A©Ò¥H§Ú­Ì­nºI¨ú©Ò¦³¦³Ãö©ó¥Ö½§ªºÃC¦âªº¤p¤pªº¼v¹³¡A§â¥¦Âà´«¦¨YESªº®y¼Ð¡C¤§«á¡A§Ú­Ì´N§âÂIµe¥X¨Ó¡A¦]¬°¨C¤@­Ópixel³£·|¦³E©MSªº­È¡A¨º§Ú­Ì´N§â¥¦¼Ð¥X­ÓÂI¨Ó¡C¨º»ò¤@¤p±iimage¸Ì­±´N·|§Î¦¨¤@­Ó±Ú¸s¡A³o¼Ë´N¥i¥Hºc¦¨¤@­Ó¹³¾ò¶êªº¹Ï§Î¡A¦]¬°§Ú­Ì¨úªº¼v¹³¤£°÷¦h¡A©Ò¥HÁÙ¤£¬O§¹¾ãªº¾ò¶ê¡A¸ò¾Ú²Î­p¾Çªº­ì²z¡A·í¨úªº·U¦h®É¡A²Î­pªº¤À¥¬·|ÁͦV¤@­Ó¦ÛµMªºfunction¡C¥t¥~¬O¤ÑªÅªºÃC¦â¡F¦bµe®aªº²´¸Ì¡A¤ÑªÅ¥D­n¬O¥Õ¦â¡BÂŦâ¡C¦ý¬O¤j®a¤£¯à¥u¬Ý³o¤G«×ªÅ¶¡¡A§Ú­ÌÁÙ­n¬Ý¥¦²Î­pªº¼Æ¥Ø¡A¦]¬°¦b¦P¤@­ÓÂI¤W¥i¯à¦³«Ü¦hÂI³£²Ö¿n¦b³o¸Ì¡C©Ò¥H­n±N¥¦¬Ý¦¨¤T«×ªÅ¶¡¡A¥¦¥i¯à¬O¤@®y¤s®ÇÃ䪺¤@®y¤p¤s¡A©Ò¥H§â¥¦«ö·Ó¼Æ¥Ø¨Ó¥­§¡¤§«á¡A´N·|²£¥Í³o¨â­Ó¾ò¶ê§Î¡C±µ¤U¨Ó´N­n¨Ó±Ð¹q¸£«ç»ò¼Ë¨Ó¤À¿ë¨v¹³ªº¤H¹³¡AÁÙ¬O¦ÛµMªº­·´ºµe¡C³o­Ó¤èªk¦³¤»­Ó¨BÆJ¡G

  1. §â¨C­ÓpixelªºRGBªº­È´«ºâ¦¨E©MSªº­È¡C
  2. Histogramªºanalysis¡A¥D­n´N¬Olambdaªº­È¡A¦]¬°¨C¤@­Ópixelªº­È³£¦³E©MSªºvalue ¡A©Ò¥H¥i¥H¥N¤J­è¤~¨º­Ó¾ò¶êªº¤èµ{¦¡¡CµM«á·|ºâ¥X¤@­Ólambdaªº­È¡C¦Ó¹ï©ó³o­Ólambdaªº­È¡A¥i¥H®³¨Ó°µ²Î­p¤Wªº¤ÀªR¡A¥i¥H§ä¨ì¥¦«D±`¶°¤¤ªº¦a¤è¡A©M¤£¶°¤¤ªº¦a¤è¡CÁöµM¨M©w¤F³Ìªìªºthreshold¡A¦ý¬O¨C±i¹Ïµeªº¯S©Ê¤£¦P¡A©Ò¥H³o­Óthreshold¬O¥i¥H¦Û¤v½Õªº¡C
  3. threshold¥¦·|¦Û¤v²¾°Ê¡A ¸ò¾Ú§Ú­Ìµ¹¥¦ªº¤@­Ó³W«h¡A¥¦·|¶}©l²¾°Ê¡CµM«á¨M©w¤@­Óthreshold value¥sta¡A¦Ó¤@¶}©lªº­È¬Otu¡C
  4. ¤@¦ý¨M©w¤F³o­Ó·sªºthresholdªº­È¤§«á¡A§âÄݩ󨺭ӱڸsªº¥þ³¡§ä¥X¨Ó¡C
  5. ¦b§ä¥X¤@­Ó±Ú¸s¤§«á¡A­Y¥¦¬OÄÝ©ó¦P¤@­Óregionªº¡A´N¥i¥H§ä¥X¥¦Ãä½tªºpixel¡C
  6. linking¡A´N¬O§â¨C¤@­ÓÃä½tªºpixel³s°_¨Ó¡A¨º§Ú­Ì´N§ä¨ì¤Fboundary¡C

¸ò¾Ú³o¨Çboundary¡A§Ú­Ì¥i¥H¿ë§O¬O¤Hª«ªºµe¹³ÁÙ¬O¦ÛµMªºµe¹³¡C­º¥ý¦³­Ódigitalªºcolor image¡A¿é¤J¹q¸£§âRGBÂà´«¦¨ES¡A¬Ý¦³´XºØºØÃþ¡A¦b³o¸Ì¡A§Ú­Ì©Ò¥Îªº¬O¥Ö½§ªºÃC¦â©M¤ÑªÅªºÃC¦â¡C¦pªG­n§ä¥Ö½§ÃC¦â¡A´N¥Î­è­è¥Ö½§ÃC¦âªºmean value©Mcovariance¨Óºâ¥Xlambda¡AµM«á§â³o¨Çlambdaµe¦¨histogramªº¹Ï¡AµM«á´N­n¶}©l¥h§ä·sªºthreshold value¡A¦]¬°°_ªì¨M©wªº¬O¦b2¡A¥i¬O¥¦¤£¤@©w­n¬O2¡A¥¦·|²¾°Ê¡C¸ò¾Ú¨C±iµeªº¯S©Ê¡A§Ú­Ì¥h§ä¤@¨Ç¥i¥H°Êªºvalue¡C§Ú­Ì¥ý§ä³Ì°ªÂI¬OpeaksÁÙ¦³¥¦ªº®p¨¦ÂIvalleys¡C§Ú­Ì½T©w¤F³oÂI¤§«á¡A§Ú­Ì´N¬Ý·íªì¤@¶}©l³]ªºinitialªºthreshold¬O¤£¬O¦b¾aªñpeaksªº¦a¤è¡A­pºâ³oÃ䪺­±¿n©M³o­ÓÂI¦b¤U¤@­Óvalleyªº­±¿n¡F¬Ý¨º¤@­Ó¤j¦Ó©¹¤jªº¤è¦V²¾°Ê¡C©Ò¥H³o­Ó¦a¤è¬O¤@­Óautomatic adaptionªºalgorithm¡C¦pªG³o­Ó¤£¬O¦b«Ü©úÅ㪺°Ï°ì¡A´N¿ï¾Ü­ì¨Óªº­È¡C³o­Ó¨BÆJ´N¨M©w¤F§Ú­Ì·sªºthresholdªº­È¡C¥Î³o­Óthresholdªº­È¡A¥i¥H¨Ó¨M©w¾ò¶ê§Îªº¤j¤p¡Athresholdªº­È¿ï¤j¤@ÂI¡A¾ò¶ê´N·|Åܤj¡A¤Ï¤§«hÅܤp¡C¦Ó«á¨Ó¤ÀÃþ¡A­Y¦³Ãø°Ï¤Àªº¦a¤è¡A´N¥Î²Î­p¾Ç¤WªºMAPªº¤èªk¨Ó¸Ñ¨Mambiguties¡C¤§«á¿ï¾Üedge pixel¡A§â¥¦link¡A¥i¥H§ä¨ìÄݩ󨺤@­Ó±Ú¸sªº°Ï°ì¡CÃö©ó¹êÅ窺µ²ªG¡A­º¥ý¦b¤H¹³³¡¤À¡A¿é¤J¹q¸£ªº¬O»X®RÄR¯½ªºµe¹³¡A³o±iµe¹³¦³¤@­Ó«Ü¯S§Oªº¦a¤è¡A´N¬O»X®RÄR¯½¬O¯¸¦b¤@´Tµe¹³ªº«e­±¡A¦Ó¨º´Tµe¤ÑªÅªºÃC¦â¡A«Ü¹³¥Ö½§ªºÃC¦â¡A©Ò¥H¥i¥H¬Ý¨ì§Ú­Ì¸ÕÅ窺µ²ªG¥i¥Hµo²{¦b¤H¹³«á­±¤]¬O«Ü¦h¥Ö½§ÃC¦âªº¡C¦b§Ú­Ì¦³¤@°Ï¤@°Ï¥Ö½§ªº°Ï°ì¤§«á¡A§Ú­Ì´N¨Ó§ä¥¦Ãä½tªºpixel¡A§ä¥X¨Ó¤§«á¡A°µ¤F¤@­Ó¤jÁxªº°²³]¡A©Ò¦³¤H¹³¹ÏµeªºÁy¡A¤@©w¦b¤W¥b³¡¡A¤£¥i¯àµe¦b¤U¥b³¡¡C¤HªºÁy¤@©w·|§Î¦¨¤@­Óclose loop¡A¦Ó¥B¥¦ªº­±¿n·|¤j©ó¾ã±i¹Ïµeªº4%¡C¦]¦¹§Ú­Ì¤jÁxªº°²³]¦pªG¦b¹Ïµeªº¤W¥b³¡§ä¨ì¤@­Ó­±¿n¶W¹L4%ªº«Ê³¬°Ï°ì¡A´N¥i¥H¿ë§O³o±iµe¬O¨v¹³µe¡C¦Ó¥t¥~¤@­Ó¨Ò¤l¬O²ö¤º¤á¥~ªºµe¡A³o±iµe¬Omarch¡A¦³¤@Ã䪺¤ÑªÅ¡A¤w³Qµe®aµeªº¤Ó¶Â¤F¡A¤£¦b§Ú­Ìtrainingªº½d³ò¡A©Ò¥H§Ú­Ì¤£¥h¨ú¥¦¡C¤U¤@¨B´N¬O§äÃä½tªºpixel¡A³o¨Çpixel¬OÄÝ©óÃäÃ䪺pixel¡C§Ú­Ì¤S°µ¤@­Ó°²³]¡A©Ò¦³ªº¤ÑªÅ·|¦û¤@­Ó«Ü¤jªº­±¿n¡A¬ù¦û20%¥H¤W¡A§Ú­Ì§ä¥X¨Ó¬O³Ìªøªº¡A¦Ó¥B±`±`¬O²×¤î¦bÃä½t¡A¤£¤@©w¬O¤@­Óopenªº¡C­Y§ä¨ì¤@­Ó³ÌªøªºÃä¬O¤ÑªÅªºÃC¦â¡A´N¨M©w´N±iµe¬O­·´ºµe¡C³o¨â­Ó°²³]³£¦³¯ÊÂI¡A¥u¯à°µ²Å¦Xªº¹Ï¡C

µ²½×¡G

³o­Óproject¦³¨â­Ó¦n³B¡A²Ä¤@­Ó¬O¥¦¥þ¬O¥Î¦Û°Ê¨Ó¿ë§Oªº¡A¤£¥Î¤H¤u¥h¿ï¾Üthreshold¡A¦Ó¦Û°Êªº¥h¿ë§O©Ò­nªº¹Ï¡C³o­Óalgorithm ¥i¥H°µ±m¦âªº¤ÀÃþ¡C¤§«á¤£¦Pªºpixel¥i¥H§â¤£¦Pªº°Ï¶¡¤À¥X¨Ó¡A©Ò¥Îªº¤èªk¬O¥u­n¬OÄÝ©ó¦P¤@­Ó°Ï°ìªºpixel¡A¥¦°ò¥»¤W¬O³sÄò©Êªº¡A¦Ó¤£·|³æ¿W¦s¦b¡A©Ò¥Hpixel·|¶°¤¤¦b¤@°Ï¡C²Ä¤G­Ó¦n³B¬O¥u¥Î¨ì¹Ïµeªº±m¦â¡A¥u¦³¦b«á­±¨BÆJ¤~¦³¥Î¨ìshape¡A©Ò¥H¥Î¥¦±m¦âªº³¡¤À¨Ó¤ÀÃþ¬O«D±`ªºÂ²³æ¡C¦Ó¥B­pºâªº®É¶¡«Üµu¡C³Ì«á¦b¿ëÃÑ¥¦¬O¨v¹³ÁÙ¬O¦ÛµM­·´º®É¡A¦³¥Î¨ì­±¿n¡A©Ò¥H¥u¥Î¨ì¥¦ªº±m¦â³¡¤À©M­±¿n³¡¤À¡C

FUTURE WORK¡G

²Ä¤@­Ó¬O¥i¥H¼W¥[§ó¦hªººØÃþ¡A©Î¬Okeyword¡C²Ä¤G­Ó¥i¥H¥[¤J§ó¦hªº¤§«e©Î¤§«áªº¼v¹³³B²z¡A¨Ó¨Ï±oµ²ªG§ó¦n¡C¦p¥i¥H§âdataÀ£ÁY¡A¦b­pºâªº®É¶¡¥i¥HÁY±o«Üµu¡C²Ä¤G­Óregion segmentation­n°µªº«Ü·Ç½T¡A¥Ø«e¨S¦³°µsimulationªº³¡¤À¡A­Y°µsimulationªº³¡¤À¡A´N­n­«·s¥Î­Ó§ó·Ç½Tªºregion segmentationªº¤èªk¡C


ºtÁ¿¸Ô²Ó¤º®e¡G

¤µ¤Ñ¨ÓÁ¿ªº¬O¦³Ãö©ó±m¦â¼v¹³°Ï°ì¤À¹jªº¤èªk¡A§Ú­Ó¤H¨«ªº¤ñ¸û¾aªñ¼v¹³³B²z(image processing)ÁÙ¦³¹q¸£µøÄ±(computer vision)³o¤è­±¡A³o¤è­±ÁÙ¬O¤@­Ó¾Ô°ê®É¥N¡F¤j®a³£¨S¦³¤@­Ó³Ì¦nªº¤èªk¨Ó¸Ñ¨M¤@­Ó°ÝÃD¡A©Ò¥H´N¬O¦Ê®aª§»ï¡A¦Ó¥B¨S¦³¤H¯à´£¥X¤@­ÓÅý¤j®a³£«ÜªA®ðªº¤èªk¡C©Ò¥H¤µ¤Ñ³o­ÓÃD¥Ø¤]¬O¤@¼Ë¡A³o­Ótopic«Ü¦h¤H¦b°µ¡A¦ý¬O¨S¦³¤H¯à°÷´£¥X³Ì¦nªº¤èªk¡A¥u¯à°÷°w¹ï¬Y¤@ºØ¯S®í¥Î³~¡A¨Ó´£¥X¦o­Ì¤@­Ó¤ñ¸û¦nªº¤èªk¡C¤µ¤Ñ§Ú³o­Óproject¥D­n¬O§Ú¦b¬ü°ê­×ºÓ¤h¯Z½Òªº¤@­Óproject¡C¦bproject¤§«á¡A§Úªº±Ð±Â»{¬°§Ú³o­Óprojectªºidea«D±`¦n¡A©Ò¥H§Ú§â¥¦implement¦¨¤@­Ópaper¡AµM«á¦b1997¥xÆW¶}ªº¤@­Ómultimediaªºconference¦³µoªí¹L¡C¥¦ªºtopic¥D­n¬O§Æ±æ¥Î±m¦â°Ï°ì¤À¹jªº¤èªk¨Ó¿ë»{ÃÀ³Nªºµe¹³¡A¥¦¬O¤@­Ó¤Hª«ªº¨v¹³©O¡HÁÙ¬O¤@­Ó¦ÛµM­·´ºªºµe¡C¦U¦ì¦P¾Ç¦pªG¦³¥ô¦ó°ÝÃD¥i¥HÀH®É¥´Â_§Ú¡A¦]¬°¤µ¤Ñ³o­ÓºtÁ¿¤ñ¸û­«­nªº¬OÅý¤j®aª¾¹D¤j·§¤@­Ó¤ëªºproject¥i¥H«ç»ò¨Ó°µ¡F¦]¬°³o­ÓªF¦è¥uªá¤F¤@­Ó¤ë¡A©Ò¥H¥¦¦³«Ü¦h»Ý­n¸É±jªº¦a¤è¡Aidea°µ¬ã¨sªº¤è¦¡¹ê»Ú¤Wxxxx(Å¥¤£²M·¡)¡C­º¥ý²Ä¤@­Ó¬O¬°¤°»ò­nµo®i³o­ÓªF¦è¡A§Úªº°Ê¾÷¬O¦pªG©p¦bºô­¶¤W¥i¥H¬Ý¨ì¤@¨Ç§O¤H¤w¸g¤Wºô¸ôªº¥sweb museum´N¬O¦bºô¸ô¤Wªº³Õª«À]¡A¨º¬O«D±`¦h«D±`¦hªºµe¹³¡BÃÀ³N«~¡A©p«ç»ò¼Ë¨Óhandle¥¦¡AµM«áÀx¦s¥¦¡F¬Æ¦Ü©p­n§ä¤@­Óµe¹³¡A¦ý©p¤£¾å±o³o´Tµe¬O½Ö§@ªº¡A©p¥uª¾¹D³o´Tµe¸Ì­±¥i¥H¦³¤@­Ó¬ü¤k¦b·L¯º¡AµM«á©p´N¥´¶i³o¨Ç·Qªk¡AµM«á¥¦´N¥i¥HÀ°©p§ä¨ì¡C¥t¥~¤@­Ó°Ê¾÷¬O±N¨Ó¤£¤@©w¬OºÞ²z¤@­Ó¬ü³NÀ]¡A¦Ó¬OºÞ²z¨ä¥¦ªF¦èªº®É­Ô¡A©Ò¥H§Ú­Ì»Ý­n¤@­Óobject-oriented environment¡C³o­ÓÀô¹Òªº¯SÂI¬O¦³«Ü¦h¸ê®Æ­n³B²z¡A²Ä¤G­Ó²{¦bªº³B²z¤è¦¡¹³¹Ï®ÑÀ]ªº¤è¦¡¬O¥Î¤å¦r¡C¨º»ò¡A§Ú­Ì·Q­n«Ø¥ß¤@­Ó¬O¥Î¥ô¦ó¦³Ãö¼v¹³ªºªF¦è¨Ó³B²z³o­Óenvironment¡F©Ò¥H§Ú­Ì¥i¥H¥Î¥¦ªºcolor¡Btexture¡Bshape©Mlayout¨Ó¤ÀªR¡C³Ì«á§Ú­Ì­n¹F¨ìªº¤@­Ó¥Î³~¬O¨Ï¥ÎªÌ¤£»Ý­n¤¤¶¡¥h¿ï¾Ü¡A¤ñ¦p«ö¶i¨º¨Ç¦r¡A¥¦´N·|¥X²{¨º¥»®Ñ¡F¦b¼v¹³¤è­±¡A§Ú­Ì§Æ±æ©p¯à°÷µ¹¥¦¤@­Óexample¡A¥¦´N¥i¥Hª½±µÀ°©p§ä¨ì¡A©p¤£»Ý­n¥h«ü¾É¹q¸£¥h§ä¨ì¥¦¡C

°ò©ó³o¨â­Ó°Ê¾÷¡A§Ú­Ì¨Ó¬Ý¦³¨º¨Ç¤H¤w¸g°µ¹L³o¤è­±ªº¬ã¨s¡C§Ú­Ì¤À¬°¨â¤è­±¡A­º¥ý¦b¸ê®Æ³B²z´£¥X¤@­Ómodelªº¤H¡A¥D­n¦o­Ì¬O·Q­n¿ë»{¦³¨v¹³¦bªºimage¡F©Ò¿×¨v¹³ªºimage´N¬O¦³­Ó¤H¡A¹³¦U¦ìªº¤jÀY·Ó¤@¼Ë¡C¨º¥¦¬O¤@­Óportrait imageªº¸Ü¡A¦³¨â­Ó¤ñ¸û¦¨¥\ªºwork¤w¸g¦³¤F¡C²Ä¤@­Ó¦o­Ì¬O¥ÎRGB¡A§Ú­Ìª¾¹D±m¦âspace¬ORGB¡A¦ý¬O§Ú­Ì¥i¥H¥ÎRGB¤T­Óspace´«¦¨¥ô¦ó¤@­Ócolor space¡C¨º»ò³o­Ó¤H¦o¬O§â¤jªºRGB space´«¦¨¤pªºrgbªºcolor space¡C³o­Ó¤H¬OMiyake¦b1990¦~¡A¦o¬O¯à°÷§â§Ú­Ì¥Ö½§ªºÃC¦â¡A¦]¬°portrait³Ì¦hªº´N¤Hªº¥Ö½§ªºÃC¦â¡C¦oµo²{¥Ö½§ªºÃC¦â¦brg³o¨â«×ªÅ¶¡ªºspace¤W­±¥i¥H§Î¦¨¤@­Ó¾ò¶ê§Îªºgroup¸Ì­±¡C

²Ä¤G­Óµo²{¬O¦³¤H§âRGB´«¦¨YES¡AYES¬O¥t¥~¤@­Ócolor space¡C³o­Ó¤H¬OSaber¦b1996¦~´£¥Xªº¡A¦oµo²{¦³¤T­Ó±Ú¸sªºÃC¦â¡A´N¬O¥Ö½§¡B¤ÑªÅ¡B©M¯ó¦aªºÃC¦â¦b"ES"³o­Ó¥­­±¤W·|§Î¦¨2-DªºGaussian functionªº¤À¥¬¡C³o½gpaper¨ä¹ê´N¬O³o­Ó¤Hpaperªº©µ¦ù¡C§Ú­Ì§Æ±æ¥Î¦oªºmodel¨Ó°µ§ó¦hªº¨Æ±¡¡A³Ì«á¹F¨ì°Ï°ì¤À¹jªº®ÄªG¡C¦bobject-oriented environment¤è­±¡A¤w¸g°µ¹Lªº¦³«Ü¦h¡A³o­Ó¤j®a³£¦³Å¥¹L¡C¹³IBM¥¦¦³µo®i¤@­Ó³nÅé¥s°µQuery-by-Image-Content¡AÄ´¦p»¡¡A©p­n§ä¤@­Ó¬õ¥Õ¬Û¶¡ªºT-shirtªº¹Ï®×¡A©p¥u­n§â³o­Ó¹Ï®×©ñ¦buser choice³oÃä¡A¥¦´N·|§ä¥X¦b¥¦database¤¤©Ò¦³¦³¬õ¥Õ¬Û¶¡ªº¹Ï®×³£µ¹©p¡F¦b1993¦~¤w¸gcommercial¡CMIT¦³¤@­Óphotobook¦b1993¦~¡CETL¦b1991¦~¦³­Ó³B²zart museumªº³nÊ^¡CUniversity of Miohigan ¦b1993¦~¤]°µ¥X¤@­Ó³nÊ^¥s°µXenemania¦b1993¦~¡A³Ìªñªº¬O1995¦~¥ÑColumbla Univ.¥X¤F¤@­ÓMultimedia testbed¡C

¶i¤J¤µ¤Ñtalkªºoutline¡A­º¥ý§Ú­Ì¤µ¤Ñ­n¤¶²Ðªº¬O²Ä¤@­Ó§Ú­Ì«ç¼Ë¥h¿ï¾Ü¤@­Ócolor space¡AµM«á¿ï¾Ü¤@­Ódata modeling¡C²Ä¤G­Ó­n¤¶²Ðªº¬O§Ú¥Îªº¤èªk³Ì«á¨Ó¸Ñ¨M§Q¥Î±m¦â°Ï°ì¤À¹jªº¤è¦¡¡C³o­Ó¤èªk¦³¤T­Ó¨BÆJ¡F²Ä¤@­Ó¨BÆJ¬O¦]¬°¦b¤@­Óiamge¸Ì­±¡A¨C­Ó³æ¦ì´N¬Opixel¡A²Ä¤@­Ó¬O¥Îpixel¨Ó¤ÀÃþ¡F´N¬O¨º­Ópixel ¥¦¬OÄݩ󤰻òÃC¦â¡A§Ú­Ì´N§â¨º­Ópixel ¤À¬°¨º¤@Ãþ¡C¦b©p§â©Ò¦³¦P­ÓÃC¦â¡AÄ´¦p»¡¤Hªº¨v¹³¨ÓÁ¿¡AÁyªº¤è­±¦pªG¯à°÷¥i¥H§â©Ò¦³¥Ö½§ªºÃC¦âpixel³£§ä¥X¨Ó¡A¤§«á¡A§Ú­Ì´N§âpixel¾ã¦X°_¨Ó¡AÅܦ¨¤@­Ó°Ï°ì¡AµM«á§â¥¦¤À¹j¥X¨Ó¤§«á¡A³Ì«á§Ú­Ì§Æ±æ§â¥¦ªºboundary¯à°÷¨M©w¡C§Ú³o½gµ¹¤j®a¬Ýªº¬O¤@­ÓÀ³¥Î¡A«ç¼ËÅý¹q¸£¦Û°Ê¦a¤À¿ë¤@­Ó¬ü³Nªº¹Ïµe¬O¤Hªº¨v¹³ÁÙ¬O¦ÛµM´ºÆ[ªº¹Ïµe¡C³Ì«á·|µ¹¤j®a¬Ý¤@¨Ç¹êÅ窺µ²ªG¡C

­º¥ý¡A§Ú­Ì´N¨Ó¿ï¾Ü¤@­Ócolor space¡F³o­Ócolor space¥sYES¡A³o­ÓYES¬O¥Ñ¤@­Ó¤é¥»ªº¾Ç·|¡A¥s°µMotion Picture and Television Engineersªº¾Ç·|©Ò´£¥X¨Óªº¡A¤]´N¬ORGB¦A¸g¹L¤@­ÓmatrixªºÂà´«¤§«á¡A¥¦·|Âà´«¦¨YES¡C³o¨Ç«Y¼Æ«ç»ò¨Óªº©O¡H°ò¥»¤W¬O¸gÅ窺¡A³o¨ÇEngineers test«Ü¦h¤£¦P·Ó©úµ{«×ªº¹Ï¤ù¡AµM«á¨Ó¿ï¥X³o­Ómatrixªº«Y¼Æ¡A¦o­Ì¨M©w³o­Ó¥sYES¡CYES³£¦³¥¦ªº·N¸q¦b¡AY:¨Æ¹ê¤W´N¬O§â¤@­Ópixel¤¤RGBªº¦¨¤À©Ò¦³¦³Ãö«G«×ªº¦¨¤À§â¥¦®³¥X¨Ó¡A©Ò¥H¥sliminance¡CE©MS°ò¥»¤W´N§â¥¦¦â±mªº¦¨¤À¡A¦³ÃC®Æªº³¡¤À¨ú¥X¨Ó¡C

§Ú¬°¤°»ò­n¿ï¾Ü³o­Ócolor space¡A¦pªG¤j®a¦³¿³½ì¥i¥H¬Ý¬Ýimage processingªºbook¡A¥¦·|¼g»¡color space´N¤@­Ó¤jtable¦³«Ü¦h¤H´£¥X¤£¦Pªº¥Î³~¡C³o­Ó¬O¤ñ¸û·sªºÃD§÷¡A©Ò¥H©pªº±Ð¬ì®Ñ§ä¤£¨ì¡C¥¦ªº¦n³B¬O¡G¤@¡B¯à°÷§âluminance ª½±µ®³¥X¨Ó¡AÄ´¦p¦P¼Ë¤@¥ó¶Â¦âªºÅ¨­m¦b¤£¦Pªº«G«×¡A¥¦ªºÃC¦â´N·|ÅܤơC©Ò¥H«Ü­«­nªº¬O«Ü®e©ö¤zÂZ§Ú­Ìªº¤èªk´N¬O¥¦ªº«G«×¡A©Ò¥H³o­Ócolor space¥¦ªº²Ä¤@­Ó¦n³B´N¬O¯à§â«G«×¨ú¥X¨Ó¡C

²Ä¤G­Ó¦n³B¬O¥i¥H«D±`¦³®Ä²v¡C¦]¬°³q±`¤@­Ó¹Ï¤ù³£¬O500x500©Î¬O500x1000 ¡A©Ò¥H§Ú­Ì­n«D±`¦³®Ä²v¨Ó°µcolor spaceªºÂà´«¡C¦]¬°¦b³o­Óresearch·í¤¤¡A§Ú­Ì¥u»Ý­nE©MS¡A§Ú­Ì¤£»Ý­nY¡A¦]¬°§Ú­Ì¥u¬O»Ý­nª¾¹D³o­ÓpixelªºÃC®Æ¡A©Ò¥H§Ú­Ì¥u­nª¾¹DE¡BS¡C©Ò¥H¥i¥Hµo²{¡AEµ¥©ó¬OR-G¦A°£¥H¤G¡A©Ò¥H«D±`ªº§Ö³t¡F¦ÓS¬O§âR©MG¥[°_¨Ó¡A°£¥H¤G¡A´î±¼B¡A¦A°£¥H¤G¡C©Ò¥H°ò¥»¤W¥¦ªº¹Bºâ«D±`²³æ¡Bª½±µ¡C

²Ä¤T­Ó¦n³B¬O¥¦¨S¦³signularities¡CSigularities¥D­n¬Oµo¥Í¦bÂà´«®É¬Ononlinearªºnatrix¡A´N¬O­è­è¤¤¶¡¨º­Ómatrix¬Ononlinear¡A´N·|¦³singularitiesªº°ÝÃD¥X²{¡F¦ý¬O¦]¬°³o¬O¤@­Ólinearªºspace¡A©Ò¥H¨S¦³³o­Ó°ÝÃD¡C©Ò¥H³o´N¬O§Ú­Ì¬°¤°»ò¿ï¾ÜYESªº­ì¦]¡C

²Ä¤G­Ó§Ú­Ì­n¿ï¾Ü«ç¼Ë¨Ómodel³o­Ódata¡F§Ú¿ï¾Ü¤FSaberªº¤èªk¡A¦]¬°¦oªºmodelingªº¥»½è¡Bfeture¬O«Ü¦X¾Aªº¡A§Ú¦Û¤v¦³check¹L¡A©Ò¥H§Ú¿ï¾Ü¦oªº¤èªk¡C©Ò¿×ªº2-D Gaussian probability density function³o´N¬O¤@­Ó¥¦¼Ð·Çªº¤@­Óprobability function¡A¤j®a¥i¥H¬Ý¨ì¡A¬O³o¼Ë¤l¡CµM«á³oÃ䦳­Óexponential¡AµM«á§Ú­Ìµo²{³oÃä¡A³o­Ó¬Omean vectorµM«ák¬Ocovariance matrix¡AµM«á³oÃä­º¥ýx¬O¥Nªí¨C¤@­Ópixelªºe­È©Ms­È¡Cmean value¬Oe¨º­Ó¶bªºmean value©Ms¶bªºmean value¡Fcovariance´N¬Oe©Msªº¬ÛÃö«Y¼Æ¡C§Ú­Ì¥i¥Hµo²{¡A¦pªG§Ú­Ì§âexponential¤W­±ªº«Y¼Æ¬O¨M©w³o­ÓGaussian function¬O¤ñ¸û­DÁÙ¬O¤ñ¸û½G¡C¨º³o­Ócoefficient¬O¨M©w³o­ÓGaussian functionªº°ª«×¡A³Ì°ªªº¨º­ÓÂI¦b¨º¸Ì¡C¦]¬°Gaussian function ¦b¤T«×ªÅ¶¡±q¥¿¤W¤è¬Ý¹L¥h¡A¥¦¬O¤@­Ó¾ò¶ê§Î¡C¨º¤W­±³o­Ócoefficient¬O¨M©w³o­Ó¾ò¶êªº¤½¦¡¡C¦pªG§Ú­Ì§â³o­Óequaltionºâ¥X¤@­Ó­È¡A§Ú­Ì­q¥¦¥slambda¡A±N¨Ó¥Î°µ¦Û°Êªºalgorithmªºprimate(¤£²M·¡)¡C

§Ú­Ì¿ï¾Ü¤F³o­Ódata¤§«á¡A³o­Ómean©Mcovarianceªº­È«ç»ò§ä¡A±µ¤U¨Ó¡A´N­n¸ò¤j®aÁ¿¡A¦]¬°³o­ÓªF¦è¥H«á¦pªG©p­Ì¦³¿³½ì°µdata¤è­±ªº³£·|¥Î¨ì¡C¦]¬°¦b³o­Óresearch§Ú¬O­n°µ¤H¹³¸ò¤á¥~­·´º¡A©Ò¥H§Ú¿ï¾Ü¨âºØºØÃþ¡A¤@­Ó¬O¥Ö½§ªºÃC¦â¡A¥t¤@­Ó¬O¤ÑªÅªºÃC¦â¡C¦pªG¤@­Ó2-DªºGaussian function±q¥¿¤W¤è¬Ý¹L¥h¡A¥¦¬O¤@­Ó¾ò¶ê§Î¡C³o´N¬O§Útraining data©Ò²£¥Íªº¸ê®Æ¡Cstar©Ò§Î¦¨ªº¬O¥Ö½§©Ò¦¨ªº±Ú¸s¡A¦Ó°é°é©Ò§Î¦¨ªº¬O¤ÑªÅ§Î¦¨ªº±Ú¸s¡C§Ú­Ì­n«ç»ò°µ©O¡H­º¥ý¡A²{¦bºô¸ô«D±`ªºµo¹F¡A¥i¥H¦bºô¸ô¤W§ä¨ì¤@¨ÇÃÀ³Nªº¤H¹³¡C¥H¤U¬O¤j®a¬Ý¨ì«Ü¼ô±xªº­±¤Õ--±ë¨¦¡C­º¥ý§Ú­Ì­n¨ú¤@¨Çtraining data ¡A©Ò¥H§Ú­Ì­nºI¨ú©Ò¦³¦³Ãö©ó¥Ö½§ªºÃC¦âªº¤p¤pªº¼v¹³¡A§â¥¦Âà´«¦¨YESªº®y¼Ð¡C¤§«á¡A§Ú­Ì´N§âÂIµe¥X¨Ó¡A¦]¬°¨C¤@­Ópixel³£·|¦³E©MSªº­È¡A¨º§Ú­Ì´N§â¥¦¼Ð¥X­ÓÂI¨Ó¡C¨º»ò¤@¤p±iimage¸Ì­±´N·|§Î¦¨¤@­Ó±Ú¸s¡C§Ú¿ï¤F¤Q­Ó¤H¹³ªºµe¡A´N¥i¥Hºc¦¨³o­Ó¡C©p¬Ý¥¦¬Ý°_¤£¹³¾ò¶ê¡A¦]¬°§Ú­Ì¨úªº¼v¹³¤£°÷¦h¡A¸ò¾Ú²Î­p¾Çªº­ì²z¡A·í¨úªº·U¦h®É¡A²Î­pªº¤À¥¬·|ÁͦV¤@­Ó¦ÛµMªºfunction¡A©Ò¥H³o­Ó­ì¦]¬O§Ú­Ì¨úªº¤£°÷¦h¡C

¥t¥~¡A³o­Ó¹Ï¬O¤ÑªÅªºÃC¦â¡F¦bµe®aªº²´¸Ì¡A¤ÑªÅ«Ü¦h³£¬O¦³¶³ªºÃC¦â¡A©Ò¥H¥D­n¬O¥Õ¦â¡BÂŦâ¡F³o­Ó©_©Çªº±Ú¸s¬O¦³¤@±iµe¡A¥¦¤ÑªÅªºÃC¦â«D±`ªºÂÅ¡C©Ò¥H·|³y¦¨³o¨â­Ó±Ú¸s¤À¶}¡C¦ý¬O¤j®a¤£¯à¥u¬Ý³o¤G«×ªÅ¶¡¡A§Ú­ÌÁÙ­n¬Ý¥¦²Î­pªº¼Æ¥Ø¡A¦]¬°¦b¦P¤@­ÓÂI¤W¥i¯à¦³«Ü¦hÂI³£²Ö¿n¦b³o¸Ì¡C©Ò¥H­n±N¥¦¬Ý¦¨¤T«×ªÅ¶¡¡A¥¦¥i¯à¬O¤@®y¤s®ÇÃ䪺¤@®y¤p¤s¡A©Ò¥H§â¥¦«ö·Ó¼Æ¥Ø¨Ó¥­§¡¤§«á¡A´N·|²£¥Í³o¨â­Ó¾ò¶ê§Î¡C

§ä¨ì³o­Ó¤§«á¡A´N¥i¥H¨M©w¡A³o¨â­ÓÂI¡A´N¬O±Ú¸sªº¤¤¤ß¡C¦Ó¾ò¶ê§Îªº¤j¤p¡A¬O­n¥Ñ¦Û¤v¥h¨M©w¡F¦pªG³o­Ó¾ò¶ê§Î¦b¤T«×ªÅ¶¡¬Ý°_¨Ó«Ü­D¡A¨º»ò³o­Ó¾ò¶ê´N­n¨ú¤j¤@ÂI¡F¦pªG¬O¤@­Ó«ÜÅÖ²Óªº¾ò¶ê¡A´N­n§ä½G¤@ÂI¡C©Ò¥H³o­n¥h¬Ý¤@¤U¤T«×ªÅ¶¡ªº¤À¥¬¡C³Ì«á·|±o¨ì¥¦ªº¥­§¡­È¡A´N¬O­è¤~¨º¨Ç¾ò¶êªº¤¤¤ß¦ì¸m¡AµM«á¸ò¥¦ªº¤j¤p¡AÁÙ¦³locationªº¤è¦V¥i¥H¨M©wcovariance¡C¾ò¶êªº¤j¤p¡A­n¨M©w¦h¤j¬O«D±`conferdant(¤£²M·¡)ªºinteval¡C¤]´N¬O»¡¡A§Ú½T©w©Ò¦³ªºÂI¦b³o­Ó¾ò¶ê¤§¤º³£¬O³o­ÓÃC¦â¡C©Ò¥H¡A¸ò¾Ú§Ú­Ó¤H©Ò¥Îªºtraining data¡A§Ú¨M©w¨â­Ó³£¬O2¡A¨º¥u¬O´ê¥©¡A¤£¤@©w­n¤@¼Ë¡C

¨M©w¤Fcolor space©Mdata modelingªº¤è¦¡¤§«á¡A³o¸Ìshow¤@¤U§Ú¥Îªº¨ä¥¦ªºtraining data¡A¦b¤H¹³¤è­±¡A§Ú¥Îªº¦³¹F¤å¦è¡B¹p¿Õº¸¡B±ë¨¦ªº¡F©Ò¥H¥¦­ÌªºÃC¦â¬O®t§O«Ü¦hªº¡C¦Ó¦b«Ç¥~ªº¦ÛµM´ºÆ[¡A§Ú©Ò¥Îªº¬O³o¨Ç¡A¹³³o­Ó¤ÑªÅ´N«D±`ªºÂÅ¡A©M®ü¤ô¤@¼Ë¡A³o´N¬O­è¤~²æÂ÷¥D­n±Ú¸sªº³¡¤À¡C¦³ªº¤ÑªÅ¥Rº¡¤F¶³¡AÁÙ¦³ªº¤ÑªÅ¬Oªx¬õ¡C

¦b¦ÛµM´ºÆ[¤è­±¡A§Ú©ÒµÛ­«ªºµe®a¬O²ö¤º¡C°ò¥»¤W¤j®a¥i¥H¦³­Óidea´N¬O±N¨Ó¥i¥H¥Î¨Ó³B²z·Ó¤ù¡A¬°¤°»ò§Ú³o­Óproject ¬O°µ¬ü³Nªºµe¡A¬O¦]¬°¦Ñ®v³W©w©Ò¦³ªºproject³£­n©M¬ü³N¦³Ãö¡A©Ò¥H¤j®a¤£¥i¥H¥Î·Ó¤ù¡C

±µ¤U¨Ó¡A´N­n¨Ó±Ð¹q¸£«ç»ò¼Ë¨Ó¤À¿ë¨v¹³ªº¤H¹³¡AÁÙ¬O¦ÛµMªº­·´ºµe¡C³o­Ó¤èªk¦³¤»­Ó¨BÆJ¡A²Ä¤@­Ó¨BÆJ´N¬O§â¨C­ÓpixelªºRGBªº­È´«ºâ¦¨E©MSªº­È¡A§Ú­Ì¤£»Ý­n¥¦ªº«G«×¡C²Ä¤G­Ó¨BÆJ´N¬OHistogramªºanalysis¡A¥D­n´N¬Olambdaªº­È¡A¦]¬°¨C¤@­Ópixelªº­È³£¦³E©MSªºvalue ¡A©Ò¥H¥i¥H¥N¤J­è¤~¨º­Ó¾ò¶êªº¤èµ{¦¡¡CµM«á·|ºâ¥X¤@­Ólambdaªº­È¡C¦Ó¹ï©ó³o­Ólambdaªº­È¡A¥i¥H®³¨Ó°µ²Î­p¤Wªº¤ÀªR¡A¥i¥H§ä¨ì¥¦«D±`¶°¤¤ªº¦a¤è¡A©M¤£¶°¤¤ªº¦a¤è¡CÁöµM¨M©w¤F³Ìªìªºthreshold¡A¦ý¬O¨C±i¹Ïµeªº¯S©Ê¤£¦P¡A©Ò¥H³o­Óthreshold¬O¥i¥H¦Û¤v½Õªº¡C©Ò¥H¤U¤@­Ó¨BÆJ´N¬Othreshold ¥¦·|¦Û¤v²¾°Ê¡A¸ò¾Ú§Ú­Ìµ¹¥¦ªº¤@­Ó³W«h¡A¥¦·|¶}©l²¾°Ê¡CµM«á¨M©w¤@­Óthreshold value¥sta¡A¦Ó¤@¶}©lªº­È¬Otu¡C¤U¤@­Ó¨BÆJ¬O¤@¦ý¨M©w¤F³o­Ó·sªºthresholdªº­È¤§«á¡A§âÄݩ󨺭ӱڸsªº¥þ³¡¦¬¶°¡B§ä¥X¨Ó¡C¤U¤@­Ó¨BÆJ¬O¦b§ä¥X¤@­Ó±Ú¸s¤§«á¡A­Y¥¦¬OÄÝ©ó¦P¤@­Óregionªº¡A´N¥i¥H§ä¥X¥¦Ãä½tªºpixel¡A¤U¤@­Ó¨BÆJ¬Olinking¡A´N¬O§â¨C¤@­ÓÃä½tªºpixel³s°_¨Ó¡A¨º§Ú­Ì´N§ä¨ì¤Fboundary¡C¸ò¾Ú³o¨Çboundary¡A§Ú­Ì¥i¥H¿ë§O¬O¤Hª«ªºµe¹³ÁÙ¬O¦ÛµMªºµe¹³¡C

­º¥ý¦³­Ódigitalªºcolor image¡A¿é¤J¹q¸£¡A¹q¸£´N§âRGBÂà´«¦¨ES¡A¬Ý¦³´XºØºØÃþ¡A¦b³o½gpaper ¤¤¡A§Ú­Ì©Ò¥Îªº¬O¥Ö½§ªºÃC¦â©M¤ÑªÅªºÃC¦â¡C¦pªG­n§ä¥Ö½§ÃC¦â¡A´N¥Î­è­è¥Ö½§ÃC¦âªºmean value©Mcovariance¨Óºâ¥Xlambda¡Aºâ¥Xlambda¤§«á¡A´N§â³o¨Çlambdaµe¦¨histogramªº¹Ï¡A³o¹Ï¬Ý°_¨Ó¡A´N¬O¹³³o¼Ë¤l¡Cºâ¥X¤F©Ò¦³ªº­È¡AµM«áµe¥X³o­Óhistogram¡A´N­n¶}©l¥h§ä·sªºthreshold value¡A¦]¬°°_ªì¨M©wªº¬O¦b2¡A¥i¬O¥¦¤£¤@©w­n¬O2 ¡A©Ò¥H¥¦·|²¾°Ê¡C¸ò¾Ú¨C±iµeªº¯S©Ê¡A§Ú­Ì¥h§ä¤@¨Ç¥i¥H°Êªºvalue¡C§Ú­Ì¥ý§ä³Ì°ªÂI¬OpeaksÁÙ¦³¥¦ªº®p¨¦ÂIvalleys¡C§Ú­Ì½T©w¤F³oÂI¤§«á¡A§Ú­Ì´N¬Ý·íªì¤@¶}©l³]ªºinitialªºthreshold¬O¤£¬O¦b¾aªñpeaksªº¦a¤è¡A§Ú­Ì·|­pºâ³oÃ䪺­±¿n©M³o­ÓÂI¦b¤U¤@­Óvalleyªº­±¿n¡F¬Ý¨º¤@­Ó¤j¦Ó©¹¤jªº¤è¦V²¾°Ê¡C©Ò¥H³o­Ó¦a¤è¬O¤@­Óautomatic adaptionªºalgorithm¡C¦pªG³o­Ó¤£¬O¦b«Ü©úÅ㪺°Ï°ì¡A´N¿ï¾Ü­ì¨Óªº­È¡C³o­Ó¨BÆJ´N¨M©w¤F§Ú­Ì·sªºthresholdªº­È¡C

¥Î³o­Óthresholdªº­È¡A¥i¥H¨Ó¨M©w¾ò¶ê§Îªº¤j¤p¡A¦]¬°§ïÅܤF³o­Óthresholdªº­È¡A¾ò¶ê´N·|ÅܤơCthresholdªº­È¿ï¤j¤@ÂI¡A¾ò¶ê´N·|Åܤj¡A¤Ï¤§«hÅܤp¡C¦Ó«á¨Ó¤ÀÃþ¡A­Y¦³Ãø°Ï¤Àªº¦a¤è¡A´N¥Î²Î­p¾Ç¤WªºMAPªº¤èªk¨Ó¸Ñ¨Mambiguties¡C¤§«á¿ï¾Üedge pixel¡A§â¥¦link¡A¥i¥H§ä¨ìÄݩ󨺤@­Ó±Ú¸sªº°Ï°ì¡C

²{¦b§Ú­Ì¨Ó¬Ý¤@¤U¹êÅ窺µ²ªG¡A­º¥ý¦b¤H¹³³¡¤À¡A¿é¤J¹q¸£ªº¬O»X®RÄR¯½ªºµe¹³¡A³o±iµe¹³¦³¤@­Ó«Ü¯S§Oªº¦a¤è¡A´N¬O»X®RÄR¯½¬O¯¸¦b¤@´Tµe¹³ªº«e­±¡A¦Ó¨º´Tµe¤ÑªÅªºÃC¦â¡A«Ü¹³¥Ö½§ªºÃC¦â¡A©Ò¥H¥i¥H¬Ý¨ì§Ú­Ì¸ÕÅ窺µ²ªG¥i¥Hµo²{¦b¤H¹³«á­±¤]¬O«Ü¦h¥Ö½§ÃC¦âªº¡C¦b§Ú­Ì¦³¤@°Ï¤@°Ï¥Ö½§ªº°Ï°ì¤§«á¡A§Ú­Ì´N¨Ó§ä¥¦Ãä½tªºpixel¡A§ä¥X¨Ó¤§«á¡A°µ¤F¤@­Ó¤jÁxªº°²³]¡A©Ò¦³¤H¹³¹ÏµeªºÁy¡A¤@©w¦b¤W¥b³¡¡A¤£¥i¯àµe¦b¤U¥b³¡¡C¤HªºÁy¤@©w·|§Î¦¨¤@­Óclose loop¡A¦Ó¥B¥¦ªº­±¿n·|¤j©ó¾ã±i¹Ïµeªº4%¡C·|¥h§äµe¤W¥i¥H³s¦¨ªº°Ï°ì¡A¦Ó¥B³o­Ó°Ï°ìªº­±¿n¬O¤j©ó4%ªº¡C¦]¦¹§Ú­Ì¤jÁxªº°²³]¦pªG¦b¹Ïµeªº¤W¥b³¡§ä¨ì¤@­Ó­±¿n¶W¹L4%ªº«Ê³¬°Ï°ì¡A´N¥i¥H¿ë§O³o±iµe¬O¨v¹³ªºµe¡C¦Ó¥t¥~¤@­Ó¨Ò¤l¬O²ö¤º¤á¥~ªºµe¡A³o±iµe¬Omarch¡A¦³¤@Ã䪺¤ÑªÅ¡A¤w³Qµe®aµeªº¤Ó¶Â¤F¡A¤£¦b§Ú­Ìtrainingªº½d³ò¡A©Ò¥H§Ú­Ì¤£¥h¨ú¥¦¡C¤U¤@¨B´N¬O§äÃä½tªºpixel¡A³o¨Çpixel¬OÄÝ©óÃäÃ䪺pixel¡C§Ú­Ì¤S°µ¤@­Ó°²³]¡A©Ò¦³ªº¤ÑªÅ·|¦û¤@­Ó«Ü¤jªº­±¿n¡A¬ù¦û20%¥H¤W¡A§Ú­Ì§ä¥X¨Ó¬O³Ìªøªº¡A¦Ó¥B±`±`¬O²×¤î¦bÃä½t¡A¤£¤@©w¬O¤@­Óopenªº¡C­Y§ä¨ì¤@­Ó³ÌªøªºÃä¬O¤ÑªÅªºÃC¦â¡A´N¨M©w´N±iµe¬O­·´ºµe¡C³o¨â­Ó°²³]³£¦³¯ÊÂI¡A¥u¯à°µ²Å¦Xªº¹Ï¡C

µ²½×¦p¤U¡A²Ä¤@­Ó¦n³B¬O¥¦¥þ¬O¥Î¦Û°Ê¨Ó¿ë§Oªº¡A¤£¥Î¤H¤u¥h¿ï¾Üthreshold¡A¦Ó¦Û°Êªº¥h¿ë§O©Ò­nªº¹Ï¡C¸g¥Ñpre-specifiedªºkeyword¡C³o­Óalgorithm ¥i¥H°µ±m¦âªº¤ÀÃþ¡A­º¥ý¥i¥H±Npixel¤À¦¨·Q­nªººØÃþ¡A¦Ó¥B­pºâ«D±`ªºÂ²³æ¡C¤§«á¥i¥H§â¤£¦Pªºpixel¥i¥H§â¤£¦Pªº°Ï¶¡¤À¥X¨Ó¡A©Ò¥Îªº¤èªk¬O¥u­n¬OÄÝ©ó¦P¤@­Ó°Ï°ìªºpixel¡A¥¦°ò¥»¤W¬O³sÄò©Êªº¡A¦Ó¤£·|³æ¿W¦s¦b¡A©Ò¥Hpixel·|¶°¤¤¦b¤@°Ï¡C²Ä¤G­Ó¬O¥u¥Î¨ì¹Ïµeªº±m¦â¡A¥u¦³¦b«á­±¨BÆJ¤~¦³¥Î¨ìshape¡A©Ò¥H¥Î¥¦±m¦âªº³¡¤À¨Ó¤ÀÃþ¬O«D±`ªºÂ²³æ¡C¦Ó¥B­pºâªº®É¶¡«Üµu¡C³Ì«á¦b¿ëÃÑ¥¦¬O¨v¹³ÁÙ¬O¦ÛµM­·´º®É¡A¦³¥Î¨ì­±¿n¡A¦ý¬OshapeÁÙ¨S¦³¥Î¨ì¡A©Ò¥H¥u¥Î¨ì¥¦ªº±m¦â³¡¤À¡AÁÙ¦³­±¿n³¡¤À¡C

³o­Óproject ¥¼¨ÓÁÙ¦³«Ü¦h¥i¥H°µ¡A²Ä¤@­Ó¬O¥i¥H¼W¥[§ó¦hªººØÃþ¡A©Î¬Okeyword¡C²Ä¤G­Ó¥i¥H¥[¤J§ó¦hªº¤§«e©Î¤§«áªº¼v¹³³B²z¡A¨Ó¨Ï±oµ²ªG§ó¦n¡C¦p¥i¥H§âdataÀ£ÁY¡A¦]¬°¹Ïªºpixel³£«Ü¦h¡A¦pªG¥i¥H§â1000x1000Åܦ¨500x500¡A¦b­pºâªº®É¶¡¥i¥HÁY±o«Üµu¡C²Ä¤G­Óregion segmentation­n°µªº«Ü·Ç½T¡A¥Ø«e¨S¦³°µsimulationªº³¡¤À¡A­Y°µsimulationªº³¡¤À¡A´N­n­«·s¥Î­Ó§ó·Ç½Tªºregion segmentationªº¤èªk¡C

«Ý·|·|show¤@¨Ç¥Ø«e¦b°µªºresearch¡A¤]¬OÃö©ósegmentation¡A°Ï°ì¤À¹jªºªF¦è¡A¦ý¬O¬O¥Î¦b¶Â¥Õ¼v¹³¤è­±¡A¦Ó¥B¬O¥Î¦b¿Ë¤ldnaŲ©wªº¹Ï¹³¤W­±¡C

¦bdna¤W­±¡A§Ú­Ì©Òµo®iªº¬O¦Û°Ê¼Æ¥X¦³¦h¤Ö±ødnaªºlane¡A¦bband¤W­±¡A§Ú­Ì­n§@ªº¬O¼Æ¥X¨C±ølane¤W­±¦³¦h¤Ö±øband¡A¦bspot¤è­±¡A­n¦Û°Ê¦a¼Æ¥X¦³¦h¤Ö­Ó´³ÂI¡C²{¦b§â¹Ï¤ù¨qµ¹¤j®a¬Ý¡A¦]¬°¸Û¦p§Ú­è¤~©Ò»¡ªº¡A©Ò¦³ªº¤èªk³£¬O°Ó·~¾÷±K¡A¤£¯àª½±µ§i¶D¤j®a¡C

³o±i¹Ïµe´N¬O©Ò¿×ªºdnaªºband¡A³o¥i¯à¬O§Aª¨ª¨ªºdna¡A¦pªG§A¬O§Aª¨ª¨ªº¤p«Äªº¸Ü¡A¨º»ò§A¤@©w­n¬O§Aª¨ª¨¸ò§A¶ý¶ýªºÁp¶°¡C©Ò¥H¦pªG§A¥X¨Ó¦³§Aª¨¶ý¨S¦³ªº¡A¨º»ò§A´N¤£¬O§Aª¨¶ýªº¤p«Ä¡C²{¦b¬°¬Æ»òdnaÀË´úªº®É¶¡­n¨º»òºCªº­ì¦]´N¬O¡A²Ä¤@­Ó­n¶]¥X³o±i¹Ï¤ù´N­nªá¤W¥b¤Ñªº®É¶¡¡A¦]¬°¥¦©ñ¦b¹³ªG­á¤W­±ªºªF¦è¡A­n¦bªG­á¤W«õÀG¹D¡A§âdna©ñ¶i³o¨ÇÀG¹DùØ­±¡AµM«á¦A©ñ¶i¹q¸Ñ¦À¡AµM«ádna´N·|®Ú¾Ú¥¦ªº¤À¤l¶qªº­«¶q¶]¡A¶]¥X¨Ó·|¤À¥¬¤£¦Pªº§Îª¬¡A¦A¤ÀªR©¼¦¹¤§¶¡ªºÃö«Y¡A³Ì­«­n´N¬Oºâ¥X³o¨Ç¤À¤l¶q¡C©Ò¥H§Ú­Ì¼v¹³³B²zªº¤H­nÀ°§U¥L­Ì±q³o¨Ç¦ì¸m¨Ó¨M©w¤À¤l¶q¡A¦]¬°¶V­«ªº¶]ªº¶VºC¡C

²Ä¤@±i¹Ï¤ù¶i¨Ó«á¡A§Ú­Ì¤j·§¥uªá¥b¬íÄÁ´N¥i¥Hª¾¹D³o¨Ç©Ò¦³bandªº¦ì¸mÁÙ¦³¥¦¦³¦h¤Öªºlane¡A©Ò¥H¹q¸£°¨¤W§i¶D§A¥¦¦³8­Ólane¡A32­Óband¡A¥Îc++¨Óimplementªº¡C¦pªG¦³¤H³ßÅwªº¸Ü¡A´N¬O§â¨C­Óband®Ø°_¨Ó¡A©ÎªÌ¬O§â¨C­ÓbandªºÃä½t³£§ä¥X¨Ó¡A¨Óºâband©Ò¦ûªº­±¿n¡A³o¥u¬OÅã¥Ü¤è¦¡ªº¤£¦P¡C

²Ä¤G­Óexample´N½ÆÂøªº¦h¡A¥¦¦³¦n´X­Óchannel¡A©Ò¥H³o¶]¥X¨Ó¦³20±ø¡A¤¤¶¡¤@±ø¬OªÅªº¡C©Ò¥H³o¤]¤j·§ªá¤@¬íÄÁ´N¥i¥H§ä¥X¥¦¦³19­Ólane¡A77­Óband¡C

²Ä¤T­Ódemoªº¼v¹³´N¬O¦³®É­Ôchannel«õ±o¬n¤F¡A©ÎªÌ¹q¸Ñ¦Àªº¹q¦ì¤£¬O¹ï±o«Ü·Ç¡A©Ò¥H·|©¹¨âÃä¶]¡C³o­Ó¤ñ¸û¤j¡A©Ò¥Hªá¤F3¬íÄÁ§ä¥X¨Ó¥¦¦³10±ølane¦³146­Óband¡C

¦bdna¬Ý§¹¤§«á¡A³o­Ó¬O¤£¦Pªº¯fµß¡A¹³¸z¯f¬r¤Ó¤p¡A§Ú­Ì¤j·§¬O§ä¤@¨Ç²Óµß¡A¹³¤j¸z±ìµß¡C§A¥i¥Hµo²{³o¬O¶Â¥Õ¼v¹³¡A¥¦¦³¶Â¦âªºÂI©M¥Õ¦âªºÂI¡C¦Ó¥B«Ü§xÃøªº¡A¦]¬°¶Â¦âªºÂI¥ú½u¤@·Ó·|¤Ï¥ú¡C©Ò¥H¹q¸£­n¼Æ¥X¥Õ¦âªº¤ñ¸û®e©ö¡A¶Â¦âªº¤ñ¸ûÃø¡C³o­Ó¤j·§¤]¬O¤@¬íÄÁ¡A¥i¥H¼Æ¥X¥¦¦³13­Ó¥ÕÂI¡C¤ñ¸û¬D¾Ô©Êªº¬O·í¨â­Ó¥ÕÂI«D±`¾aªñªº®É­Ô¡A¦Ó¥B¤@¤j¤@¤p¡A«ç»ò¼Ë¤À¥X¨Ó¡CµM«á¶Â¦âÂI³s¦b¤@°_ªº®É­Ô¡A¥¦­n«ç»ò¼Æ¥X¨Ó¡A©Ò¥H¼Æ¥X¨Ó¦³16­Ó¶ÂÂI¡C

¸ò¤j®a³zÅS¤@ÂI°Ó·~¾÷±K¡A¦]¬°§Ú­Ìªº«È¤á³£ª¾¹D¡A¦ý¬O§A­Ì¤ñ¸ûÁo©ú¡A©Î³\Å¥¤F´N·|§@¡C

¦bdna¤è­±¡A§Ú­Ì©Ò¥Î¨ìªº¬O¤@«×ªÅ¶¡ªºpeak detection¡A´N¬O§ä¥¦ªºpeak¡C§â¨C­Óband¦b¤T«×ªÅ¶¡µ¹model¤@­Ó¤p¤s®p¡AµM«á¦A¥Îtraversal linking¡A¨Ìpixel³s°_¨Ó¡C¨º»ò¦bspot¤W­±¡A¬O¤G«×ªÅ¶¡ªºpeak detection¡A¦pªG§A­Ì§â¼v¹³ªºªø«×¬Ý¦¨¤T«×ªÅ¶¡ªº¸Ü¡A´N¬O2-dimension of peak detection¡A¤@¼Ë§âÃä½t³s°_¨Ó¡C¶Â¦âªºÂI´Ntricky¤@ÂI¡A¦]¬°­n¦Ò¼{¨ì¶Â¦â·Ó¨ì¿O¥ú·|¤Ï¥ú¡C

¦b­è­èÁ¿ªº±m¦â¤è­±¡A§Úªº¤½¥q¤]§Y±Nµo®i¤@¨Ç±m¦â³B²zªº»ö¾¹¡C­º¥ý­n¦p¦ó¥[±j±m¦â¼v¹³¡A¦]¬°±m¦â¼v¹³³£¦³noise¡C©Ò¥H¥ªÃ䦳noiseªº¼v¹³¸g¹L³B²z¥H«á¡A©Ò¦³ªº´³ÂI¥þ³¡¤£¨£¤F¡C

¦A¨q¤@­Ó¤ÀÃþ¨Ò¤l¡AÄ´¦p»¡¦³¤TºØ²Ó­M©Î¤TºØ¤£¦P¯f¬rªºÃC¦â¡C¹³§Ú­Ì­nÂŦ⪺²Ó­M¦Ó¤w¡A§Ú­Ì´N§äÂŦ⪺¡C¥Î­è­è±Ðµ¹¤j®aªº¤è¦¡¡A§ä¨ì¥¦ªºÃä½t¡A´N¥i¥H­pºâ¥X¨C¤@ÂŦâ²Ó­Mªº¯S©Ê¡G¥¦ªº­±¿n¡A¦h¶ê¡A³Ì¤j¸ò³Ì¤p­È¡A´N¬OùØ­±ªºpixel value¡C

©Ò¥H³oÃäµ²§ô§Ú¤µ¤Ñ¸ò¤j®a§@ªºpresentation¡A§Æ±æ¤j®a·|³ßÅw¡AÁÂÁ¤j®a¡I

--

q:¾ò¶ê¤£·|¦³¥æ¶°¶Ü?
a:
¦pªG¦³¥æ¶°®É¡A´N­n¥Î²Î­p¾Çªº¤@­Ó¤èªk¡A¦³¤@ÂI¦b¨â­Ó¾ò¶ê¥æ¶°ªº¦a¤è¡F¨Æ¹ê¤W¡A³o­Ó¾ò¶êªº¤j¤p¤]¬O¥Ñ©p¨Ó¨M©w¡C­n«ö·Ó¸ê®Æªº¯S©Ê¨Ó¿ï¾Ü¾ò¶êªº¤j¤p¡C³o¥u¬O¬Y­Ó°Ï¶¡¡A¦]¬°Gaussian function ÁÙ·|Ä~Äò¦Ü¨â­Ó¤@©w·|¦³¥æ¶°ªº¦a¤è¡C¦b¥æ¶°ªº¦a¤è¡A´N§âpixel®³¥X¨Ó¡A¬Ý¥¦¤ñ¸û¾aªñ¨º¤@­Ó¾ò¶ê¡A´N¬O¨º¤@­Óªº±Ú¸s¡C

q:³o¼Ë§P¿ë´N·|¦³¿ù»~¤F?
a:
·íµM·|¦³error¡A·í©p«D±`confuse®É¡A©Ò¥H¥i¥H¤£­n¿ï¾Ü¨º­Ópixel¡A³q±`ambiguousªºpixel ¬O¦b¤ñ¸ûdiscreteªº¦a¤è¡A¦]¬°§Ú­Ì­n§äªº¬O¤@­Óregion¡A©Ò¥H¥i¥H¤£­n¥¦¡C

q:¹ï©ó·Ó¤ù¡A¤]¬O¤@¼Ë¶Ü?
a:
¤]¬O­n¥h§ä«Ü¦h·Ó¤ù¨Ótraining¡A¨Ó§â¥¦Âà´«¦¨E©MS¡A«á§ä¥X¥¦ªº±Ú¸s¥X¨Ó¡C

q:¨º·Ó¤ù¤£¬O·|¤ñ¸ûxx(¤£²M·¡)¡A¦]¬°µe®aªº???
a:
¹ï¡A¤µ¤Ñ§Ú©Ò®i²{ªº«Y¼Æ¡A¥¦¬O¤£¯à¥Î¦bnatureªº¼v¹³¤è­±¡C¥¦¥u¬O¦b«Y¼Æªº®t§O¡A¦ý¤èªk¬O¥i¥HÀ³¥Î¦b·Ó¤ùªº¡C­Y¥H«á¦Û¤v­n°µ¡A­ì¨Ó§@ªÌSaber ¬O°µ¦b·Ó¤ù¤W¥Ö½§ªºÃC¦â¡A¥i¬OSaber ³o­Ó¤H¥i¯à¦³ºØ±Úª[µø¡A¦o¥u°µ¥Õ¤Hªº¥Ö½§¡A©Ò¥H¦oªºtest images¥þ³¡³£¬O¥Õ¤H¡C©Ò¥H­n¦Û¤v°µªº´N¬O³o­Ó³¡¤À¡C·|±o¨ì¤£¦P°Ï°ìªº¾ò¶ê§Î¡C¦ý¬O­ì²z¬O¤@¼Ëªº¡C¦Ó¥B·í¨â­Ó¾ò¶ê¾aªñªº®É­Ô¡A´N­n¥Î²Î­p¾Ç¤W¡A¬Ý¨º¤@­Óprobability¤ñ¸û¤j¡A¨Ó¨M©w¥¦¬OÄݩ󨺤@­Ó±ÚÃþ¡C

q:­è¤~¦³»¡¨ìColumbla Univ.ªº¦³°µ¤@¨Ç¨S¦³ºØ±Úª[µøªº¡A«á¨Óµo²{¦U­Ó¤HºØ³£¦³¤@¨Ç¦U¦Û¦@¦Pªº«Y¼Æ¡C
a:°ò¥»¤W½§¦â¸ò·Ó©ú«Ü¦³Ãö«Y¡A©Ò¥H¦³¨Ç¤H¥Ö½§«Ü¶Â¡A±j¥ú¤@¥´ÁÙ¬O·|Åܱo«Ü¥Õ¡C¹³±B¯½·Ó¡A¦³¨Ç¤H«Ü¶Â¡A¥i¬O¬Ý°_¨Ó«o«Ü¥Õ¡A´N¬O¦]¬°·Ó©úªºÃö«Y¡C©Ò¥H§Úı±oÃC®Æ¦³¥¦ªº¥»½è¡A©Ò¥HYES ³o­Ócolor space¬O¤j®a¥i¥H¦Ò¼{ªº¤@­Óspace¡C¥i¬O¥¦¤£¬O°ß¤@ªº¡C©Ò¥H²{¦b¦³«Ü¦hªºengineersÁÙ¦b§ä¾A¦X¦o­ÌÀ³¥Îªºcolor scaoe¡C§Ú¤µ¤Ñ¥u¬O´£¥X¤@­Óidea¦Ó¤w¡C

q:¤£¦Pµe®aªºµe­·¡A¥i¤£¥i¥H¥Ñ³]©w¥X¨Óªº«Y¼Æ¬Ý¥X¨Ó?
a:
¥i¥H¡A¹³¹F¤å¦èµeªº¬ü¤k³£«Ü»a¥Õ¡A¦Ó¹p¿Õº¸µeªº´N¤ñ¸û°·±d¡C·|§Î¦¨¤£¦Pªº±Ú¸s¡A¦ý¬O·|«Ü¾aªñ¡A¦]¬°°ò¥»¤W¡A³£¬O¥Ö½§ªºÃC¦â¡C©Ò¥H¦b¥Ö½§¤è­±¤ñ¸ûÃø¤À¿ë¡A¦ýµy·L¥i¥H¬Ý¥X¨Ó¡C¹³¤ÑªÅ¤è­±¡A¦pªG¬OÂÅÂŪº¡A¹³¨ººØ«C¦âªº¤ÑªÅ¡A¥¦´N¯S§Oªº²æÂ÷¥Õ¦âªº°Ï°ì¡C¦]¬°¤j³¡¤Àªº¤ÑªÅ¡A¶³³£«Ü¦h¡A©Ò¥H·|¤ñ¸û¾aªñ¥Õ¦âªº°Ï°ì¡C¸g¥Ñ§ë¼v©M­ì¨ÓªºÃC¦â¤£¤@¼Ë¡A©Ò¥H³o¬OÂà´«¹Lµ{¤¤ªº¤@­Ódistortion¡C

q:¿ù»~²v¤W¡A¦³¦h¤Ö?
a:
¥Ø«e¥u¦³¨â­Ótest¡A©Ò¥H¨S¦³¿ìªk«Ü¦³®Äªº´ú¥X¡A¦ý¬O³o­Óalgorithm¦b«á­±°²³]¦³«Ü¤jªºº|¬}¡A¦p­Y¬O¨â­Ó¤HÁyªº¨v¹³¡A¨º­±¿n´N¥i¯à·|´î¤Ö¡A©Ò¥H¥²¶·°µ§ó¦hªº¦Ò¶q¡C¦pªG¥Ö½§ÃC¦â©M¤ÑªÅÃC¦â¡A¨S¦³¶W¥X§Ú°V½mªº¼v¹³ªº½d³ò¤§¥~¡A´N¥i¥H¤À¿ëªº¥X¡C¦ý¬Oµe®a·|±`¥Î¤£¦PÃC¦â¨Óªí²{¡A¦p²¦¥d¯Á´N¦³¤@±iµµ¦âªºÁy¡A¨º´N§¹¥þ´ú¤£¥X¨Ó¡C©Ò¥H¦bÃC¦â¤è­±¬OÆZrobust¡A¥i¬O¦pªG§ó¦hªº¤H¡A©Î¬OÃC¦âªºÅܰʡA´N¨S¦³¿ìªk¤F¡C


¬ÛÃö¸ê®Æ

Introduction

In text, There are some important topics that have been introducted and we don't repeat it again. But some topics that don't introducte may make some mistake if somebody never learn about video or multimedia. We will talk about it's detail and where we find them out in the world wide web. Just as the speaking that there some topics and technicality but that don't touch it detial and what we will introduce are four topics - Color space, RGB, CMYK and Region-based segentation - There are also some relative topcs such as Query-by-Image-Content that can find out the digital image form your query.

Colorspace

1.what is colorspace

Color Space Definitions

Chromaticity Diagrams

The CIE 1931 xy

This color space is a mathematical representation of the human eye's response to color. Note that this color space is non-linear! It's a 3-D model that is "unfolded" like a Mecador projection map of the earth.

CIELAB

LAB color is the same model revised to be a linear color space. This model works best to define color on a computer because of its linear nature. A and B define the hue and saturation of the color where the L value or "Luminance" comprises the Z axis through the middle. The 0,0 point of the diagram is the "white reference" and defined as the color temperature. With the xyY defination of LAB (See "Printing Inks Setup" in Photoshop), the Y value can be calculated through a simple equation.

2.RGB

Exact color reproduction is and has always been the goal that is difficult to achieve ¡CMany applications use the Intenational de l¡¦Eclairage¡¦s CIE XYZ color model (or color space) as a basis to prepare a color rendering before it is output to a device¡CThe CIE XYZ color model provides a mathematical representation of all visible colors and is hardware-device-independent¡CHowever¡Aeven a device-independent color model does not work well in practice for following reasons:

  1. The color range of the printer and the display do not match since monitors use the RGB color model and printer use the CMYK color model¡Athe printer driver make the best suited for monitors since screen phosphors generate color by additive mixing¡COn the other hand ¡Athe CMYK color model is best suited for printers since cyan ¡Amagenta ¡Ayellow ¡Aand black use subtractie mixing to create colors on paper¡C
  2. The color ranger of a device varies with the technology as each technology has its own color range ¡C
  3. Even after the devices are calibrated¡Athe colors vary when printed¡C

So what is the solution ? The monitor screen colors are output to a printer consistenty ; that is ¡Athe application outputs the same color every time a paticular color is used¡CThe devices are regularly calibrated to ensure that the color rendering of the printer matches the screen output¡C

3.CMYK

  1. Introduction one CMYK publishes the very best work from students in advertising, design, illustration and photography. As judged by leaders in each field.CMYK is then sent to individual creative directors, art directors and principals at over 15,000 advertising agencies and design firms across the U.S. and abroad. Along with agency and design firm features, top student and school profiles , hints and tips from the pros, and much more, CMYK allows creative leaders the opportunity to see who's hot, and who to hire ¡V and gives students the chance to see what their competition is up to. With a total circulation of 30,000, CMYK is also used as the curriculum at the nation's art schools and universities, and can be found in leading book chains throughout the U.S., Canada, Japan and Sweden at Barnes and Noble, Borders, Crown Books, Tower Books and Keplers.
  2. Introduction two CMYK offers Scitex imagesetting for superior film output. Our Scitex system includes 2 fully equipped workstations and a large format Dolev 800 imagesetter for plotting films up to 32"x44". We also have an AGFA 7000 imagesetter for films up to 22"x26". We can provide you with positives or negatives at any line screen up to 625 lpi. Custom screens ara available, including Scitex FULLtone stochastic screening. For screenprinters, we offer custom screen angles and line screens as at 55 lpi or lower.We offer trapping with Scitex Full Auto Frames to provide automatic RIP trapping of your entire desktop lie on our Scitex workstations.

Save money by stripping your files electronically! We can provide paginations for QuarkXpress and Pagemaker files. We offer the QuarkXpress Xtension InPosition for printer spreads, 4-up's and 8-up's, or we can provide manual impositions.

4.Relation

The Commission International L'Eclairage (CIE) diagram of color space shows the range of visible color and the much smaller gamut ranges of RGB and YMCK color space. This is a 2-dimensional representation of something that is better described by a 3-dimensional model. The concept of three dimensions of color spaces was first articulated by Albert Munsell in 1914. Munsell defined even visual steps of color in hue, value (saturation), and chroma (equivalent grayness). Other early work in describing three-dimensional color space was contributed by Ostwald, who conceived that pure color hues should be lightened by adding white and darkened by adding black. His 3-D model took the shape of two equal cones joined at the base. In 4-color process, less dot density allows more light to be reflected from white paper, and colors are darkened by adding black, which is neutral and does not affect cast.

In 1931, the CIE refined the mathematical notation system for the three dimensions of color space by adding the concept of a standard observer with standard illumination. As the electronic pre-press industry seeks desirable standardization, a consensus formed using CIElab notation values (with 1,a,b being the coordinates for hue, saturation, and brightness, the evolutionary way of describing the three dimensions of color space originally detailed by Munsell). In the patented TRUMATCH approach to 4-color process, lighter values of a hue are achieved by reducing dot density and allowing more "white" paper to show through; darker colors are achieved achromatically by adding black. TRUMATCH affords the graphic designer a proportionately gradated color palette through the ability to select YMCK output values to 1% targets using desktop software and imagers. The CIE diagram illustrates the range of visible color space. Notice that RGB space and YMCK space are not congruent. Many RGB colors are not reproducible in YMCK space, and the reverse is also true. Furthermore, many solid-ink swatches fall beyond both gamut ranges. Publishing in 4-color process with the TRUMATCH Colorfinder for color selection will give an achievable palette of over 2,000 colors, blanketing the YMCK gamut range in smooth, proportionate steps.

Color space http://www.barco-usa.com/colorspa.htm

CMYK

  1. http://www.digitalcmyk.com/imaging.html
  2. http://www.cmykmag.com/

Relation http://www.trumatch.com/articles/anchor.htm

Region Segmentation

General Approach

There are two classes of techniques for color indexing: indexing by (1) global color distribution and (2) local or region color. An important distinction between these techniques is that indexing by global distribution enables only whole images to be compared while regional indexing enables matching between localized regions within images. Both techniques are very useful for retrieval of images and videos but are suited for different types of queries.

Color indexing by global distribution is most useful when user provides a sample image for the query. For example, if the user is interested in finding panoramic shots of a football game between two teams, then by providing one sample of the image others can be found that match the query image. Color indexing by global color distribution works well in this case because the user is not concerned with the positions of colored objects or regions in the images. However, when the user is interested in finding things within images the global color distribution does not provide the means for resolving the spatially localized color regions from the global distribution.

On the other hand, color indexing by localized or regional color provides for partial or sub-image matching between images. For example, if the user is interested in finding all images of sunsets where the sun is setting in the upper left part of the image, then regional indexing enables this query to be answered. Localized or regional color indexing is generally more difficult because it requires the effective extraction and representation of local regions. In both cases a system is needed for the automated extraction and efficient representation of color so that the query system provides for effective image and video retrieval. In this paper we present a technique that applies to both global and localized color region extraction and indexing.

The feature extraction system consists of two stages: (1) extraction of regions and (2) extraction of region properties as shown in Figure 1. An automated system performs both stages without human assistance. It is typically stage 1 that is either manual or semi-automatic when the system is not fully automated. The representation of region color is often more robust when the region extraction is not automated.

 

figure22


Figure 1:   General approach for color image feature extraction.

Region Extraction/Segmentation

The goal of the region extraction system is to obtain the spatial boundaries of the regions that will be of most interest to the user. The process of region extraction differs from image segmentation. Segmentation corresponds to a complete partitioning of the image such that each image point is assigned to one segment. With region extraction an image point may be assigned to many regions or to none. Conceptually, this is more desirable than segmentation because it supports an object-oriented representation of image content. For example, an image point corresponding to the wheel of a car can simultaneously belong to the two different regions that encapsulate respectively the wheel and the car as a whole.

There are several techniques for region extraction. The least complex method involves (1) manual or semi-automated extraction. In this process the images are evaluated by people and the pertinent information is confirmed or identified visually. This is extremely tedious and time-consuming for large image and video databases. Another procedure for region extraction utilizes a (2) fixed block segmentation of the images. By representing color content of small blocks independently there is greater likelihood that matches between regions can be obtained. However, it is difficult to pick the scale at which images should be best blocked. A third technique involves (3) color segmentation. There have been several techniques recently proposed for this such as color pairs [CLP94] and foreground object color extraction tools [HCP95].

We propose a new technique which partly employs the color histogram (4) back-projection developed by Swain and Ballard for matching images [SB91][SC95]. The basic idea behind the back-projection algorithm is that the most likely location of a spatially localized color histogram within an image is found by the back-projection onto the image of the quotient of the query histogram and the image histogram. More specifically, given query histogram g[m] and image histogram h[m], let tex2html_wrap_inline1282 . Then replace each point in the image by the corresponding confidence score B[m,n] = s[I[m,n]]. After convolving B[m, n] with a blurring mask, the location of the peak value corresponds to the most likely location of the model histogram within the image. In small image retrieval applications this computation is performed at the time of query to find objects within images [SB91][EM95]. However, for a large collection it is not feasible to compute the back-projection on the fly. A faster color indexing method is needed.

We extend the back-projection to the retrieval from large databases by precomputing for all images in the database the back-projections with predefined color sets. By processing these back-projections ahead of time, the system returns the best matches directly and without new computation at the time of the query. More specifically, we modify the back-projection algorithm so that it back-projects binary color sets onto the images. Instead of blurring the back-projected images we use morphological filtering to identify the color regions. This process is described in greater detail in later sections.

Region Feature Extraction

Once the image regions are identified, each region is characterized and represented using a feature set. The goal of color representation is to accurately capture the salient color characteristics of the region in a low-dimensional vector. Two choices for simple color features are (1) mean color and (2) dominant color of the region. They both use only a 3-D vector to represent the color of the region. The comparison measurement is fairly easy to compute but the discrimination is inadequate. An alternative is to use (3) color histograms. A color histogram is a high-dimensional feature vector typically having greater than 100 dimensions and the comparison of histograms is computationally intensive. They are best suited for representation of global color rather than local color regions because of storage requirements and the large number of computations required at query time.

We propose a new approach for representation of color content which is well matched to the binary color set back-projection region extraction. We represent region color using a (4) binary color set that selects only those colors which are sufficiently present in the region. Since the color sets are binary vectors, they can be indexed efficiently, i.e., using a binary tree, rather than requiring substantial computations to compare histograms at query time. The binary set representation of color regions and indexing is described in detail in the following sections.

Color Image Retrieval Systems

Recently, several systems have appeared for the content-based retrieval of images. The QBIC system [FFN93] provides for the retrieval of images by color, texture and shape. It supports two classes of color content - global and local. In the QBIC system the extraction of local regions is handled manually by requiring a person to draw region boundaries using a mouse. Both the global and local color information is represented by mean color and color histogram. The QBIC system also uses a quadratic distance metric for comparing histograms. Because the quadratic distance measure is very computationally intensive, the mean color distance is used as a pre-filter for color queries [HSE95].

VisualSEEk

We are currently developing VisualSEEk: a content-based image/video retrieval system for the World Wide Web. VisualSEEk supports query by color and texture and spatial-layout. Other features such as shape, motion and embedded text will be incorporated in the near future. The system (see Figure 2) provides the user with a highly functional and platform independent Java user-interface which collects the user's query. The query is communicated to the VisualSEEk server on the World Wide Web through the Common Gateway Interface (CGI). The server answers the user's query by accessing the extracted meta-data that describes the images and videos in the archive.

  figure63
Figure 2:   VisualSEEk external system operation.

The VisualSEEk system emphasizes several research goals - automated extraction of localized regions and features, efficient representation of features, preservation of spatial properties, extraction from compressed data [Cha95][CS95] and fast indexing and retrieval. The color indexing component of the system uses the binary color set back-projection algorithm to extract color regions. The user may search for images using both global and local features. For a local color region query the user indicates the locations of the color regions by positioning regions on the query grid (see Figure 3(a)). The returned images can be used for a global query whereby the user selects one image and uses the give me more feature of VisualSEEk to retrieve images that best match the selected one in a specified way (see Figure 3(b)).

  figure75
Figure 3:   (a) VisualSEEk Java user interface and (b) query returns with give me more feature of VisualSEEk.

Terminology

Thresholding

Thresholding converts a grey level image into a binary image by setting all pixel values above a threshold to 1 and all those below to zero.

Adaptive Thresholding

The process of thresholding with the automatic local selection of a theshold value. Generally this value is estimated from local image content.

Region Growing

An intermediate step in the process of many "image segmentation" algorithms which aims to merge adjacent regions with similar characteristics in order to obtain a simpler and hopefully more correct interpretation of the data.

Image Segmentation

The process of assigning classification groups to an image (often on a pixel by pixel basis) in order to isolate (segment) particular regions of scene structure. Often done as a precursor to higher level interpretation such as recognition or measurement (achieved via the processes of "representation" and "classification").

Split and Merge

A particualr pixel based approach to "image segmentation" whereby the best image segmentation is arrived at by iteratively combining the processes of "region growing" and sub-division.

Histograms

A histogram is an array of non negative integer counts from a set of data, which represents the frequency of occurance of values within a set of non-overlapping regions. For example, the image histogram is an array of the frequency of occurrence of grey levels within a particular set of grey level ranges.

Grey and Binary Scale Levels

A grey level is a single scalar value associated with a particular location in an image. For optical or photographic sensors this value is proportional (or at least monotonically related to) the measured signal. Some sensors or image processing algorithms require multi-valued data such as complex (two valued) or "colour" (three valued) images. Grey level images generally have integer values in the range 0-64, 0-256 or 0-1024, corresponding to 6-bit, 8-bit, or 10-bit digitisation. The binary image requires less storage, one bit per pixel, and is generally used to represent the presence or absence of a particular "feature" at each point in the image (eg: "edges");.

Grey Level Processing

Grey level processing is the general term given to image processing of "grey level" image data. As distict from "binary processing" which refers to processing of binary (single bit per pixel) images.

References

Cha95
S.-F. Chang. Compressed-domain techniques for image/video indexing and manipulation. In I.E.E.E. International Conference on Image Processing, October 1995.
CLP94
T.-S. Chua, S.-K. Lim, and H.-K. Pung. Content-based retrieval of segmented images. In ACM Multimedia 1994, October 1994.
CS95
S.-F. Chang and J. R. Smith. Extracting multi-dimensional signal features for content-based visual query. In SPIE Symposium on Visual Communications and Image Processing, May 1995. Best paper award.
EM95
F. Ennesser and G. Medioni. Finding waldo, or focus of attention using local color information. I.E.E.E. Transactions on Pattern Analysis and Machine Intelligence, 17(8), August 1995.
FFN93
C. Faloutsos, M. Flickner, W. Niblack, D. Petovic, W. Equitz, and R. Barber. Efficient and effective querying by image content. IBM RJ 9453 (83074), August 1993.
HCP95
W. Hsu, T. S. Chua, and H. K. Pung. An integrated color-spatial approach to content-based image retrieval. In ACM Multimedia 1995, November 1995.
HSE95
J. Hafner, H. S. Sawhney, W. Equitz, M. Flickner, and W. Niblack. Efficient color histogram indexing for quadratic form distance functions. I.E.E.E. Transactions on Pattern Analysis and Machine Intelligence (PAMI), July 1995.
Hun89
R. W. G. Hunt. Measuring Color. John Wiley and Sons, 1989.
Jon81
K. S. Jones. Information Retrieval Experiment. Butterworth and Co., 1981.
Rus95
J. C. Russ. The Image Processing Handbook. CRC Press, Boca Raton, 1995.
SB91
M. J. Swain and D. H. Ballard. Color indexing. International Journal of Computer Vision, 7:1 1991.
SC95
J. R. Smith and S.-F. Chang. Single color extraction and image query. In I.E.E.E. International Conference on Image Processing, October 1995.
WS82
G. Wyszecki and W. S. Stiles. Color Science: Concepts and Methods. John Wiley and Sons, 1982.

All of the information about Region Segmentation above is from these links :

  1. http://www.ctr.columbia.edu/~jrsmith/html/pubs/tatfcir/color.html
  2. http://noodle.med.yale.edu/chakrab/pap1/pap1.html
  3. http://luxor.stanford.edu/~ramani/thesis/fullthesis.html
  4. http://www.wiau.man.ac.uk/courses/cvmsc/RegSeg.htm

Image query

Image can be query from color¡Btexture¡Bshape and layout.On-line collections of images are growing larger and more common, and tools are needed to efficiently manage, organize, and navigate through them. We have developed the QBIC system which lets you make queries of large image databases based on visual image content -- properties such as color percentages, color layout, and textures occurring in the images. Such queries use the visual properties of images, so you can match colors, textures and their positions without describing them in words. Content based queries are often combined with text and keyword predicates to get powerful retrieval methods for image and multimedia databases.

Abstract

In this paper we propose a method for automatic color extraction and indexing to support color queries of image and video databases. This approach identifies the regions within images that contain colors from predetermined color sets. By searching over a large number of color sets, a color index for the database is created in a fashion similar to that for file inversion. This allows very fast indexing of the image collection by color contents of the images. Furthermore, information about the identified regions, such as the color set, size, and location, enables a rich variety of queries that specify both color content and spatial relationships of regions. We present the single color extraction and indexing method and contrast it to other color approaches. We examine single and multiple color extraction and image query on a database of 3000 color images.

Introduction

 There is an increasing need for ways to organize and filter the growing collections of image and video data. It is an extremely time consuming task to assign text descriptions to images and the inadequacy of textual annotations for visual data has been recognized. Recently, researchers have begun to investigate "content-based" techniques for indexing images using features such as color, texture and shape [1][2]. A successful content-based image database system requires the following components:

 * identification and utilization of intuitive visual features

 * effective feature representation and discrimination

 * automatic extraction of spatially localized features

 * techniques for efficient indexing

 In this paper we investigate the use of color for organizing and retrieving images and videos from databases. We maintain that color is an intuitive feature for which it is possible to utilize an effective and compact representation. Our approach automatically extracts the color content of isolated regions within images and builds efficient indexes to retrieve the regions over a large collection of images. The spatial localization of the color regions also allows for queries to include spatial positions and relationships between color regions. This gives a great power of expression for database queries that include both specification of color sets and relative and absolute spatial locations.

 Queries supported by the single color technique include the following examples: give me all images containing...

 a.) a large dark green area near top of image, i.e., trees

 b.) a yellowish-orange spot surrounded by blue, i.e., a sunset

 c.) a region composed of red, white and blue, i.e., a flag

 d.) an area with red and white in equal amounts, i.e., a checkered table cloth.

 We will explain how these queries can be answered using color sets with one or more colors, and/or by specifying spatial relationships and composition of regions.

Color Features

 Color may be one of the most straight-forward features utilized by humans for visual recognition and discrimination. However, people show the natural ability of using different levels of color specificity in different contexts. For example, people would typically describe an apple as being `red', probably implying some type of reddish hue. But in the context of describing the color of a car a person may choose to be more specific instead using the terms `dark red' or `maroon'. Color extraction by computer is performed without benefit of a context. Lack of knowledge also makes it difficult to cull the color information from the color distortion. The appearance of the color of real world objects is generally altered by surface texture, lighting and shading effects, and viewing conditions. Image database systems that use color retrieval must grapple with these problems of automated color image analysis.

Color histogram

 One common method for characterizing image content is to use color histograms. The color histogram for an image is constructed by counting the number of pixels of each color. Retrieval from image databases using color histograms has been investigated in [1][2][3]. In these studies the formulations of the retrieval algorithms follow a similar progression: (1) selection of a color space, (2) quantization of the color space, (3) computation of histograms, (4) derivation of the histogram distance function, (5) identification of indexing shortcuts. Each of these steps may be crucial towards developing a successful algorithm. But there has been no consensus about what are the best choices for these parameters. In [1] we evaluated the retrieval performance when several of these parameters were varied on a database of 500 color images.

 There are several difficulties with histogram based retrieval. The first of these is the high dimensionality of the color histograms. Even with drastic quantization of the color space, image histogram feature spaces can occupy over 100 dimensions in real valued space. This high dimensionality ensures that methods of feature reduction, pre-filtering and hierarchical indexing must be implemented. The large dimensionality also increases the complexity and computation of the distance function. It particularly complicates `cross' distance functions that include the perceptual distance between histogram bins. Another challenge with the use of color histograms is to enable the extraction of localized features.

Color image segmentation

 The extraction of spatially localized features is an extremely important aspect of image indexing. The isolated regions of interest within images should be identified and extracted independently from other regions in the image. For example, an image should be retrieved even when the user can describe only part of the image. If each image is represented by a single color histogram, this aspect of retrieval performance declines significantly. This is because extraneous information such as background colors may dominate the histogram.

Several attempts have been made to improve performance. In [1] images were segmented into fixed blocks and each block was indexed separately. In this way some blocks may still retain a reasonable characterization of objects of interest. On the other hand, the QBIC system [2] requires manual segmentation of images. In QBIC the color histograms are computed as attributes of the regions that have been cut out by hand. This reduces the potential contribution of background and other irrelevant colors but requires extensive human involvement in creation of the indexed data. Automated segmentation of images using color histograms may eventually provide useful results but has not yet been integrated into large image retrieval systems.

Single Color extraction

 The goal of the single color extraction method is to reduce the dimensionality of the color feature space while gaining the ability to localize color information spatially within images. Illustrated in Figure 1, we accomplish this through the following means: reduction of the full gamut of colors to a set of manageable size (~100 carefully selected colors). We avoid mapping unacceptably dissimilar colors into the same bins. We also allow higher tolerance in color lightness and color saturation while reserving the most fine quantization for hue. We utilize a `colorizing' algorithm to paint the color images using the reduced palette and a broad brush. This ensures that the most dominant colors and regions are emphasized. The processed images retain a visibly acceptable and compact representation of the color content. After this processing, we search over the set of colors remaining in the image, and map the regions that sufficiently contain the selected colors into a database index. The next section discusses the process in more detail.

Color space

 The RGB color format is the most common color format for digital images. The primary reason for this is because it retains compatibility with computer displays. However, the RGB space has the major drawback in that it is not perceptually uniform. Because of this, uniform quantization of RGB space gives perceptually redundant bins and perceptual holes in the color space. Furthermore, ordinary distance functions defined in RGB space will be unsatisfactory because perceptual distance is a function of position in RGB space.

Other color spaces, such as CIE-LAB, CIE-LUV and Munsell offer improved perceptual uniformity [4]. In general they represent with equal emphasis the three color variants that characterize color: hue, lightness, and saturation. This separation is attractive because color image processing performed independently on the color channels does not introduce false colors [5]. Furthermore, it is easier to compensate for many artifacts and color distortions. For example, lighting and shading artifacts will typically be isolated to the lightness channel. In general, these color spaces are often inconvenient due to the basic non-linearity in forward and reverse transformations with RGB space. For color extraction we utilize the more tractable HSV color space because it has the above mentioned characteristics and the transformation from RGB space is non-linear but easily invertible.

 The next issue after color space selection is quantization. The HSV color space can be visualized as a cone. The long axis represents value: blackness to whiteness. Distance from the axis represents saturation: amount of color present. The angle around the axis is the hue: tint or tone. Quantization of hue requires the most attention. The hue circle consists of the primaries red, green and blue separated by 120 degrees. A circular quantization at 20 degree steps sufficiently separates the hues such that the three primaries and yellow, magenta and cyan are represented each with three sub-divisions. Saturation and value are each quantized to three levels yielding greater perceptual tolerance along these dimensions. The quantized HSV space appears in Figure 3.

Color processing

 To identify color regions, the images are transformed to the quantized HSV space with 166 color bins and subsampled to approximately 196x196 such that correct aspect ratio is preserved. This generally reduces the image content to less than 50 colors. Even after the transformation it is still premature to isolate color regions because small details and spot noises interfere. We reduce most of this insignificant detail by using a colorizing algorithm. This processing is accomplished using a 5x5 median filter on each of the HSV channels. This non-linear filtering in HSV space does not introduce false hues. The color image is then converted back to an indexed RGB space. Table 1 reports the statistics of color processing of 3000 color images.

Color region labelling

 The next step involves the extraction of the color regions from the images. This is done by systematically selecting from the colors present in the image one at a time, and in multiples, each time generating a bi-level image. The levels correspond to the selected and un-selected pixels for the specified color set. Refer to Figure 2 and Table 2 for an illustration of region extraction and representation of the Butterfly color image. Next follows a sequential labelling algorithm that identifies the isolated regions within the image. The characteristics of each color region are evaluated in regards to several thresholds to determine whether the region will be added to the database. The first threshold is one for region size. In our system the region must contain more than 64 pels to be significant. This value still allows for sufficiently small regions to be indexed.

If more than one color is represented in the color set we utilize two additional thresholds. The first threshold is the absolute contribution of each color. If a color does not contribute at least 64 pels to the region, the region is not added. Furthermore, the relative contribution is also measured. All colors must contribute to at least 20% of the region area. Notice that this produces a firm limit of 5 colors per color region although, we use only up to 3 colors at a time. If a color region does not pass one of these thresholds then it will not be indexed by that color set. If a region is rejected because one of the colors from the color set is not sufficiently represented, the region still has a chance to be extracted using a reduced color set leaving out the under-represented color. Enforcing the color set thresholds prevents the unnecessary and redundant proliferation of indexed multiple-color regions.

Color image mining

 Even with the reasonably small color gamut it is necessary to search systematically for multiple color regions. Otherwise, it will require 2m passes over the image to test all combinations of m colors. We utilize a heuristic similar to that used for database mining [6]. The algorithm makes multiple passes over each image, expanding only the color sets that meet minimum support constraints. A color set Ci of binary colors is explored for an image only if for all colors k in Ci, where Ci[k]=1, there are at least t0 pixels in the image of color k such that t1 pixels of color k have not yet been allocated to a color region. We use t0 and t1 = 64. If t0 is not met then Ci will have colors that cannot be represented sufficiently by any color regions. Exploring this color set and all supersets of it would be futile. If t0 is met while t1 is not, then a color region containing all of the colors in Ci can alternatively be reconstructed using subsets of Ci and spatial composition. Therefore, exploration of Ci and its supersets generate redundant information.

 Figure 4 illustrates an example of the extraction of an American flag in the San Francisco color image. The region was extracted in whole while searching over color sets in the extraction process. The region and color set met the constraints to allow the region to be extracted. The users request for a {red, white, blue} region is answered with the minimum bounding rectangle in Figure 4(d) that represents the region.

Color specification and spatial positions

 The color characteristics specified by the user are represented using the m-dimensional binary color vector. The values may be obtained by picking colors from a color chooser, by navigating visually through 3-D color space, or by textual specification. The binary color vector will be quickly matched to region data because we allow only up to three colors per color set for each indexed region. This sparse binary vector representation of the color sets makes it far easier to index the color distributions than that needed for the color histogram techniques.

After the color characteristics of the regions have been determined, the spatial positions and relationships between regions can be specified by the user. The spatial characteristics of the color region query can be handled using one of several techniques that have been devised for representing and querying spatial information [7][8].

Conclusions and future work

 Single color query is an extremely useful content-based query tool for users of image and video databases. We proposed a method for automatically extracting the single and multiple color regions within images. The color extraction approach allows interesting queries to be formulated based on size, shape and spatial relations of the color regions. The single color approach allows the user to specify the color content and spatial positions of region within images. Single color extraction and indexing is supported in the Content-Based Visual Query System being developed at Columbia University for a variety of image and video applications.

References

 [1.] John R. Smith and Shih-Fu Chang, "Tools and Techniques for Color Image Retrieval," submitted to ACM Multimedia 95.

 [2.] C. Faloutsos, et. al., "Efficient and Effective Querying by Image Content," IBM RJ 9453 (83074), August 3, 1993.

 [3.] M. Swain and D. Ballard, "Color Indexing," International Journal of Computer Vision, 7:1, 1991, p. 11 -- 32.

 [4.] G. Wyszecki and W. S. Stiles, Color Science: Concepts and Methods, John Wiley & Sons, 1982.

 [5.] John C. Russ, The Image Processing Handbook, IEEE Press, 1995.

 [6.] R. Agrawal, et. al, "Mining Association Rules between Sets of Items in Large Database," ACM SIGMOD-93, Washington, DC, May, 1993.

 [7.] T. Gevers and A.W.M. Smeulders, "An Approach to Image Retrieval for Image Databases," Database and Expert System Applications (DEXA-93), 1993.

 [8.] S. K. Chang, et. al., "An Intelligent Image Database System," I.E.E.E. Transactions on Software Engineering, Vol. 14, No. 5, May 1988.

[9.]http://wwwqbic.almaden.ibm.com/

[10.]http://www.ctr.columbia.edu/~jrsmith/html/pubs/ICIP-95-2/single1.html

 To appear at the International Conference on Image Processing (ICIP-95), Washington, DC, Oct. 1995

   

FIGURE 1. Car color Image: (1) conversion to HVS color space, (2) quantization of HVS space, (3) color median filtering, (4) conversion to indexed RGB space, (5) the processed color image has dominant color regions emphasized.

   
   

FIGURE 2. (a) Butterfly color image, (b) processed color image with 30 colors, (c) pixels from image (b) belonging to color set Ci, (d) minimum bounding rectangles (MBRs) for extracted regions used to index the image collection.

   

FIGURE 3. Quantized HSV color space, 18 hues, 3 saturations and 3 values + 4 grays = 166 colors.

   
   

FIGURE 4. (a) San Francisco color image, (b) processed image with 73 colors, (c) pixels belonging to color set = {red, white, blue}, (d) extracted color region as present in index.


¬ÛÃöºô¯¸

  1. By-Her Wang, "Home Page" : http://robotics.stanford.edu/~bwang/
  2. By-Her Wang, "Project in Color Segmentation" : http://robotics.stanford.edu/~bwang/ee368.html
  3. John R. Smith and Shih-Fu Chang, "Tools and Techniques for Color Image Retrieval" : http://www.ctr.columbia.edu/~jrsmith/html/pubs/tatfcir/color.html
  4. "Annotated Computer Vision Bibliography: Table of Contents" : http://cmp.felk.cvut.cz/~kraus/Mirrors/iris.usc.edu/Vision-Notes/bibliography/contents.html
  5. Pascal Bertolino, "TIMC-IMAG, GRENOBLE" : http://www-timc.imag.fr/~bertolin/index.shtml
  6. "Region Segmentation" : http://www.wiau.man.ac.uk/courses/cvmsc/RegSeg.htm
  7. Ramani Pichumani, "Construction of a Three-Dimensional Geometric Model for Segmentation and Visualization of Cervical Spine Images" : http://luxor.stanford.edu/~ramani/thesis/fullthesis.html
  8. Amit Chakraborty, L.H. Staib, James S. Duncan, "Deformable Boundary Finding Influenced by Region Homogeneity" : http://noodle.med.yale.edu/chakrab/pap1/pap1.html
  9. Sven Schröter, "Automatic Calibration of Lookup-Tables for Color Image Segmentation" : http://ls7-www.informatik.uni-dortmund.de/~schroete/paper/html/erlangen97/paper.html
  10. John R. Smith, "Integrated Spatial and Feature Image Systems: Retrieval, Analysis and Compression" : http://disney.ctr.columbia.edu/jrsthesis/thesissmall.html
  11. Computer Graphics System Development Corporation : http://www.cgsd.com/
  12. "CGSD Gamma Correction and Color Space": http://cgsd.com/papers/gamma_colorspace.html
  13. "YCC color space": http://www.aols.com/colorite/yccspace.html
  14. "A Guided Tour of Color Space": http://www.inforamp.net/~poynton/papers/Guided_tour/abstract.html
  15. "The CIE Lab color space": http://www.linocolor.com/colorman/sp_ciela_2.htm
  16. "Enter color space studio ":http://www.tinymind.com/bigb/colorspace/index.html
  17. "Anchoring Color Space": http://www.trumatch.com/articles/anchor.htm
  18. "RGB Technology": http://www.rgbtec.com/