Article thumbnail
Location of Repository


By 林順喜


[[abstract]]For the reason of complexity and versatility in current multimedia digital archives, it's not easy to have a good searching interface for people to access content easily without backend metadata text labeling. It's a tedious and labor-intensive work for professional editors to enrich the library content. For the efficient query, leaving the labeled metadata alone could also improve the searching exactly. The major goal in this project is to construct a useful tool. It could help in content-based indexing and automatic content-based categorization by utilizing state-of-the-art technologies. The technologies include keyword extraction, thesaurus construction, pattern recognition, speech recognition, and content-based image retrieval. Second, by utilizing the indexing mechanism built in former processing stage, a multi-modal intelligent interactive search interface has been designed in this project. Users could access multimedia content by entering input they prefer, like text, image or speech. The backend technologies developed in this project can not only help to organize multimedia content but also help to categorize content-based information which is usually confusing in metadata. In retrieval interface system, to lead people accessible to abundant library content, a novel and convenient multi-modal human interface integrating technologies of speech recognition, content-based image retrieval and metadata text search has been established.

Topics: 數位典藏;多模式;語音辨識;圖像檢索;文字資訊檢索;語音資訊檢索, Digital Libraries;Multi-Modal;Speech Recognition;Image Retrieval;Text Information Retrieval;Speech Information Retrieval, [[classification]]42
Year: 2011
OAI identifier:
Download PDF:
Sorry, we are unable to provide the full text but you may find it at the following location(s):
  • (external link)
  • Suggested articles

    To submit an update or takedown request for this paper, please submit an Update/Correction/Removal Request.