SAPIR: towards Large Scale Multimedia Content Search
MetadataShow full item record
Existing web search technologies are limited to text-based search, yet still 99% of the information on the web consists of audio-visual content that is searchable only by associated metadata and not by its actual content. This amazing restriction has raised the question of how search technologies can tap into the potential reservoir of information. SAPIR (Search on Audio-visual content using Peer-to-peer Information Retrieval) has developed cutting-edge technology to break the barriers and enable search engines to search large-scale, audio-visual information, by content. SAPIR’s distributed P2P technology has proven to be able to effectively deal with the fundamental scalability issue. The main aim of SAPIR has been to develop theories and technologies for nextgeneration search techniques that would effectively and efficiently deliver relevant information in the presence of exponentially growing (i.e., dynamic) volumes of distributed multimedia data. Fundamental to our approach is the development of scalable solutions that address the requirements of future generations of massively distributed data produced in a variety of applications and available on the Web.