Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
2001
Video database and video on demand represent important applications of the evolving Global Information Infrastructure. Conventional database management systems (DBMS) are not well-suited for video querying and retrieval because they are designed to operate on alphanumeric data where selection conditions are easier to specify. Defining similarity between video data is difficult since the similarity involves semantics of the video, which is inherent in the data itself. We are planning to integrate two approaches to modeling, indexing and querying video data. The first one is based on textual annotations and the second is based on object motion and digital video processing techniques, in order to support content based and intelligent access. In addition to this novel data model, we also plan to develop a s ystem prototype that can synergistically use multiple media modes to retrieve data using fuzzy information fusion.
IEEE Transactions on Multimedia, 2003
In the past few years, modeling and querying video databases have been a subject of extensive research to develop tools for effective search of videos. In this paper, we present a hierarchal approach to model videos at three levels, object level ( ), frame level ( ), and shot level ( ). The model captures the visual features of individual objects at , visual-spatio-temporal (VST) relationships between objects at , and time-varying visual features and time-varying VST relationships at . We call the combination of the time-varying visual features and the time-varying VST relationships a Content trajectory which is used to represent and index a shot. A novel query interface that allows users to describe the time-varying contents of complex video shots such as those of skiers, soccer players, etc., by sketch and feature specification is presented. Our experimental results prove the effectiveness of modeling and querying shots using the content trajectory approach.
Multimedia Tools and Applications, 2005
With the advances in information technology, the amount of multimedia data captured, produced, and stored is increasing rapidly. As a consequence, multimedia content is widely used for many applications in today's world, and hence, a need for organizing this data, and accessing it from repositories with vast amount of information has been a driving stimulus both commercially and academically. In compliance with this inevitable trend, first image and especially later video database management systems have attracted a great deal of attention, since traditional database systems are designed to deal with alphanumeric information only, thereby not being suitable for multimedia data. In this paper, a prototype video database management system, which we call BilVideo, is introduced. The system architecture of BilVideo is original in that it provides full support for spatio-temporal queries that contain any combination of spatial, temporal, object-appearance, external-predicate, trajectory-projection, and similaritybased object-trajectory conditions by a rule-based system built on a knowledge-base, while utilizing an objectrelational database to respond to semantic (keyword, event/activity, and category-based), color, shape, and texture queries. The parts of BilVideo (Fact-Extractor, Video-Annotator, its Web-based visual query interface, and its SQL-like textual query language) are presented, as well. Moreover, our query processing strategy is also briefly explained.
2002
Advances in compression techniques, decreasing cost of storage, and high-speed transmission have facilitated the way video is created, stored and distributed. As a consequence, video is now being used in many application areas. The increase in the amount of video data deployed and used in today's applications not only caused video to draw more attention as a multimedia data type, but also led to the requirement of efficient management of video data. Management of video data paved the way for new research areas, such as indexing and retrieval of videos with respect to their spatio-temporal, visual and semantic contents. In this paper, semantic content of video is studied, where video metadata, activities, actions and objects of interest are considered within the context of video semantic content. A data model is proposed to model video semantic content, which is extracted from video data by a video annotation tool. The work in this paper constitutes a part of a video database system to provide support for semantic queries.
Multimedia Tools and Applications, 1997
The increasing development of advanced multimedia applications requires new technologies for organizing and retrieving by content databases of still digital images or digital video sequences. To this aim image and image sequence contents must be described and adequately coded. In this paper we describe a system allowing content-based annotation and querying in video databases. No user action is required during the database population step. The system automatically splits a video into a sequence of shots, extracts a few representative frames (said r-frames) from each shot and computes r-frame descriptors based on color, texture and motion. Queries based on one or more features are possible. Very interesting results obtained during the severe tests the system was subjected to are reported and discussed.
Information Sciences, 2002
We propose a novel architecture for a video database system incorporating both spatio-temporal and semantic (keyword, event/activity and category-based) query facilities. The originality of our approach stems from the fact that we intend to provide full support for spatio-temporal, relative object-motion and similarity-based objecttrajectory queries by a rule-based system utilizing a knowledge-base while using an object-relational database to answer semantic-based queries. Our method of extracting and modeling spatio-temporal relations is also a unique one such that we segment video clips into shots using spatial relationships between objects in video frames rather than applying a traditional scene detection algorithm. The technique we use is simple, yet novel and powerful in terms of effectiveness and user query satisfaction: video clips are segmented into shots whenever the current set of relations between objects changes and the video frames, where these changes occur, are chosen as keyframes. The directional, topological and third-dimension relations used for shots are those of the keyframes selected to represent the shots and this information is kept, along with frame numbers of the keyframes, in a knowledge-base as Prolog facts. The system has a comprehensive set of inference rules to reduce the number of facts stored in the knowledge-base because a considerable number of facts, which otherwise would have to be stored explicitly, can be derived by rules with some extra effort.
Proceedings of the seventh ACM international conference on Multimedia (Part 1), 1999
Multimedia databases have been the subject of extensive research for the last several years. In particular, semantic modeling of video data spurred tremendous interest and produced various formalisms for content-based retrieval of video data. In this paper, we propose the use of Hierarchical Petri-nets (HPN's) for multi-level representation of video data with each level consisting of an augmented Petrinet structure. We show how such representation is able to capture video data semantics, from very low level semantics such as scene change and object movements to higher level semantics involving highlevel textual description of events. A Petri-net based method for query processing of multi-object spatiotemporal video-data queries is presented.
2002
Current advances in multimedia technology enable ease of capturing and encoding digital video. As a result, video data is rapidly growing and becoming very important in our life. It is because video can transfer a large amount of knowledge by providing combination of text, graphics, or even images. Despite the vast growth of video, the effectiveness of its usage is very limited due to the lack of a complete technology for the organization and retrieval of video data. To date, there is no "perfect" solution for a complete video data-management technology, which can fully capture the content of video and index the video parts according to the contents, so that users can intuitively retrieve specific video segments. We have found that successful content-based video datamanagement systems depend on three most important components: key-segments extraction, content descriptions and video retrieval. While it is almost impossible for current computer technology to perceive the content of the video to identify correctly its key-segments, the system can understand more accurately the content of a specific video type by identifying the typical events that happens just before or after the key-segments (specific-domainapproach). Thus, we have proposed a concept of customisable video segmentation module, which integrates the suitable segmentation techniques for the current type of video. The identified key-segments are then described using standard video content descriptions to enable content-based retrievals. For retrieval, we have implemented XQuery, which currently is the most recent XML query language and the most powerful compared to older languages, such as XQL and XML-QL.
2003
Abstract The BilVideo video database management system provides integrated support for spatiotemporal and semantic queries for video. A knowledge base, consisting of a fact base and a comprehensive rule set implemented in Prolog, handles spatio-temporal queries. These queries contain any combination of conditions related to direction, topology, 3D relationships, object appearance, trajectory projection, and similarity-based object trajectories.
Workshop on Multimedia Information Systems, 2002
The most useful environments for advancing research and development in video databases are those that provide complete video database management, including (1) video preprocessing for content representation and indexing, (2) storage management for video, metadata and indices, (3) image and semantic -based query processing, (4) real- time buffer management, and (5) continuous media streaming. Such environments support the entire process
1999
... The sharing and reuse of the video material is supported by the idea of logical VideoStream. ... 231 be extracted automatically from the visual content of the video data if machine vision and image processing techniques were more sophisticated. 3. Information retrieval ...
2002
In this paper we describe the first full implementation of a content-based indexing and retrieval system for MPEG-2 and MPEG-4 videos. We consider a video as a collection of spatiotemporal segments called video objects; each video object is a sequence of video object planes. A set of representative video object planes is used to index each video object. During the database population, the operator, using a semi-automatic outlining tool we developed, manually selects video objects and insert some semantical information.
Proceedings 18th International Conference on Data Engineering, 2002
Multimedia Tools and Applications, 1998
We present an effective technique for automatic extraction, representation, and classification of digital video, and a visual language for formulation of queries to access the semantic information contained in digital video. We have devised an algorithm that extracts motion information from a video sequence. This algorithm provides a low-cost extension to the motion compensation component of the MPEG compression algorithm. In this paper, we present a visual language called VEVA for querying multimedia information in general, and video semantic information in particular. Unlike many other proposals that concentrate on browsing the data, VEVA offers a complete set of capabilities for specifying relationships between the image components and formulating queries that search for objects, their motions and their other associated characteristics. VEVA has been shown to be very expressive in this context mainly due to the fact that many types of multimedia information are inherently visual in nature.
2004
In this thesis, we have extended a video database management system, called BilVideo, with semantic querying capability. Our work is based on a video data model for the extraction and storage of the semantic contents of videos, and a query language to support semantic queries on video data. The Web based query interface of BilVideo has also been modified to handle semantic queries both visually and textually.
Lecture Notes in Computer Science, 2000
A constraint of existing content-based video data models is that each modeled semantic description must be associated with time intervals exactly within which it happens and semantics not related to any time interval are not considered. Consequently, users are provided with limited query capabilities. This paper is aimed at developing a novel model with two innovations: (1) Semantic contents not having related time information can be modeled as ones that do (2) Not only the temporal feature of semantic descriptions, but also the temporal relationships among themselves are components of the model. The query system is by means of reasoning on those relationships. To support users' access, a video algebra and a video calculus as formal query languages, which are based on semantic relationship reasoning, are also presented.
Lecture Notes in Computer Science, 2000
Modeling video data poses a great challenge since they do not have as clear an underlying structure as traditional databases do. We propose a graphical object-based model, called VideoGraph, in this paper. This scheme has the following advantages: (1) In addition to semantics of video individual events, we capture their temporal relationships as well. (2) The inter-event relationships allow us to deduce implicit video information. (3) Uncertainty can also be handled by associating the video event with a temporal Boolean-like expression. This also allows us to exploit incomplete information. The above features make VideoGraph very exible in representing various metadata types extracted from diverse information sources. To facilitate video retrieval, we a l s o i n troduce a formalism for the query language based on path expressions. Query processing involves only simple traversal of the video graphs.
Faculty of Science and Technology, 2002
Current advances in multimedia technology enable ease of capturing and encoding digital video. As a result, video data is rapidly growing and becoming very important in our life. It is because video can transfer a large amount of knowledge by providing combination of text, graphics, or even images. Despite the vast growth of video, the effectiveness of its usage is very limited due to the lack of a complete technology for the organization and retrieval of video data. To date, there is no "perfect" solution for a complete video data-management technology, which can fully capture the content of video and index the video parts according to the contents, so that users can intuitively retrieve specific video segments. We have found that successful content-based video datamanagement systems depend on three most important components: key-segments extraction, content descriptions and video retrieval. While it is almost impossible for current computer technology to perceive the content of the video to identify correctly its key-segments, the system can understand more accurately the content of a specific video type by identifying the typical events that happens just before or after the key-segments (specific-domainapproach). Thus, we have proposed a concept of customisable video segmentation module, which integrates the suitable segmentation techniques for the current type of video. The identified key-segments are then described using standard video content descriptions to enable content-based retrievals. For retrieval, we have implemented XQuery, which currently is the most recent XML query language and the most powerful compared to older languages, such as XQL and XML-QL.
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.