CA2507248A1 - System and method for annotating multi-modal characteristics in multimedia documents - Google Patents
System and method for annotating multi-modal characteristics in multimedia documents Download PDFInfo
- Publication number
- CA2507248A1 CA2507248A1 CA002507248A CA2507248A CA2507248A1 CA 2507248 A1 CA2507248 A1 CA 2507248A1 CA 002507248 A CA002507248 A CA 002507248A CA 2507248 A CA2507248 A CA 2507248A CA 2507248 A1 CA2507248 A1 CA 2507248A1
- Authority
- CA
- Canada
- Prior art keywords
- video
- audio
- selection
- multimedia content
- permitting
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims description 21
- 230000003936 working memory Effects 0.000 claims description 12
- 238000002955 isolation Methods 0.000 abstract description 2
- 238000001514 detection method Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000001939 inductive effect Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/16—Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/48—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
- G06F40/169—Annotation, e.g. comment data or footnotes
Abstract
A manual annotation system of multi-modal characteristics in multimedia files.
There is provided an arrangement for selection an observation modality of video with audio, video without audio, audio with video, or audio without video, to be used to annotate multimedia content. While annotating video or audio features is isolation results in less confidence in the identification of features, observing both audio and video simultaneously and annotating that observation results in a higher confidence level.
There is provided an arrangement for selection an observation modality of video with audio, video without audio, audio with video, or audio without video, to be used to annotate multimedia content. While annotating video or audio features is isolation results in less confidence in the identification of features, observing both audio and video simultaneously and annotating that observation results in a higher confidence level.
Description
SYSTEM AND METHOD FOR ANNOTATING MULTI-MODAL
CHARACTERISTICS IN MULTIMEDIA DOCUMENTS
Field of the Invention The present invention relates to the computer processing of multimedia files.
More specifically, the present invention relates to the manual annotation of multi-modal events, objects, scenes, and audio occurring in multimedia files.
Background of the Invention Multimedia content is becoming more common both on the World Wide Web and local computers. As the corpus of multimedia content increases, the indexing of to features within the content becomes more and more important. Observing both audio and video simultaneously and annotating that observation results in a higher confidence level.
Existing multimedia tools provide capabilities to annotate either audio or video separately, but not as a whole. (An example of a video-only annotation tool is 15 the IBM MPEG7 Annotation Tool, inventors J. Smith et al., available through [http://]www.alphaworks.ibm.com/tech/videoannex. Other conventional arrangements are described in: Park et al, "iMEDIA-CAT: Intelligent Media Content Annotation Tool", Proc. International Conference on Inductive Modeling (ICIM) 2001, South Korea, November, 2001; and Minka et al., "Interactive Learning using a Society of Models," Pattern Recognition, Vol. 30, pp. 565, 1997, TR #349.
It has long been recognized that annotating video or audio features in isolation results in a less confidence of the identification of the features.
In view of the foregoing, a need has been recognized in connection with providing improved systems and methods for observing and annotating multi-modal events, objects, scenes, and audio occurring in multimedia files.
Summary of the Invention In accordance with at least one presently preferred embodiment of the present 1o invention, there are broadly contemplated multimedia annotation systems and methods that permit users to observe solely video, video with audio, solely audio, or audio with video and to annotate what has been observed.
In one embodiment, there is provided a computer system which has one or more multimedia files that are stored in a working memory. The mufti-modal 15 annotation process displays a user selected multimedia file, permits the selection of a mode or modes to observe the file content, annotates the observations; and saves the annotations in a working memory (such as a MPEG-7 XML file).
In summary, one aspect of the invention provides an apparatus for managing multimedia content, the apparatus comprising: an arrangement for supplying multimedia content; an input interface for permitting the selection, for observation, of at least one of the following modes associated with the multimedia content: an audio portion that includes video; and a video portion that includes audio; and an arrangement for annotating observations of a selected mode.
A further aspect of the invention provides a method of managing multimedia content, the method comprising the steps of: supplying multimedia content;
permitting the selection, for observation, of at least one of the following modes associated with the multimedia content: an audio portion that includes video;
and a video portion that includes audio; and annotating observations of a selected mode.
to Furthermore, an additional aspect of the invention provides a program storage device readable by machine, tangibly embodying a program of instructions executable by the machine to perform method steps for managing multimedia content, the method comprising the steps of: supplying multimedia content; permitting the selection, for observation, of at least one of the following modes associated with the multimedia 15 content: an audio portion that includes video; and a video portion that includes audio;
and annotating observations of a selected mode.
For a better understanding of the present invention, together with other and further features and advantages thereof, reference is made to the following description, taken in conjunction with the accompanying drawings, and the scope of 20 the invention will be pointed out in the appended claims.
Brief Descriution of the Drawings Figure 1 is a block diagram depicting a multi-modal annotation system.
Figure 2 is an illustration of a system annotating video scenes, objects, and events.
Figure 3 is an illustration of a system annotating audio with video.
Figure 4 is an illustration of a system annotating audio without video.
Description of the Preferred Embodiments Figure 1 is a block diagram of one preferred embodiment of a multi-modal annotation system in accordance with the present invention. The multimedia content and previous annotations are stored on the storage medium 100. When a user 130 selects a multimedia file via the annotation tool from the storage medium 100, it is loaded into working memory 110 and portions of it displayed in the annotation tool 120. At any time, the user 130 may also request that previously saved annotations associated with the current multi-modal file be loaded from the storage medium into working memory 110. The user 100 views the multimedia data by making requests through the annotation tool 120. The user 130 then annotates his observations and the annotation tool 120 saves these annotations in working memory 110. The user can at anytime request the annotation tool 120 to save the annotation on the storage medium 100.
Figure 2 is an illustration of a system annotating video scenes, objects, and events. (Simultaneous reference should also be made to Fig. l.) The multimedia data has been loaded from the storage medium 100 into working memory 110. A video tab 290 has been- selected. The multimedia video has been segmented using scene 5 changed detection into shots. A shot list window 200 displays a portion of the shots in the multimedia. Here, the user 130 has selected a shot 2I0 which is highlighted in the shot list window 200. A key frame 220, which is a representative shot in the frames of a shot, is preferably displayed. In addition, the frames of the shot maybe viewed in the video window 230 using play controls 240. The video can be viewed with or without audio depending upon the selection of a mute button 250. The user 130 may select annotations for this shot by clicking the boxes in events 260, static scenes 270, or key objects 280 lists of boxes. Any significant observations which are not contained in the check boxes can be noted in a keywords text box 300.
Figure 3 is an illustration of the system annotating audio with video.
(Simultaneous reference should also be made to Fig. 1.) The multimedia data has been loaded from the storage medium 100 into working memory I 10. The audio with video tab 370 has been selected. The multimedia video has been segmented using scene change detection into shots. The shot list window 200 displays a portion of the shots in the multimedia. The shot 210 associated with the current audio position is highlighted in the shot list window 200. The audio data is displayed in the window 390. A segment of audio 340 has been delimited for annotation; that is, the limits or bounds of the audio has been fixed for subsequent annotation. The video associated with the audio is shown in 230. As the user 130 uses the play controls 360, the audio data display 390 is updated to display the current audio data and the video window 230 changes to reflect the current video frame. Thus, the user 130 may observe the video and simultaneously hear the audio while making audio annotations. The user 130 preferably uses the buttons 350 to delimit audio segments. Check boxes corresponding to the foreground sounds (320) (the most prominent sounds in the segment) and background sounds (330) (sounds which are present but are secondary to other sounds) may be checked to indicated sounds heard within the audio segment 340. Any significant observations which are not contained in the check boxes can be noted in keywords text box 300.
Figure 4 is an illustration of the system annotating audio without video.
(Simultaneous reference should be made to Fig. 1.) The multimedia data has been loaded from the storage medium 100 into working memory 110. Audio-without-video tab 400 has been selected. The audio data is displayed in the window 390. A
segment of audio 340 has been delimited for annotation. As the user 130 uses the play controls 360, the audio data display 390 is updated to display the current audio data.
Thus, the user 130 may only hear the audio while making audio annotations. The user 130 uses the buttons 350 to delimit audio segments. The check boxes for foreground sounds 320 and background sounds 330 may be checked to indicate sounds heard within the audio segment 340. Any significant observations which are not contained in the check boxes can be noted in the keywords text box 300.
It is to be understood that the present invention, in accordance with at least one presently preferred embodiment, includes an arrangement for supplying multimedia content, an input interface for permitting the selection, for observation, of a mode associated with the multimedia content, and an arrangement for annotating observations of a selected mode. Together, these elements may be implemented on at least one general-purpose computer running suitable software programs. These may also be implemented on at least one Integrated Circuit or part of at least one Integrated l0 Circuit. Thus, it is to be understood that the invention may be implemented in hardware, software, or a combination of both.
If not otherwise stated herein, it is to be assumed that all patents, patent applications, patent publications and other publications (including web-based publications) mentioned and cited herein are hereby fully incorporated by reference 15 herein as if set forth in their entirety herein.
Although illustrative embodiments of the present invention have been described herein with reference to the accompanying drawings, it is to be understood that the invention is not limited to those precise embodiments, and that various other changes and modifications may be affected therein by one skilled in the art without 20 departing from the scope or spirit of the invention.
CHARACTERISTICS IN MULTIMEDIA DOCUMENTS
Field of the Invention The present invention relates to the computer processing of multimedia files.
More specifically, the present invention relates to the manual annotation of multi-modal events, objects, scenes, and audio occurring in multimedia files.
Background of the Invention Multimedia content is becoming more common both on the World Wide Web and local computers. As the corpus of multimedia content increases, the indexing of to features within the content becomes more and more important. Observing both audio and video simultaneously and annotating that observation results in a higher confidence level.
Existing multimedia tools provide capabilities to annotate either audio or video separately, but not as a whole. (An example of a video-only annotation tool is 15 the IBM MPEG7 Annotation Tool, inventors J. Smith et al., available through [http://]www.alphaworks.ibm.com/tech/videoannex. Other conventional arrangements are described in: Park et al, "iMEDIA-CAT: Intelligent Media Content Annotation Tool", Proc. International Conference on Inductive Modeling (ICIM) 2001, South Korea, November, 2001; and Minka et al., "Interactive Learning using a Society of Models," Pattern Recognition, Vol. 30, pp. 565, 1997, TR #349.
It has long been recognized that annotating video or audio features in isolation results in a less confidence of the identification of the features.
In view of the foregoing, a need has been recognized in connection with providing improved systems and methods for observing and annotating multi-modal events, objects, scenes, and audio occurring in multimedia files.
Summary of the Invention In accordance with at least one presently preferred embodiment of the present 1o invention, there are broadly contemplated multimedia annotation systems and methods that permit users to observe solely video, video with audio, solely audio, or audio with video and to annotate what has been observed.
In one embodiment, there is provided a computer system which has one or more multimedia files that are stored in a working memory. The mufti-modal 15 annotation process displays a user selected multimedia file, permits the selection of a mode or modes to observe the file content, annotates the observations; and saves the annotations in a working memory (such as a MPEG-7 XML file).
In summary, one aspect of the invention provides an apparatus for managing multimedia content, the apparatus comprising: an arrangement for supplying multimedia content; an input interface for permitting the selection, for observation, of at least one of the following modes associated with the multimedia content: an audio portion that includes video; and a video portion that includes audio; and an arrangement for annotating observations of a selected mode.
A further aspect of the invention provides a method of managing multimedia content, the method comprising the steps of: supplying multimedia content;
permitting the selection, for observation, of at least one of the following modes associated with the multimedia content: an audio portion that includes video;
and a video portion that includes audio; and annotating observations of a selected mode.
to Furthermore, an additional aspect of the invention provides a program storage device readable by machine, tangibly embodying a program of instructions executable by the machine to perform method steps for managing multimedia content, the method comprising the steps of: supplying multimedia content; permitting the selection, for observation, of at least one of the following modes associated with the multimedia 15 content: an audio portion that includes video; and a video portion that includes audio;
and annotating observations of a selected mode.
For a better understanding of the present invention, together with other and further features and advantages thereof, reference is made to the following description, taken in conjunction with the accompanying drawings, and the scope of 20 the invention will be pointed out in the appended claims.
Brief Descriution of the Drawings Figure 1 is a block diagram depicting a multi-modal annotation system.
Figure 2 is an illustration of a system annotating video scenes, objects, and events.
Figure 3 is an illustration of a system annotating audio with video.
Figure 4 is an illustration of a system annotating audio without video.
Description of the Preferred Embodiments Figure 1 is a block diagram of one preferred embodiment of a multi-modal annotation system in accordance with the present invention. The multimedia content and previous annotations are stored on the storage medium 100. When a user 130 selects a multimedia file via the annotation tool from the storage medium 100, it is loaded into working memory 110 and portions of it displayed in the annotation tool 120. At any time, the user 130 may also request that previously saved annotations associated with the current multi-modal file be loaded from the storage medium into working memory 110. The user 100 views the multimedia data by making requests through the annotation tool 120. The user 130 then annotates his observations and the annotation tool 120 saves these annotations in working memory 110. The user can at anytime request the annotation tool 120 to save the annotation on the storage medium 100.
Figure 2 is an illustration of a system annotating video scenes, objects, and events. (Simultaneous reference should also be made to Fig. l.) The multimedia data has been loaded from the storage medium 100 into working memory 110. A video tab 290 has been- selected. The multimedia video has been segmented using scene 5 changed detection into shots. A shot list window 200 displays a portion of the shots in the multimedia. Here, the user 130 has selected a shot 2I0 which is highlighted in the shot list window 200. A key frame 220, which is a representative shot in the frames of a shot, is preferably displayed. In addition, the frames of the shot maybe viewed in the video window 230 using play controls 240. The video can be viewed with or without audio depending upon the selection of a mute button 250. The user 130 may select annotations for this shot by clicking the boxes in events 260, static scenes 270, or key objects 280 lists of boxes. Any significant observations which are not contained in the check boxes can be noted in a keywords text box 300.
Figure 3 is an illustration of the system annotating audio with video.
(Simultaneous reference should also be made to Fig. 1.) The multimedia data has been loaded from the storage medium 100 into working memory I 10. The audio with video tab 370 has been selected. The multimedia video has been segmented using scene change detection into shots. The shot list window 200 displays a portion of the shots in the multimedia. The shot 210 associated with the current audio position is highlighted in the shot list window 200. The audio data is displayed in the window 390. A segment of audio 340 has been delimited for annotation; that is, the limits or bounds of the audio has been fixed for subsequent annotation. The video associated with the audio is shown in 230. As the user 130 uses the play controls 360, the audio data display 390 is updated to display the current audio data and the video window 230 changes to reflect the current video frame. Thus, the user 130 may observe the video and simultaneously hear the audio while making audio annotations. The user 130 preferably uses the buttons 350 to delimit audio segments. Check boxes corresponding to the foreground sounds (320) (the most prominent sounds in the segment) and background sounds (330) (sounds which are present but are secondary to other sounds) may be checked to indicated sounds heard within the audio segment 340. Any significant observations which are not contained in the check boxes can be noted in keywords text box 300.
Figure 4 is an illustration of the system annotating audio without video.
(Simultaneous reference should be made to Fig. 1.) The multimedia data has been loaded from the storage medium 100 into working memory 110. Audio-without-video tab 400 has been selected. The audio data is displayed in the window 390. A
segment of audio 340 has been delimited for annotation. As the user 130 uses the play controls 360, the audio data display 390 is updated to display the current audio data.
Thus, the user 130 may only hear the audio while making audio annotations. The user 130 uses the buttons 350 to delimit audio segments. The check boxes for foreground sounds 320 and background sounds 330 may be checked to indicate sounds heard within the audio segment 340. Any significant observations which are not contained in the check boxes can be noted in the keywords text box 300.
It is to be understood that the present invention, in accordance with at least one presently preferred embodiment, includes an arrangement for supplying multimedia content, an input interface for permitting the selection, for observation, of a mode associated with the multimedia content, and an arrangement for annotating observations of a selected mode. Together, these elements may be implemented on at least one general-purpose computer running suitable software programs. These may also be implemented on at least one Integrated Circuit or part of at least one Integrated l0 Circuit. Thus, it is to be understood that the invention may be implemented in hardware, software, or a combination of both.
If not otherwise stated herein, it is to be assumed that all patents, patent applications, patent publications and other publications (including web-based publications) mentioned and cited herein are hereby fully incorporated by reference 15 herein as if set forth in their entirety herein.
Although illustrative embodiments of the present invention have been described herein with reference to the accompanying drawings, it is to be understood that the invention is not limited to those precise embodiments, and that various other changes and modifications may be affected therein by one skilled in the art without 20 departing from the scope or spirit of the invention.
Claims (19)
1. An apparatus for managing multimedia content, said apparatus comprising:
an arrangement for supplying multimedia content;
an input interface for permitting the selection, for observation, of at least one of the following modes associated with the multimedia content: an audio portion that includes video; and a video portion that includes audio; and an arrangement for annotating observations of a selected mode.
an arrangement for supplying multimedia content;
an input interface for permitting the selection, for observation, of at least one of the following modes associated with the multimedia content: an audio portion that includes video; and a video portion that includes audio; and an arrangement for annotating observations of a selected mode.
2. The apparatus according to Claim 1, wherein said input interface permits the selection, for observation, of both of the following associated with the multimedia content: an audio portion that includes video; and a video portion that includes audio.
3. The apparatus according to Claim 1, wherein said input interface additionally permits the selection, for observation, of solely a video portion of multimedia content.
4. The apparatus according to Claim 1, wherein said input interface additionally permits the selection, for observation, of solely an audio portion of multimedia content.
5. The apparatus according to Claim 1, wherein said arrangement for supplying multimedia content comprises a working memory which stores multimedia files.
6. The apparatus according to Claim 1, wherein said input interface is adapted to: first permit the selection of a multimedia file and then permit the selection of said at least one of: an audio portion simultaneously with video; and a video portion simultaneously with audio.
7. The apparatus according to Claim 1, further comprising a working memory for saving the annotated observations of a selected mode.
8. The apparatus according to Claim 1, wherein said input interface is adapted to permit the selection, for observation, at least the following mode associated with the multimedia content: a video portion that includes audio.
9. The apparatus according to Claim 8, wherein said input interface comprises:
an arrangement for permitting the selection, for observation, of a video mode of multimedia content; and an arrangement for selectably adding audio to the video mode for observation.
an arrangement for permitting the selection, for observation, of a video mode of multimedia content; and an arrangement for selectably adding audio to the video mode for observation.
10. A method of managing multimedia content, said method comprising the steps of:
supplying multimedia content;
permitting the selection, for observation, of at least one of the following modes associated with the multimedia content: an audio portion that includes video;
and a video portion that includes audio; and annotating observations of a selected mode.
supplying multimedia content;
permitting the selection, for observation, of at least one of the following modes associated with the multimedia content: an audio portion that includes video;
and a video portion that includes audio; and annotating observations of a selected mode.
11. The method according to Claim 10, wherein said step of permitting selection comprises permitting the selection, for observation, of both of the following associated with the multimedia content: an audio portion that includes video;
and a video portion that includes audio.
and a video portion that includes audio.
12. The method according to Claim 10, wherein said step of permitting selection additionally comprises permitting the selection the selection, for observation, of solely a video portion of multimedia content.
13. The method according to Claim 10, wherein step of permitting selection comprises permitting the selection, for observation, of solely an audio portion of multimedia content.
14. The method according to Claim 10, wherein said step of supplying multimedia content comprises providing a working memory which stores multimedia files.
15. The method according to Claim 10, wherein said step of permitting selection comprises: first permitting the selection of a multimedia file and then permitting the selection of said at least one of: an audio portion simultaneously with video; and a video portion simultaneously with audio.
16. The method according to Claim 10, further comprising the step of providing a working memory for saving the annotated observations of a selected mode.
17. The method according to Claim 10, wherein said step of permitting selection comprises permitting the selection, for observation, at least the following mode associated with themultimedia content: a video portion that includes audio.
18. The method according to Claim 17, wherein said step of permitting selection comprises:
permitting the selection, for observation, of a video mode of multimedia content; and thereafter enabling the addition of audio to the video mode for observation.
permitting the selection, for observation, of a video mode of multimedia content; and thereafter enabling the addition of audio to the video mode for observation.
19. A program storage device readable by machine, tangibly embodying a program of instructions executable by the machine to perform method steps for managing multimedia content, said method comprising the steps of:
supplying multimedia content;
permitting the selection, for observation, of at least one of the following modes associated with the multimedia content: an audio portion that includes video;
and a video portion that includes audio; and annotating observations of a selected mode.
supplying multimedia content;
permitting the selection, for observation, of at least one of the following modes associated with the multimedia content: an audio portion that includes video;
and a video portion that includes audio; and annotating observations of a selected mode.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/325,061 US20040123231A1 (en) | 2002-12-20 | 2002-12-20 | System and method for annotating multi-modal characteristics in multimedia documents |
US10/325,061 | 2002-12-20 | ||
PCT/US2003/040726 WO2004059536A2 (en) | 2002-12-20 | 2003-12-19 | System and method for annotating multi-modal characteristics in multimedia documents |
Publications (1)
Publication Number | Publication Date |
---|---|
CA2507248A1 true CA2507248A1 (en) | 2004-07-15 |
Family
ID=32593641
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA002507248A Abandoned CA2507248A1 (en) | 2002-12-20 | 2003-12-19 | System and method for annotating multi-modal characteristics in multimedia documents |
Country Status (9)
Country | Link |
---|---|
US (2) | US20040123231A1 (en) |
EP (1) | EP1581893A2 (en) |
JP (1) | JP4347223B2 (en) |
KR (1) | KR100781623B1 (en) |
CN (1) | CN1726496A (en) |
AU (1) | AU2003297416A1 (en) |
CA (1) | CA2507248A1 (en) |
TW (1) | TW200416600A (en) |
WO (1) | WO2004059536A2 (en) |
Families Citing this family (43)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7082572B2 (en) * | 2002-12-30 | 2006-07-25 | The Board Of Trustees Of The Leland Stanford Junior University | Methods and apparatus for interactive map-based analysis of digital video content |
US20050027664A1 (en) * | 2003-07-31 | 2005-02-03 | Johnson David E. | Interactive machine learning system for automated annotation of information in text |
US7797328B2 (en) * | 2004-12-21 | 2010-09-14 | Thomas Lane Styles | System and method of searching for story-based media |
US20060184553A1 (en) * | 2005-02-15 | 2006-08-17 | Matsushita Electric Industrial Co., Ltd. | Distributed MPEG-7 based surveillance servers for digital surveillance applications |
DE102005022550A1 (en) * | 2005-05-17 | 2006-11-23 | Siemens Ag | Method for postprocessing at least one film shot created during an investigation |
DE102005025903A1 (en) * | 2005-06-06 | 2006-12-28 | Fm Medivid Ag | Device for annotating motion pictures in the medical field |
US8805929B2 (en) | 2005-06-20 | 2014-08-12 | Ricoh Company, Ltd. | Event-driven annotation techniques |
US8015482B2 (en) * | 2005-09-14 | 2011-09-06 | Microsoft Corporation | Dynamic anchoring of annotations to editable content |
US20070204238A1 (en) * | 2006-02-27 | 2007-08-30 | Microsoft Corporation | Smart Video Presentation |
CA2652986A1 (en) * | 2006-05-19 | 2007-11-29 | Sciencemedia Inc. | Interactive learning and assessment platform |
US8301995B2 (en) * | 2006-06-22 | 2012-10-30 | Csr Technology Inc. | Labeling and sorting items of digital data by use of attached annotations |
US7739304B2 (en) * | 2007-02-08 | 2010-06-15 | Yahoo! Inc. | Context-based community-driven suggestions for media annotation |
US7788099B2 (en) * | 2007-04-09 | 2010-08-31 | International Business Machines Corporation | Method and apparatus for query expansion based on multimodal cross-vocabulary mapping |
US8583267B2 (en) | 2007-08-17 | 2013-11-12 | The Invention Science Fund I, Llc | Selective invocation of playback content supplementation |
US8990400B2 (en) * | 2007-08-17 | 2015-03-24 | The Invention Science Fund I, Llc | Facilitating communications among message recipients |
US7733223B2 (en) * | 2007-08-17 | 2010-06-08 | The Invention Science Fund I, Llc | Effectively documenting irregularities in a responsive user's environment |
KR101387521B1 (en) * | 2007-10-30 | 2014-04-23 | 엘지전자 주식회사 | Mobile terminal, method of transmitting information therein, and method of palying back data therein |
US8898558B2 (en) * | 2007-12-07 | 2014-11-25 | Hewlett-Packard Development Company, L.P. | Managing multimodal annotations of an image |
US8140973B2 (en) * | 2008-01-23 | 2012-03-20 | Microsoft Corporation | Annotating and sharing content |
US20090319883A1 (en) * | 2008-06-19 | 2009-12-24 | Microsoft Corporation | Automatic Video Annotation through Search and Mining |
US8839327B2 (en) * | 2008-06-25 | 2014-09-16 | At&T Intellectual Property Ii, Lp | Method and apparatus for presenting media programs |
US9015778B2 (en) | 2008-06-25 | 2015-04-21 | AT&T Intellectual Property I. LP | Apparatus and method for media on demand commentaries |
US9276761B2 (en) * | 2009-03-04 | 2016-03-01 | At&T Intellectual Property I, L.P. | Method and apparatus for group media consumption |
US8275623B2 (en) | 2009-03-06 | 2012-09-25 | At&T Intellectual Property I, L.P. | Method and apparatus for analyzing discussion regarding media programs |
US8380866B2 (en) | 2009-03-20 | 2013-02-19 | Ricoh Company, Ltd. | Techniques for facilitating annotations |
US8266652B2 (en) | 2009-10-15 | 2012-09-11 | At&T Intellectual Property I, L.P. | Apparatus and method for transmitting media content |
US9830605B2 (en) * | 2009-10-30 | 2017-11-28 | At&T Intellectual Property I, L.P. | Apparatus and method for product marketing |
KR20110047768A (en) | 2009-10-30 | 2011-05-09 | 삼성전자주식회사 | Apparatus and method for displaying multimedia contents |
DE102009060687A1 (en) * | 2009-11-04 | 2011-05-05 | Siemens Aktiengesellschaft | Method and device for computer-aided annotation of multimedia data |
US8224756B2 (en) * | 2009-11-05 | 2012-07-17 | At&T Intellectual Property I, L.P. | Apparatus and method for managing a social network |
US8760469B2 (en) | 2009-11-06 | 2014-06-24 | At&T Intellectual Property I, L.P. | Apparatus and method for managing marketing |
US9031379B2 (en) * | 2009-11-10 | 2015-05-12 | At&T Intellectual Property I, L.P. | Apparatus and method for transmitting media content |
US8316303B2 (en) | 2009-11-10 | 2012-11-20 | At&T Intellectual Property I, L.P. | Method and apparatus for presenting media programs |
US10708663B2 (en) | 2009-11-13 | 2020-07-07 | At&T Intellectual Property I, L.P. | Apparatus and method for media on demand commentaries |
US8387088B2 (en) * | 2009-11-13 | 2013-02-26 | At&T Intellectual Property I, Lp | Method and apparatus for presenting media programs |
US8839306B2 (en) | 2009-11-20 | 2014-09-16 | At&T Intellectual Property I, Lp | Method and apparatus for presenting media programs |
US9100550B2 (en) | 2009-11-20 | 2015-08-04 | At&T Intellectual Property I, L.P. | Apparatus and method for managing a social network |
US8373741B2 (en) | 2009-11-20 | 2013-02-12 | At&T Intellectual Property I, Lp | Apparatus and method for collaborative network in an enterprise setting |
US9094726B2 (en) | 2009-12-04 | 2015-07-28 | At&T Intellectual Property I, Lp | Apparatus and method for tagging media content and managing marketing |
US20120159329A1 (en) * | 2010-12-16 | 2012-06-21 | Yahoo! Inc. | System for creating anchors for media content |
US20140164366A1 (en) * | 2012-12-12 | 2014-06-12 | Microsoft Corporation | Flat book to rich book conversion in e-readers |
US10318572B2 (en) * | 2014-02-10 | 2019-06-11 | Microsoft Technology Licensing, Llc | Structured labeling to facilitate concept evolution in machine learning |
US10438499B2 (en) * | 2014-08-01 | 2019-10-08 | International Business Machines Corporation | Identification of comprehension burden in multimedia content |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5148154A (en) * | 1990-12-04 | 1992-09-15 | Sony Corporation Of America | Multi-dimensional user interface |
US5524193A (en) * | 1991-10-15 | 1996-06-04 | And Communications | Interactive multimedia annotation method and apparatus |
US5442744A (en) | 1992-04-03 | 1995-08-15 | Sun Microsystems, Inc. | Methods and apparatus for displaying and editing multimedia information |
US5583980A (en) * | 1993-12-22 | 1996-12-10 | Knowledge Media Inc. | Time-synchronized annotation method |
US5600775A (en) * | 1994-08-26 | 1997-02-04 | Emotion, Inc. | Method and apparatus for annotating full motion video and other indexed data structures |
US5838313A (en) * | 1995-11-20 | 1998-11-17 | Siemens Corporate Research, Inc. | Multimedia-based reporting system with recording and playback of dynamic annotation |
US6154600A (en) * | 1996-08-06 | 2000-11-28 | Applied Magic, Inc. | Media editor for non-linear editing system |
JPH10177511A (en) | 1996-12-18 | 1998-06-30 | Fuji Xerox Co Ltd | File system |
US6006241A (en) * | 1997-03-14 | 1999-12-21 | Microsoft Corporation | Production of a video stream with synchronized annotations over a computer network |
US7111009B1 (en) * | 1997-03-14 | 2006-09-19 | Microsoft Corporation | Interactive playlist generation using annotations |
US6449653B2 (en) * | 1997-03-25 | 2002-09-10 | Microsoft Corporation | Interleaved multiple multimedia stream for synchronized transmission over a computer network |
US6546405B2 (en) * | 1997-10-23 | 2003-04-08 | Microsoft Corporation | Annotating temporally-dimensioned multimedia content |
US6173287B1 (en) * | 1998-03-11 | 2001-01-09 | Digital Equipment Corporation | Technique for ranking multimedia annotations of interest |
US6956593B1 (en) * | 1998-09-15 | 2005-10-18 | Microsoft Corporation | User interface for creating, viewing and temporally positioning annotations for media content |
US7051275B2 (en) * | 1998-09-15 | 2006-05-23 | Microsoft Corporation | Annotations for multiple versions of media content |
US7647555B1 (en) | 2000-04-13 | 2010-01-12 | Fuji Xerox Co., Ltd. | System and method for video access from notes or summaries |
US8028314B1 (en) * | 2000-05-26 | 2011-09-27 | Sharp Laboratories Of America, Inc. | Audiovisual information management system |
-
2002
- 2002-12-20 US US10/325,061 patent/US20040123231A1/en not_active Abandoned
-
2003
- 2003-10-27 TW TW092129818A patent/TW200416600A/en unknown
- 2003-12-19 CN CNA2003801064602A patent/CN1726496A/en active Pending
- 2003-12-19 WO PCT/US2003/040726 patent/WO2004059536A2/en active Application Filing
- 2003-12-19 KR KR1020057009874A patent/KR100781623B1/en not_active IP Right Cessation
- 2003-12-19 US US10/539,890 patent/US7793212B2/en not_active Expired - Lifetime
- 2003-12-19 JP JP2004563862A patent/JP4347223B2/en not_active Expired - Lifetime
- 2003-12-19 AU AU2003297416A patent/AU2003297416A1/en not_active Abandoned
- 2003-12-19 CA CA002507248A patent/CA2507248A1/en not_active Abandoned
- 2003-12-19 EP EP03814249A patent/EP1581893A2/en not_active Withdrawn
Also Published As
Publication number | Publication date |
---|---|
US20040123231A1 (en) | 2004-06-24 |
EP1581893A2 (en) | 2005-10-05 |
AU2003297416A1 (en) | 2004-07-22 |
US7793212B2 (en) | 2010-09-07 |
JP4347223B2 (en) | 2009-10-21 |
KR100781623B1 (en) | 2007-12-07 |
CN1726496A (en) | 2006-01-25 |
TW200416600A (en) | 2004-09-01 |
US20060218481A1 (en) | 2006-09-28 |
KR20050085251A (en) | 2005-08-29 |
JP2006512007A (en) | 2006-04-06 |
WO2004059536A2 (en) | 2004-07-15 |
WO2004059536A3 (en) | 2004-12-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7793212B2 (en) | System and method for annotating multi-modal characteristics in multimedia documents | |
US10031649B2 (en) | Automated content detection, analysis, visual synthesis and repurposing | |
US7506262B2 (en) | User interface for creating viewing and temporally positioning annotations for media content | |
US6546405B2 (en) | Annotating temporally-dimensioned multimedia content | |
US7725829B1 (en) | Media authoring and presentation | |
US7181757B1 (en) | Video summary description scheme and method and system of video summary description data generation for efficient overview and browsing | |
US6636238B1 (en) | System and method for linking an audio stream with accompanying text material | |
US7739255B2 (en) | System for and method of visual representation and review of media files | |
US20050229118A1 (en) | Systems and methods for browsing multimedia content on small mobile devices | |
US8930308B1 (en) | Methods and systems of associating metadata with media | |
WO2001027876A1 (en) | Video summary description scheme and method and system of video summary description data generation for efficient overview and browsing | |
KR20030007727A (en) | Automatic video retriever genie | |
US20040181545A1 (en) | Generating and rendering annotated video files | |
Messer et al. | SeeNSearch: A context directed search facilitator for home entertainment devices | |
EP1683044A1 (en) | Screen-wise presentation of search results | |
US20020059303A1 (en) | Multimedia data management system | |
Tjondronegoro et al. | Content-based indexing and retrieval using MPEG-7 and X-query in video data management systems | |
EP1405212B1 (en) | Method and system for indexing and searching timed media information based upon relevance intervals | |
KR100369222B1 (en) | Digital Video Searching And Authoring Tool | |
Hirzalla et al. | A multimedia query specification language | |
Chang et al. | The characteristics of digital video and considerations of designing video databases | |
Tseng et al. | Video personalization and summarization system | |
Mu et al. | Enriched video semantic metadata: Authorization, integration, and presentation | |
Tseng et al. | Personalized video summary using visual semantic annotations and automatic speech transcriptions | |
US20140250055A1 (en) | Systems and Methods for Associating Metadata With Media Using Metadata Placeholders |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
EEER | Examination request | ||
FZDE | Discontinued | ||
FZDE | Discontinued |
Effective date: 20101220 |