WO1996008107A1 - System and method for recording and playing back multimedia events - Google Patents
System and method for recording and playing back multimedia events Download PDFInfo
- Publication number
- WO1996008107A1 WO1996008107A1 PCT/US1995/009959 US9509959W WO9608107A1 WO 1996008107 A1 WO1996008107 A1 WO 1996008107A1 US 9509959 W US9509959 W US 9509959W WO 9608107 A1 WO9608107 A1 WO 9608107A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- tracks
- track
- event
- user
- original
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 40
- 230000002452 interceptive effect Effects 0.000 claims abstract description 17
- 230000001360 synchronised effect Effects 0.000 claims abstract description 8
- 238000013507 mapping Methods 0.000 claims description 36
- 238000003860 storage Methods 0.000 claims description 24
- 238000009826 distribution Methods 0.000 claims description 14
- 238000012545 processing Methods 0.000 claims description 8
- 230000001419 dependent effect Effects 0.000 claims description 3
- 230000001131 transforming effect Effects 0.000 claims 1
- 230000000007 visual effect Effects 0.000 abstract description 7
- 238000012549 training Methods 0.000 description 47
- 230000011664 signaling Effects 0.000 description 10
- 230000006835 compression Effects 0.000 description 9
- 238000007906 compression Methods 0.000 description 9
- 238000013459 approach Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 5
- 238000012546 transfer Methods 0.000 description 5
- 230000008901 benefit Effects 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 238000012552 review Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 206010017577 Gait disturbance Diseases 0.000 description 2
- 230000006837 decompression Effects 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000007781 pre-processing Methods 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 238000005520 cutting process Methods 0.000 description 1
- 230000003292 diminished effect Effects 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 238000011144 upstream manufacturing Methods 0.000 description 1
- 230000001755 vocal effect Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/16—Analogue secrecy systems; Analogue subscription systems
- H04N7/173—Analogue secrecy systems; Analogue subscription systems with two-way working, e.g. subscriber sending a programme selection signal
- H04N7/17309—Transmission or handling of upstream communications
- H04N7/17318—Direct or substantially direct transmission and handling of requests
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/34—Indicating arrangements
Definitions
- This invention relates to a system and method for recording and playing back multimedia events.
- a shared communications experience is an event at which groups of communicants engage in the exchange of information and ideas. Such events include business and technical presentations, training sessions, lectures, and meetings. A shared communications experience might also include a teleconferencing session between participants located remotely from one another. Where the event entails the exchange of different forms (e.g., video, text, and audio), the event is commonly characterized as a multimedia event. The exchange of information may be assisted by both non-electronic devices (e.g., view graphs, white boards, and notepads) and electronic devices (e.g., video, computers, and electronic white boards).
- non-electronic devices e.g., view graphs, white boards, and notepads
- electronic devices e.g., video, computers, and electronic white boards.
- Fig. 1 depicts an example of a multimedia event and electronic and non ⁇ electronic devices that might be employed to facilitate communication.
- a lecturer 110 who presents information to trainees 11 1 - 114 seated in the training room 100.
- the lecturer utilizes a view graph machine 120, which projects an image on to a projection screen 125.
- the lecturer also utilizes an electronic white board 127 for drawing handwritten notes and diagrams.
- Shared communications experiences are often recorded to allow persons unable to attend the event to benefit from the exchange of information and ideas.
- the recording also permits attendees to review the event and clarify details that were missed during the live event.
- a technique for recording such events may include using written notes, such as those that might be taken by one of the trainees 11 1-1 14 at the training session.
- Written notes result in a subjective interpretation of the event that may lack important visual information, such as the speaker's body language and trainees' non-verbal responses to lecture materials as well as audible information such as the speaker's tone of voice.
- Another drawback is that the written notes are also a subjective interpretation of the event by the scrivener.
- Another technique involves recording the audio of the event on audio cassettes. Pure audio recordings can be a more accurate and complete record of the experience than written notes, but are difficult to browse or search. Therefore, if a user of the recording were only interested in hearing a portion of the event where the lecturer 110 answered a question asked by trainee 111 , the listener must randomly advance and reverse in the audio recording until stumbling upon the desired segment. As with written notes, visual cues from the event are lost.
- a commonly preferred technique for recording events is using a video camcorder operated by human operator.
- a video recording provides the most accurate representation of the event, but often does not provide adequate coverage of the whole experience because of the limited field of view of conventional video camcorders.
- the view of the event captured in the video recording has been subjectively selected by the human operator, and this view may not coincide with the view of the event desired by a user of the video recording.
- the camera operator may focus on the view graphs projected onto the projection screen 125 while the lecturer 1 10 is emphasizing a point and miss the gestures of the lecturer 110 used to illustrate the point. Consequently, users of the recording will never see these gestures.
- both the view graphs and the lecturer should be captured such that users of the recording have the option to choose which view should be presented.
- users of the video recording are limited to the view subjectively selected by the camera operator.
- the video recording can be generated using a more sophisticated technique employing multiple cameras and other equipment as described in U.S. Patent 5,187,571, which cameras and equipment facilitate the capturing and contiguous merging of multiple views of an event to form an aggregate field view.
- cameras 130, 140, and 150 could be used to capture Regions A, B, and C, respectively, of the training session as shown in Fig. 1.
- VHS Video Home System
- SMPTE Society of Motion Picture and Television Engineers
- VCRs video cassette recorders
- locating a point of interest requires that the user randomly advance and reverse in the video recording until stumbling upon the point of interest. Furthermore, once the point of interest has been located, the viewer has limited flexibility, if any, to control the resolution of an aspect of the recording or to disable an unwanted aspect of the recording. For example, under conventional approaches, a viewer would not have the flexibility of increasing the resolution of the audio portion of a recording to increase the audible clarity of a question from trainee 111 to the lecturer 110, or totally disabling the view from camera 140 of Region B.
- Another consideration when selecting a mechanism for recording an event is the method of delivering the recorded event to the user.
- One delivery alternative is to transmit the recording from a service provider to a user over a network in real time.
- a limitation of conventional video recordings is that such recordings are bandwidth intensive and often exceed the bandwidth capabilities of the network and the user's facilities, thereby precluding this alternative as a method of delivering recordings to users.
- our objective is a system and method for recording and playing back an event such that a user of the recording can control the aspects of the event being presented.
- a further objective is to facilitate browsing and searching through a recording to locate a point of interest in the recorded event.
- Another objective is to record events such that the aspects of the event delivered to the user can be adapted to meet constraints of the network, such as bandwidth limitations, and the user's equipment.
- aspects of our invention are a system and method for recording and playing back multimedia events, such as business and technical presentations, training sessions, lectures, and meetings.
- Our inventive system includes recording sources, a preprocessor, event mass storage, a delivery processor, a distribution network, and user control units.
- the recording sources are devices, such as audio recorders and video camcorders, each of which captures an aspect of the multimedia event, which aspect is referred to herein as a multimedia track.
- the recording source transmits the track to the preprocessor where each track is received, coded into digital blocks, if necessary, and time stamped. Digitized blocks of tracks that are simultaneously received at the preprocessor are assigned the same relative time and therefore synchronized in time.
- some tracks of the multimedia event are also compressed to reduce the bandwidth required for transmitting a multimedia event to a user.
- the digitization and compression techniques employed for each track are dictated by entries in a track mapping table, which entries are based on the type of information carried by the track.
- tracks carrying predominately textual information may be digitized and compressed differently than tracks carrying full motion video.
- the tracks of the multimedia event may be preprocessed to generate other representations of the multimedia event, known as synthetic tracks, which further assist users in searching through the multimedia event.
- Our inventive system also includes an event mass storage where recordings of multiple multimedia events are stored for delivery to requesting users.
- the tracks of a multimedia event which tracks each represents different aspects of the event, are maintained as separate and distinct tracks.
- Another component of our inventive system is the delivery processor which services requests for multimedia events from multiple users, retrieves requested events from the event mass storage, and delivers events to requesting users according to the users' interactive commands.
- the delivery processor is connected via the distribution network to multiple user control units, where requests and interactive commands from the user are received and multimedia events from the delivery processor are received for presentation the user.
- the delivery processor may also identify the aspects of the event to be delivered to a user control unit based upon the bandwidth limitations of the distribution network and the capabilities of the user equipment.
- the delivery processor also generates stacked representations of the tracks of the multimedia event, which representations provide visual cues to assist the user in conducting intelligent searches through the multimedia event and in locating points of interest.
- the user control unit is connected to user output devices, which convert the tracks of the multimedia event received from the delivery processor into a form suitable for the user.
- the user control unit is also connected to user input devices, where users input requests and interactive commands.
- Our inventive method for recording and playing back multimedia events entails capturing various aspects of a multimedia event, such as a training session, using multiple recording sources.
- the tracks generated by the recording sources are received at the preprocessor, where the tracks are coded into digitized blocks, time stamped, compressed, preprocessed, and transferred to and stored at the event mass storage.
- the delivery processor responds by retrieving the requested multimedia event from mass storage and delivering the tracks of the event, which tracks represent aspects of the event, to the user control unit. If only certain aspects of the multimedia event were requested by the user or if the distribution network or user output devices is not capable of supporting all aspects of the event, the delivery processor will cause those aspects that are requested and capable of being received to be delivered.
- the tracks are then presented by the user control unit at the user output devices.
- the user can search through any of the individual tracks of the multimedia events to locate a point of interest in the multimedia event and can control the presentation of the tracks at the user output devices using traditional VCR-type commands, as well as other play back features.
- FIG. 1 depicts an illustrative example of a training session.
- Fig. 2 depicts an illustrative embodiment of a system for recording and playing back multimedia events in accordance with our invention.
- Fig. 3A depicts a preprocessor unit in accordance with an aspect of our invention.
- Fig. 3B depicts a track mapping table, which provides instructions for digitizing, compressing, and preprocessing tracks of the multimedia event in accordance with an aspect of our invention.
- Fig. 3C depicts a synthetic track generator of the preprocessor in accordance with an aspect of our invention.
- Fig. 4 illustrates the arrangement of multimedia events in mass storage in accordance with an aspect of our invention.
- Fig. 5 shows a delivery processor for playing back multimedia events in accordance with an aspect of our invention.
- Fig. 6 depicts a user control unit in accordance with an aspect of our invention.
- Fig. 7 A depicts an illustrative example of a training session, wherein recording sources capture multimedia tracks of various aspects of the session, in accordance with our invention.
- Fig. 7B depicts the user's view of the training session, as displayed on the user's monitor, in this embodiment of our invention.
- Fig. 7C depicts the user's view of a stacked representation of a track of the training session, as displayed on the user's monitor, in this embodiment of our invention.
- Fig. 7D depicts the user's view of a high resolution version of a track of the training session, as displayed on the user's monitor, in this embodiment of our invention.
- FIGs. 8 A, 8B, and 8C depict an illustrative flow diagram of our recording and play back method, in accordance with our invention.
- Multimedia is a combination of different types of information (e.g., still images, text, graphics, audio, and motion video) integrated into a single presentation.
- Multimedia events include business and technical presentations, training sessions, lectures, and meetings, comprising such integrated combinations of different forms of information.
- Our invention is a system and method for recording and playing back multimedia events such that a user of the recorded multimedia event can review a desired aspect of the event and maximally benefit from the exchange of information that took place at the live event.
- Our system also permits a user to intelligently search through the recorded event and control the play back of the recorded event in a manner conducive to the information needs of the user and within the limitations of the user's facilities.
- FIG. 2 A specific illustrative embodiment of our inventive system 200 is depicted in Fig. 2 and comprises recording sources 201-1 through 201-N, a preprocessor 300, an event mass storage 390, a delivery processor 400, a distribution network 450, and user control units 500-1 through 500-M.
- the recording sources 201 are devices that capture and transmit an aspect of a multimedia event, which aspect is referred to herein as a multimedia track. Such devices include audio recorders and camcorders. Recording sources may also include a device located internal to a presentation device, such as an electronic white board, which records texts and drawings written thereon.
- recording sources may include a computerized view graph machine, which projects view graphs stored in computer memory onto a projection screening, and technologies such as electronic white boards that output a digital stroke list, resulting in higher visual quality and lower processing and storage requirements than a video recording of a conventional white board.
- each of the recording sources 201-1 through 201-N transmits the captured track to the preprocessor 300, which synchronizes the individual tracks in time and processes the tracks for storage at the event mass storage and review by users at a later time.
- the preprocessor 300 as shown in Fig.
- Input interface 310 includes source receivers 301-1 through 301 -N which receive tracks from recording sources 201-1 through 201-N, respectively.
- the track mapping table held at store 335 includes instructions for digitizing, compressing, decompressing, and generating synthetic tracks for each track of a multimedia event. This table describes specific instructions relating to the recording and play back of a training session, discussed infra.
- the preprocessor 300 also includes a digitizer 320, which codes tracks into digitized blocks.
- the technique employed for each track is dictated by the recording source generating the track as indicated at the track mapping table store 335. Some tracks may have been outputted from the recording sources in digitized block form. Thus, the signals are passed through digitizer 320 unchanged.
- the same relative time is assigned to each digitized block of each track simultaneously received at the time stamp processor 322 such that the tracks generated from the recording sources are synchronized in time.
- the time stamp processor 322 of the synchronization processor 300 retrieves the relative time from local clock 325 for synchronizing the tracks.
- the time stamp processor 322 appends the relative time information to each digitized block in the block header.
- Another component of the preprocessor 300 is the compressor 330, which receives tracks from the time stamp processor and compresses each individual track according to a technique indicated in the track mapping table store 335. As indicated in the track mapping table shown in Fig.
- the compression technique is dependent upon the type of information (e.g., audio, text, and video) being captured by the recording source.
- the bandwidth required for transmitting a multimedia event to a user is reduced.
- a track containing information, that remains essentially unchanged over a period of time e.g., video tracks of view graphs and slides
- the track mapping table may indicate that a track should not be compressed, in which case, the track is passed through the compressor 330 unchanged.
- the preprocessor 300 also includes a synthetic track generator 350, which processes tracks received from compressor 330 to generate other tracks, which are referred to herein as synthetic tracks, as distinguished from original tracks, and which are digitized and time stamped representations of information transmitted from the recording sources.
- a synthetic track comprises information contained in one or more tracks of a multimedia event which information has been converted into a form that may be more useful to users in searching and browsing through a multimedia event.
- the synthetic track generator 350 includes a character recognition device 360, a speaker identification device 370, a resolution device 380, and a track distributor 385.
- the character recognition device 360 extracts strings of characters identified in a track of a multimedia event and generates a synthetic track, which can be utilized by the user to locate a point of interest in the multimedia event based on strings of characters.
- a track of a multimedia event which represents a video recording of view graphs shown on a projection screen could be processed at the character recognition device 360 to generate a synthetic track of strings of characters identified in the text of the view graphs.
- the character recognition device 350 could extract titles of view graphs from an original track and generate a synthetic track comprising an index of view graph titles corresponding to the original track.
- the synthetic track generator 350 also includes a speaker identification device 370 which processes tracks of a multimedia event containing voices of different speakers and generates synthetic tracks which represent changes in speakers' voice patterns during the multimedia event.
- a speaker identification device 370 which processes tracks of a multimedia event containing voices of different speakers and generates synthetic tracks which represent changes in speakers' voice patterns during the multimedia event.
- the synthetic track generator 350 may also include a resolution device 380 which alters the resolution of an original track to generate a synthetic track comprising a representation of the original track having a higher or lower resolution.
- Other processing devices may be implemented at the synthetic track generator for processing original tracks to generate synthetic tracks. Such other devices could generate synthetic tracks which represent an original track at different frame rates or which provide speaker face recognition information derived from an original track.
- the blocks of the synthetic tracks also contain relative time information, which is carried over from the blocks of the underlying original track during processing.
- the synthetic track generator 350 transfers the original and synthetic tracks of the multimedia event to the output interface 340, which transfers the tracks to the event mass storage 390 where multiple multimedia events are stored for delivery to requesting users.
- a copy of the track mapping table corresponding to the multimedia event is also stored at the event mass storage 390.
- Fig. 4 depicts the layout of multimedia events in storage 390 and illustrates that each track of a multimedia event is separately maintained.
- Another component of our inventive system is a delivery processor 400, which, on receiving a request from a user, delivers the requested multimedia event to the user according to interactive commands issued by the user.
- An illustrative embodiment of our delivery processor 400 is shown in Fig.
- the delivery processor further includes an event store 404, where multimedia events are stored once retrieved from mass storage 390, and a user signalling store 406 for holding interactive signalling received from the user control units 500.
- a delivery controller 403 which causes the event, including the tracks and the track mapping table, to be retrieved from mass storage 390 into the event store 404 via interface 401.
- the delivery controller 403 also controls the tracks of the multimedia event transferred from event store 404 to event processor 405 for delivery to the user.
- the event processor 405 is also controlled by the delivery controller 403.
- the event processor 405 processes, in real time, the individual tracks of the multimedia event, in response to the user's interactive command for a stacked representation of a video track to facilitate rapid and intelligent searching through the track.
- the output of the event processor 405 is delivered to the user control unit of the requesting user via interface 402.
- the control of the event store 404 and the event processor 405 by the delivery controller 403 may be based on the user's interactive commands and also upon other information, such as the capabilities of the user's output devices and the bandwidth constraints of the distribution network 450 interconnecting the delivery processor 400 and the user control unit. Such capabilities and constraints may be stored in a user profile that is accessible by the delivery controller for each user serviced by the delivery controller or may be sent to the delivery controller by the user at the time a request for a multimedia event is made.
- a key function of the event processor 405 is to generate a stacked representation of an original or synthetic track, in real time, to assist the user in searching through a track of a multimedia event to locate a point of interest.
- a stacked representation is constructed by stacking individual frames of a track of a multimedia event on top of one another, with each frame slightly offset from the previous (see Fig. 7D, discussed infra). The patterns along the top and side of the streamer may provide visual cues.
- the frames of the stacked representation also contain relative time information. The relative time associated with each is determined by the earliest relative time associated with any block of the track comprising the frame.
- the delivery controller 403 On receiving an interactive command from a user requesting a stacked representation of a track, the delivery controller 403 causes the track to be transmitted from event store 404 to the event processor 405, and the event processor to generate the stacked representation.
- the stacked representation could appear at the user's monitor as an overlay of the track from which the stacked representation is generated, or as a separate window.
- the user can manipulate a pointer along the stacked representation by interacting with the user input devices connected to the user control unit. By moving the pointer along the representation, individual frames of a track can be displayed allowing the user to rapidly browse from frame to frame.
- the stacked representation of a track does not have to be constructed from full video frames.
- the event processor 405 could crop the frames of the track such that changes in the information contained in the track can be readily observable at the top and side edges of the frames.
- the frames of the track could be cropped by cutting off the upper part of each frame just above the top of the heads of trainees seated in the room, such that, if a trainee stood up or raised her hand to ask a question, this change would be easy to spot by looking at the pattern in the top of the stacked representation.
- Another function of the event processor 405 is to generate signalling messages to be presented to users in overlay fashion to acknowledge the users' requests or commands to provide status information, or to present stacked representations. Such messages are generated by the event processor on instructions by the delivery controller and are transmitted to the user control unit 500.
- the delivery processor 400 and the user control units 500- 1 through 500-M are interconnected by a distribution network 450, which comprises transmission equipment and transmission links of fiber optic, coaxial, or copper cable or other suitable transmission medium.
- the distribution network could be the distribution portion of a public switched telephony network (PSTN), which employs Integrated Services Digital Network (ISDN) technology operating at a basic rate of 128 Kb/s or asymmetrical digital subscriber line (ADSL) technology, which provides transport at approximately 1.5 Mb/s downstream and a few kilobits/second upstream overlaid on a telephony channel.
- PSTN public switched telephony network
- ISDN Integrated Services Digital Network
- ADSL digital subscriber line
- the distribution network 450 could also include a local area network.
- FIG. 6 An illustrative embodiment of a user control unit 500 is shown in Fig. 6.
- the user control unit is connected to user output devices such as a computer monitor or a display screen, on which tracks of the multimedia event containing textual and video information is displayed, or an audio system, at which tracks containing audio information are presented.
- the user control unit 500 is also connected to user input devices, at which user requests for multimedia events and interactive commands are inputted.
- the user input devices could include a computer keyboard, a computer mouse, or a combination thereof.
- each user control unit 500 is connected to the delivery processor 400 via the distribution network 450. As shown in Fig.
- the user control unit is comprised of a user network interface (UNI) 510, a track mapping table store 515, a user control interface (UCI) 520, a demultiplexer 530, a decompressor 535, buffer store 537, a presentation controller 550, a graphic overlay processor 540, and a local clock 555.
- the user input devices are connected to the UCI 520, and on receiving a request for an event or an interactive command being input by the user, the UCI 520 generates signalling to the UNI 510, which forwards the user's requests and commands to the delivery processor 400 to be serviced.
- the user can select one or more of the original or synthetic tracks of a multimedia event for presentation at the user output devices and control the presentation of the tracks to the user output devices using traditional VCR-type commands and other play back features. Also, the user can search through individual tracks of the multimedia events to locate a point of interest in the multimedia event and request stacked representations of different tracks.
- the UNI 510 receives signalling from the UCI 520 representing the user's requests and interactive commands and forwards this signalling to the delivery processor 400.
- the UNI 510 also receives the track mapping table and the tracks of the multimedia event from the delivery processor 400 to be presented at the user output devices.
- the track mapping table is transferred by the UNI into store 515.
- the UNI Upon receiving the tracks of the multimedia event, the UNI passes the tracks to the demultiplexer 530, at which the tracks are separated and each track is outputted from a distinct channel of the demultiplexer 530 and transferred to the decompressor 535.
- each track is decompressed according to information contained in the track mapping table, which information is retrieved by the decompressor from the store 515. As described above, this table assigns a compression and decompression technique for tracks of the multimedia event. In the case where the track is not to be decompressed, the track passes through the compressor unchanged.
- Tracks outputted from the decompressor 535 are transferred to a buffer store 537, where the tracks are held for presentation to the user.
- Tracks of a multimedia event to be presented at the user output devices in overlay fashion are received from decompressor 535 at the graphics overlay processor 540.
- the presentation controller 550 retrieves the blocks of each track from the buffer store 537 and the graphics overlay processor 540, and presents blocks of each track to the user output devices when the relative time associated with each block matches the the relative time indicated at the local clock 555, thereby permitting the blocks of each track to be presented in a synchronized manner.
- the presentation controller also resets the relative time at the local clock 555 according to adjustments made by the user in the current position in the multimedia event.
- the user control unit 500 may include a graphics overlay processor 540 which receives signalling information from the delivery processor via the UNI 510. This signalling information controls the overlay of text and graphics on to the tracks of the multimedia events, which text and graphics are presented to the user output devices by the presentation controller 550.
- a graphics overlay processor 540 which receives signalling information from the delivery processor via the UNI 510. This signalling information controls the overlay of text and graphics on to the tracks of the multimedia events, which text and graphics are presented to the user output devices by the presentation controller 550.
- Fig. 7a depicts a training room, wherein a training session is being conducted by lecturer 1 10 and attended by trainees 1 11-1 14.
- Fig. 7A depicts a training session being captured using the recording sources of our inventive system
- Figs. 8A, 8B, and 8C show an illustrative flow diagram of our recording and play back method.
- Recording sources 127, 601, 602, and 603 comprise the recording sources 201 of Fig.
- the recording sources include a first camera 601 which may be mounted in the room ceiling and pointed at the lecturer to capture in full motion video a view of the lecturer 110 during his presentation at the training session; a second camera 602 which may also be mounted in the ceiling and is pointed similarly to capture a view of trainees 1 1 1 through 114, and third ceiling mounted camera 603, which captures a view of the projection screen 125 where view graphs are presented to the trainees.
- Other recording sources include an audio capturing system 610 comprising a lavaliere microphone 604, which is worn by the lecturer and captures the audio of the lecturer's presentation, and a ceiling microphone 605, which captures audio from within the training room, such as questions from the trainees.
- white board 127 is a recording source which captures texts and drawings written thereon.
- the audio capturing system 610 also includes a mixer (not shown) which combines the audio from the lavaliere and ceiling microphones, and the combined audio is treated as one track of the multimedia event.
- a mixer not shown
- the track mapping table held at store 335 for the training session is shown in Fig. 3B. This table includes the digitization, compression/decompression, and preprocessing instructions for each track of the training session.
- Each of the recording sources transmits its captured track to the input interface 310 of the preprocessor 300 where each track is received by a source receiver.
- the track captured by the audio capturing system 610 is received at source receiver 301-1.
- Video tracks captured by the first, second, and third cameras 601 through 603, are each received at a source receiver and the track from the electronic white board is also received at source receiver.
- the tracks received at the source receivers are transmitted to the digitizer 320, where each track is transformed into digitized blocks according to instructions contained in the track mapping table 335 (Step 810).
- the track received from the white board is already in digitized block form, and therefore, this track requires no digitization as indicated in the track mapping table.
- the digitized blocks are then transferred to the time stamp processor 322 where each block of each track is time stamped with a relative time retrieved from local clock 325 (Step 815). Specifically, each digitized block of each track received simultaneously at the frame synchronization processor 320 are time stamped with the same relative time, which is retrieved from local clock 325.
- the time stamped, digitized blocks representing a track received at each source receiver comprise original tracks 1 through 5 of the multimedia event.
- each track is compressed according to the technique specified in the track mapping table held at store 335, as shown in Fig. 3B (Step 820).
- track 1 comprising the combined audio detected from the audio capturing system
- ADPCM Adaptive Differential Pulse Coded Modulation
- Original tracks 2 and 3 comprising video of the lecturer and trainees, respectively, are compressed using the Motion Picture Expert Group (MPEG) technique with a 100 to 1 compression ratio.
- Original track 4 comprising video of the view graphs are compressed using the Motion Joint Photographic Expert Group (M-JPEG) technique with a 5 to 1 compression ratio.
- M-JPEG Motion Joint Photographic Expert Group
- track 5 comprising the contents of the white board remains uncompressed.
- the original tracks are transferred to the synthetic track generator 350 where the original tracks are preprocessed according to instructions contained in the track mapping table to generate synthetic tracks of the multimedia event (Step 825).
- track 1 is preprocessed at the speaker identification device 370 to generate synthetic track 6 which includes a bar graph representing changes in speaker voice patterns during the training session.
- Original tracks 2 and 3 are preprocessed at the resolution device 380 to generate synthetic tracks 7 and 8, which each represent a high resolution version of the video information of tracks 2 and 3 respectively.
- Original track 4 is preprocessed at the character recognition device 360 to generate synthetic track 9 comprising an index of titles of the view graphs captured in original track 4.
- original track 5 comprising the contents of the white board is also preprocessed at the character recognition device 360 to generate synthetic track 10, comprising strings of characters from original track 5.
- the track mapping table and the original and synthetic tracks of the training session are transferred by the synthetic track generator 350 to the event mass storage 390 (Step 830 of Fig. 8B).
- a user 501-1 served from user control unit 500-1 wishes to review the training session and that the user has user input devices and output devices integrated in a computer system (not shown).
- a mouse and a keyboard of the computer system serve as user input devices and the monitor and audio presentation system serve as user output devices.
- the user places a request for the training session by interacting with the keyboard, and the user's request is transferred by the user control unit 500-1 to the delivery processor 400, via the distribution network 450.
- the delivery controller 403 responds by retrieving the training session, comprising tracks 1 through 10 and the track mapping store, from the event mass storage 390 and causes the session to be stored in event store 404 (Step 835). Based upon additional information contained in the user's request or in the user's profile at the delivery processor 400, the delivery controller 403 determines that all original tracks 1 through 5 should be delivered to the user and causes these tracks to be transferred to the event processor 405 for delivery to the user (Step 840). The delivery controller also causes the track mapping table to be transferred to the event processor 405, which forwards the table to the user control unit 500-1 (Step 840). The table is received at the UNI 510 and held at store 515.
- the delivery controller 403 instructs the event processor 405 to send to the user control unit 500-1 a "READY" signal, indicating that the delivery processor 400 is ready to deliver the requested multimedia event.
- the UNI 510 of the user control unit, Fig. 6 transfers the signal to the graphics overlay processor 540 which causes a "READY” message to be displayed on the user's monitor.
- the delivery controller 403 commences the delivery of the training session by instructing the event processor 405 to deliver original tracks 1 through 5 to the user control unit 500-1 (Step 845).
- the original tracks 1 through 5 are demultiplexed into separate channels (Step 850) and decompressed as instructed in the track mapping table held at store 515 (Step 855 of Fig. 8C).
- the decompressed tracks are then transferred to the presentation controller 550, which controls the presentation of the tracks to the user's computer system.
- the presentation controller 550 transmits blocks of each track to the computer in a synchronized manner by utilizing relative time information contained in each block of each track (Step 860).
- the presentation controller 550 of the user control unit 500- 1 thus controls the presentation of blocks of each individual tracks based upon the relative time at the local clock 555. Therefore, all tracks are presented in a synchronized manner since the blocks of each track representing concurrent aspects of the training session are time stamped with the same relative time.
- Fig. 7B depicts the user's view of the training session as displayed on the user's monitor. As shown, each track of the training session appears in a separate window of the monitor along with interactive commands available to the user.
- Window A contains track 1. Since track 1 contains audio information only, Window A displays a message to the user in graphical overlay fashion indicating that the track contains non- visual information.
- Window B displays track 2 comprising video of the lecturer 1 10
- Window C displays track 3 comprising video of the trainees 11 1 - 1 14
- Window D displays track 4 comprising video of view graphs on the projection screen 125
- Window E displays track 5 comprising information written on the white board 127.
- the user at control unit 500-1 desires to locate a specific point of interest in the training session.
- the user desires to see the facial expression of trainee 114 on hearing the lecturer's response to a question from trainee 113.
- the lecturer 1 10 has limited options in searching through the training session to locate this point of interest. These options include using VCR-type commands to advance or reverse through the session captured by a conventional system. However, such conventional options do not permit a user to rapidly and intelligently search for a particular point of interest based upon details known to the user.
- the user has multiple options for rapidly and intelligently searching through the training session to locate the point of interest described above, which options include searching through any or all original tracks of the training session, synthetic tracks generated from original tracks of the training session, and stacked representations of original or synthetic tracks.
- the option selected by the user is dependent upon the knowledge of the user about the point of interest desired. Specifically, if the user recognizes the appearance of trainee 113, the user could search through original track 3 comprising video of the trainees, through track 8, comprising a high resolution version of track 3, or through a stacked representation of tracks 3 or 8. If the user does not recognize the appearance of trainee 113, but does recognize his voice, the user could search through original track 1, comprising audio of the training session, or a stacked representation of track 1.
- the user manipulates the keyboard or computer mouse.
- the user can manipulate the pointer displayed on the monitor, as shown in Fig. 7B, using the computer mouse to select Window C.
- Window C Once Window C is selected, the user can access a menu of searching options corresponding to track 3, which options are described above, by selecting the MENU command.
- the user selects the stacked representation of track 3 extending from the current relative time to the end time of the training session.
- the user control unit 500-1 detects the user's selection and transmits it to the delivery processor 400.
- the delivery controller 403 of the delivery processor 400 responds by causing the event processor 405 to pause delivery of all tracks to the user control unit 500-1, generate a stacked representation of track 3 over the relative time period specified, and deliver the stacked representation to the user control unit 500- 1 (Step 865 of Fig. 8).
- the stacked representation is illustratively presented on the user's monitor in overlay fashion over track 3, as shown in Fig. 7C.
- the user can reverse or advance frame by frame in the training session. Therefore, as the user moves the pointer along the frames of the stacked representation, the frame pointed to appears elevated with respect to the other frames of the stacked representation.
- the user On locating a frame in the stacked representation depicting the point of interest for which the user is searching, the user enters the "PLAY" command, which is detected at the UCI 520.
- the UCI 520 retrieves from the presentation controller 550 the relative time associated with the frame pointed to and transfers the "PLAY" command coupled with this relative time to the delivery processor 400.
- the delivery controller 403 On receiving the "PLAY" command at the delivery processor 400, the delivery controller 403 causes the event processor 405 to discontinue the delivery of the stacked representation and resume the delivery of all tracks previously being delivered, but from the relative time associated with the "PLAY" command.
- the delivery controller 403 functions similarly on receiving a "PLAY" command from the user after the user has searched through any of the tracks or stacked representations using VCR-type commands. Specifically, the delivery controller 403 causes the event processor 405 to resume delivery of all tracks previously being delivered, but from the relative time associated with the "PLAY" command.
- the user could also have selected from the menu of options associated with track 3 the high resolution version (track 8).
- the delivery controller 403 On receiving the user's selection of this menu option at the delivery processor 400, the delivery controller 403 causes the event store 404 to transfer track 8 to the event processor 405. The delivery controller 403 then causes the event processor 405 to discontinue delivery of track 3 and then commence delivery of track 8 to the user control unit 500-1 (Step 870).
- track 8 is displayed on the user's monitor in a new, larger Window F.
- the user has a better view of the trainees and can read the expression on the face of trainee 114 as he listens to the lecturer's answer to the question posed by trainee 1 13.
- VCR-type commands include stop (STOP), pause (PAUSE), fast rewind (FAST-REW), fast forward (FAST-FWD), forward search (FWD-SEARCH), and reverse search (REV-SEARCH). Any of these commands can be selected by the user from the keyboard or the computer mouse.
- the delivery controller 403 causes the event processor 405 to adjust the delivery of the track selected (Step 875) and to send a message to the user control unit 500-1, which message appears on the user's monitor in overlay fashion, and acknowledges the command selected by the user. If no track is selected when VCR command is selected, all tracks are adjusted according to the command selected by the user.
- the delivery controller 403 When the "STOP" command is selected with respect to a window, the delivery controller 403 causes the event processor 405 to discontinue delivery of the track presented in the window. When the "PAUSE” command is selected, the delivery controller 403 causes the event processor 405 to continually re-deliver the current frame of each track being delivered. Audio tracks being delivered are temporarily discontinued. Selection of the "FAST REW” and “FAST-FWD” commands causes the delivery of the tracks presented in the window selected to be rapidly reversed or advanced, respectively.
- the delivery controller 403 In response to the selection of either of these commands, the delivery controller 403 causes the event processor 405 to halt the delivery of all tracks and send a signal to the user control unit 500-1 of a viewing position indicator (e.g., horizontal bar graph) which reflects the change in position in the training session. Upon reaching the desired new position in the training session, the user enters the "PLAY" command. On receiving the "PLAY" command at the delivery processor 400, the delivery controller 403 causes the event processor 405 to resume delivery of all tracks previously being delivered, but at the relative time associated with the new position in the training session (Step 880).
- a viewing position indicator e.g., horizontal bar graph
- the delivery controller 403 causes the event processor 405 to deliver blocks comprising frames from multiple groups of frames in a track to be delivered to the user control unit 500-1 in the reverse or forward direction, respectively. Delivery of these frames creates the perception that the track is being presented at a rate faster than real time. However, in fact the transmission rate from the delivery processor 400 to the user control unit 500- 1 does not change, but only frames from groups of frames in the track are being delivered.
- Our system and method for recording and playing back multimedia events entail capturing multimedia tracks representing various aspects of a multimedia event, coding the tracks into digitized blocks, and time stamping each block such that the tracks of the multimedia event are synchronized in time.
- the tracks are also compressed and preprocessed as instructed in a track mapping table.
- the tracks of the multimedia event are transmitted to the user as requested, and the delivery of the tracks is adjusted based upon relative time information associated with a new position established when searching through in the multimedia event.
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CA002198699A CA2198699C (en) | 1994-09-02 | 1995-08-07 | System and method for recording and playing back multimedia events |
EP95928757A EP0779007A1 (en) | 1994-09-02 | 1995-08-07 | System and method for recording and playing back multimedia events |
JP08509490A JP3143125B2 (en) | 1994-09-02 | 1995-08-07 | System and method for recording and playing multimedia events |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US08/300,824 | 1994-09-02 | ||
US08/300,824 US5613032A (en) | 1994-09-02 | 1994-09-02 | System and method for recording, playing back and searching multimedia events wherein video, audio and text can be searched and retrieved |
Publications (1)
Publication Number | Publication Date |
---|---|
WO1996008107A1 true WO1996008107A1 (en) | 1996-03-14 |
Family
ID=23160741
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US1995/009959 WO1996008107A1 (en) | 1994-09-02 | 1995-08-07 | System and method for recording and playing back multimedia events |
Country Status (5)
Country | Link |
---|---|
US (1) | US5613032A (en) |
EP (1) | EP0779007A1 (en) |
JP (1) | JP3143125B2 (en) |
CA (1) | CA2198699C (en) |
WO (1) | WO1996008107A1 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH1055391A (en) * | 1996-08-12 | 1998-02-24 | Fuji Xerox Co Ltd | Information reproducing device and material display device |
WO2008070105A2 (en) * | 2006-12-06 | 2008-06-12 | Carnegie Mellon University | System and method for capturing, editing, searching, and delivering multi-media content |
WO2008107686A2 (en) * | 2007-03-08 | 2008-09-12 | John Richard Mcintosh | Video imagery display system and method |
US7831916B2 (en) | 2001-06-15 | 2010-11-09 | Fry-Altec, Llc | Method, system, and program for creating, recording, and distributing digital stream contents |
US20230154187A1 (en) * | 2021-11-12 | 2023-05-18 | Unisys Corporation | Method of decomposing a lecture into events |
Families Citing this family (203)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6400996B1 (en) | 1999-02-01 | 2002-06-04 | Steven M. Hoffberg | Adaptive pattern recognition based control system and method |
US10361802B1 (en) | 1999-02-01 | 2019-07-23 | Blanding Hovenweep, Llc | Adaptive pattern recognition based control system and method |
US8352400B2 (en) | 1991-12-23 | 2013-01-08 | Hoffberg Steven M | Adaptive pattern recognition based controller apparatus and method and human-factored interface therefore |
US6418424B1 (en) | 1991-12-23 | 2002-07-09 | Steven M. Hoffberg | Ergonomic man-machine interface incorporating adaptive pattern recognition based control system |
US5903454A (en) | 1991-12-23 | 1999-05-11 | Hoffberg; Linda Irene | Human-factored interface corporating adaptive pattern recognition based controller apparatus |
US6850252B1 (en) | 1999-10-05 | 2005-02-01 | Steven M. Hoffberg | Intelligent electronic appliance system and method |
US7631343B1 (en) | 1993-03-24 | 2009-12-08 | Endgate LLC | Down-line transcription system using automatic tracking and revenue collection |
US5369704A (en) * | 1993-03-24 | 1994-11-29 | Engate Incorporated | Down-line transcription system for manipulating real-time testimony |
US7249026B1 (en) * | 1993-03-24 | 2007-07-24 | Engate Llc | Attorney terminal having outline preparation capabilities for managing trial proceedings |
US5708796A (en) * | 1994-11-18 | 1998-01-13 | Lucent Technologies Inc. | Method of retrieving continuous and non-continuous media data from a file system |
AU5027796A (en) * | 1995-03-07 | 1996-09-23 | Interval Research Corporation | System and method for selective recording of information |
JPH08249348A (en) * | 1995-03-13 | 1996-09-27 | Hitachi Ltd | Method and device for video retrieval |
EP0737979B1 (en) * | 1995-04-14 | 2006-10-25 | Kabushiki Kaisha Toshiba | Reproducing apparatus for reproducing a recording medium |
JP3645619B2 (en) * | 1995-06-23 | 2005-05-11 | 富士写真フイルム株式会社 | How to record and play back image data |
JPH0916457A (en) * | 1995-06-28 | 1997-01-17 | Fujitsu Ltd | Multimedia data retrieval system |
EP0838068B1 (en) * | 1995-07-10 | 2005-10-26 | Sarnoff Corporation | Method and system for rendering and combining images |
US6332147B1 (en) | 1995-11-03 | 2001-12-18 | Xerox Corporation | Computer controlled display system using a graphical replay device to control playback of temporal data representing collaborative activities |
US5786814A (en) * | 1995-11-03 | 1998-07-28 | Xerox Corporation | Computer controlled display system activities using correlated graphical and timeline interfaces for controlling replay of temporal data representing collaborative activities |
US6118925A (en) * | 1995-11-14 | 2000-09-12 | Hitachi Denshi Kabushiki Kaisha | Method of and system for confirming program materials to be broadcasted and then broadcasting the program materials, and recording medium having recorded therein a procedure for implementing the method |
US5751885A (en) * | 1995-12-19 | 1998-05-12 | O'loughlin; Maureen | Centralized video system |
US6049363A (en) * | 1996-02-05 | 2000-04-11 | Texas Instruments Incorporated | Object detection method and system for scene change analysis in TV and IR data |
US5826206A (en) * | 1996-03-12 | 1998-10-20 | Training Inovations Group, Llc | Debriefing systems and methods for retrieving and presenting multiple datastreams with time indication marks in time synchronism |
JP3534368B2 (en) * | 1996-04-03 | 2004-06-07 | 株式会社東芝 | Moving image processing method and moving image processing apparatus |
US6229953B1 (en) * | 1996-04-03 | 2001-05-08 | Nikon Corporation | Information input apparatus |
US7058286B2 (en) * | 1996-04-03 | 2006-06-06 | Nikon Corporation | Information input apparatus |
JP4456675B2 (en) * | 1996-04-17 | 2010-04-28 | 株式会社日立製作所 | Video recording / reproducing apparatus and video recording / reproducing method |
JP3230569B2 (en) * | 1996-04-24 | 2001-11-19 | 富士ゼロックス株式会社 | Information storage device, information storage method, and information storage and playback device |
US6160950A (en) * | 1996-07-18 | 2000-12-12 | Matsushita Electric Industrial Co., Ltd. | Method and apparatus for automatically generating a digest of a program |
JP3758754B2 (en) * | 1996-08-02 | 2006-03-22 | 富士ゼロックス株式会社 | Information storage / reproduction apparatus and information storage / reproduction method |
JPH10145784A (en) * | 1996-09-10 | 1998-05-29 | Sony Corp | Data decoding system, data decoding method, transmitting method, transmitter, receiver and receiving method |
US20010033732A1 (en) * | 1996-09-20 | 2001-10-25 | Masahiro Juen | Image reproduction apparatus for reproducing multiple image files |
US6961700B2 (en) * | 1996-09-24 | 2005-11-01 | Allvoice Computing Plc | Method and apparatus for processing the output of a speech recognition engine |
JP3360548B2 (en) * | 1996-10-25 | 2002-12-24 | 日本電気株式会社 | Information playback device |
US6031573A (en) * | 1996-10-31 | 2000-02-29 | Sensormatic Electronics Corporation | Intelligent video information management system performing multiple functions in parallel |
US5893062A (en) | 1996-12-05 | 1999-04-06 | Interval Research Corporation | Variable rate video playback with synchronized audio |
US6263507B1 (en) | 1996-12-05 | 2001-07-17 | Interval Research Corporation | Browser for use in navigating a body of information, with particular application to browsing information represented by audiovisual data |
JP3512098B2 (en) * | 1996-12-13 | 2004-03-29 | ソニー株式会社 | Information recording apparatus and method, and information reproducing apparatus and method |
JP3733984B2 (en) * | 1997-01-29 | 2006-01-11 | 富士ゼロックス株式会社 | Information storage device and information storage method |
US6614988B1 (en) * | 1997-03-28 | 2003-09-02 | Sharp Laboratories Of America, Inc. | Natural language labeling of video using multiple words |
US6735253B1 (en) | 1997-05-16 | 2004-05-11 | The Trustees Of Columbia University In The City Of New York | Methods and architecture for indexing and editing compressed video over the world wide web |
JP3988205B2 (en) * | 1997-05-27 | 2007-10-10 | ソニー株式会社 | Video signal recording / reproducing apparatus, video signal recording / reproducing method, video signal reproducing apparatus, and video signal reproducing method |
US7284187B1 (en) * | 1997-05-30 | 2007-10-16 | Aol Llc, A Delaware Limited Liability Company | Encapsulated document and format system |
US6625383B1 (en) * | 1997-07-11 | 2003-09-23 | Mitsubishi Denki Kabushiki Kaisha | Moving picture collection and event detection apparatus |
US6907450B1 (en) * | 1997-07-25 | 2005-06-14 | Siemens Aktiengesellschaft | Method and apparatus for the synchronized representation of network contents |
US7295752B1 (en) * | 1997-08-14 | 2007-11-13 | Virage, Inc. | Video cataloger system with audio track extraction |
US6360234B2 (en) | 1997-08-14 | 2002-03-19 | Virage, Inc. | Video cataloger system with synchronized encoders |
US6463444B1 (en) | 1997-08-14 | 2002-10-08 | Virage, Inc. | Video cataloger system with extensibility |
US6567980B1 (en) | 1997-08-14 | 2003-05-20 | Virage, Inc. | Video cataloger system with hyperlinked output |
US6665490B2 (en) * | 1998-04-01 | 2003-12-16 | Xerox Corporation | Obtaining and using data associating annotating activities with portions of recordings |
US6163510A (en) | 1998-06-30 | 2000-12-19 | International Business Machines Corporation | Multimedia search and indexing system and method of operation using audio cues with signal thresholds |
KR100324512B1 (en) * | 1998-07-14 | 2002-06-26 | 구자홍 | Real-time data recording and playback device and its control method |
US6243713B1 (en) | 1998-08-24 | 2001-06-05 | Excalibur Technologies Corp. | Multimedia document retrieval by application of multimedia queries to a unified index of multimedia data for a plurality of multimedia data types |
US6833865B1 (en) * | 1998-09-01 | 2004-12-21 | Virage, Inc. | Embedded metadata engines in digital capture devices |
JP3396639B2 (en) * | 1998-09-30 | 2003-04-14 | 株式会社東芝 | Hierarchical storage device and hierarchical storage control method |
US7900823B1 (en) | 1998-10-09 | 2011-03-08 | Diebold, Incorporated | Banking system controlled by data bearing records |
US7533806B1 (en) | 1998-10-09 | 2009-05-19 | Diebold, Incorporated | Reading of image data bearing record for comparison with stored user image in authorizing automated banking machine access |
US7143434B1 (en) | 1998-11-06 | 2006-11-28 | Seungyup Paek | Video description system and method |
WO2000028440A1 (en) * | 1998-11-06 | 2000-05-18 | The Trustees Of Columbia University In The City Ofnew York | Systems and methods for interoperable multimedia content descriptions |
US7254285B1 (en) | 1998-11-06 | 2007-08-07 | Seungup Paek | Image description system and method |
US7174083B1 (en) * | 1998-12-30 | 2007-02-06 | Xerox Corporation | Systems and methods for using structured representations to index recordings of activity |
US6934461B1 (en) * | 1999-01-05 | 2005-08-23 | Interval Research Corporation | Low attention recording, with particular application to social recording |
US6825875B1 (en) * | 1999-01-05 | 2004-11-30 | Interval Research Corporation | Hybrid recording unit including portable video recorder and auxillary device |
US7966078B2 (en) | 1999-02-01 | 2011-06-21 | Steven Hoffberg | Network media appliance system and method |
US6227862B1 (en) | 1999-02-12 | 2001-05-08 | Advanced Drivers Education Products And Training, Inc. | Driver training system |
US7142777B1 (en) * | 1999-02-23 | 2006-11-28 | Canon Kabushiki Kaisha | Recording and reproducing apparatus and method generating recording location table for plurality of programs received in multiplexed data train |
US6636238B1 (en) * | 1999-04-20 | 2003-10-21 | International Business Machines Corporation | System and method for linking an audio stream with accompanying text material |
JP2000350156A (en) * | 1999-06-09 | 2000-12-15 | Hitachi Ltd | Method for storing moving picture information and recording medium recording the information |
AU5484200A (en) * | 1999-06-18 | 2001-01-09 | Intel Corporation | Systems and methods for enhanced visual presentation using interactive video streams |
US7797730B2 (en) * | 1999-06-24 | 2010-09-14 | Engate Llc | Downline transcription system using automatic tracking and revenue collection |
US6871107B1 (en) * | 1999-07-01 | 2005-03-22 | Ftr Pty, Ltd. | Digital audio transcription system |
JP3895892B2 (en) | 1999-09-22 | 2007-03-22 | 株式会社東芝 | Multimedia information collection management device and storage medium storing program |
US7155735B1 (en) | 1999-10-08 | 2006-12-26 | Vulcan Patents Llc | System and method for the broadcast dissemination of time-ordered data |
US8250617B2 (en) | 1999-10-29 | 2012-08-21 | Opentv, Inc. | System and method for providing multi-perspective instant replay |
US7050110B1 (en) * | 1999-10-29 | 2006-05-23 | Intel Corporation | Method and system for generating annotations video |
US6564003B2 (en) * | 1999-11-04 | 2003-05-13 | Xm Satellite Radio Inc. | Method and apparatus for composite data stream storage and playback |
US7412643B1 (en) * | 1999-11-23 | 2008-08-12 | International Business Machines Corporation | Method and apparatus for linking representation and realization data |
US6694352B1 (en) * | 1999-12-16 | 2004-02-17 | Microsoft Corporation | Methods and systems for notifying clients concerning live electronic presentations |
US7051110B2 (en) * | 1999-12-20 | 2006-05-23 | Matsushita Electric Industrial Co., Ltd. | Data reception/playback method and apparatus and data transmission method and apparatus for providing playback control functions |
US6757682B1 (en) | 2000-01-28 | 2004-06-29 | Interval Research Corporation | Alerting users to items of current interest |
US7962948B1 (en) * | 2000-04-07 | 2011-06-14 | Virage, Inc. | Video-enabled community building |
US7260564B1 (en) | 2000-04-07 | 2007-08-21 | Virage, Inc. | Network video guide and spidering |
US7222163B1 (en) | 2000-04-07 | 2007-05-22 | Virage, Inc. | System and method for hosting of video content over a network |
US8171509B1 (en) | 2000-04-07 | 2012-05-01 | Virage, Inc. | System and method for applying a database to video multimedia |
US7194186B1 (en) * | 2000-04-21 | 2007-03-20 | Vulcan Patents Llc | Flexible marking of recording data by a recording unit |
JP3380212B2 (en) * | 2000-05-18 | 2003-02-24 | 明宏 川村 | Multiple listening training apparatus and recording medium for multiple listening training |
US20020016719A1 (en) * | 2000-06-19 | 2002-02-07 | Nemeth Louis G. | Methods and systems for providing medical data to a third party in accordance with configurable distribution parameters |
US8595372B2 (en) * | 2000-09-12 | 2013-11-26 | Wag Acquisition, Llc | Streaming media buffering system |
US7716358B2 (en) | 2000-09-12 | 2010-05-11 | Wag Acquisition, Llc | Streaming media buffering system |
US6766376B2 (en) | 2000-09-12 | 2004-07-20 | Sn Acquisition, L.L.C | Streaming media buffering system |
US6421358B1 (en) * | 2000-09-15 | 2002-07-16 | Mediaone Group, Inc. | Method and system for delivering a synchronized stream of broadcast event data, live event data, and static event data over a hybrid fiber coax network |
US20020042834A1 (en) * | 2000-10-10 | 2002-04-11 | Reelscore, Llc | Network music and video distribution and synchronization system |
US6916521B2 (en) * | 2000-12-13 | 2005-07-12 | Fuji Photo Film Co., Ltd. | Cleaning medium for magnetic recording apparatus |
CA2328795A1 (en) | 2000-12-19 | 2002-06-19 | Advanced Numerical Methods Ltd. | Applications and performance enhancements for detail-in-context viewing technology |
US20020133520A1 (en) * | 2001-03-15 | 2002-09-19 | Matthew Tanner | Method of preparing a multimedia recording of a live presentation |
US20020188772A1 (en) * | 2001-04-02 | 2002-12-12 | Mark Radcliffe | Media production methods and systems |
CA2345803A1 (en) | 2001-05-03 | 2002-11-03 | Idelix Software Inc. | User interface elements for pliable display technology implementations |
US8416266B2 (en) | 2001-05-03 | 2013-04-09 | Noregin Assetts N.V., L.L.C. | Interacting with detail-in-context presentations |
US9760235B2 (en) * | 2001-06-12 | 2017-09-12 | Callahan Cellular L.L.C. | Lens-defined adjustment of displays |
WO2002101534A1 (en) * | 2001-06-12 | 2002-12-19 | Idelix Software Inc. | Graphical user interface with zoom for detail-in-context presentations |
US7203693B2 (en) * | 2001-06-12 | 2007-04-10 | Lucent Technologies Inc. | Instantly indexed databases for multimedia content analysis and retrieval |
US7084886B2 (en) | 2002-07-16 | 2006-08-01 | Idelix Software Inc. | Using detail-in-context lenses for accurate digital image cropping and measurement |
FR2828044A1 (en) * | 2001-07-26 | 2003-01-31 | Net Television Factory | Multimedia Internet process/method broadcasting enriched digital word stream having structured relationship user/moderator/presentation services with levels managing enriched words/routing server set exchanges/managing broadcast servers. |
US7778925B2 (en) * | 2001-09-18 | 2010-08-17 | Sony Corporation | Audio and video digital content delivery |
US20030065655A1 (en) * | 2001-09-28 | 2003-04-03 | International Business Machines Corporation | Method and apparatus for detecting query-driven topical events using textual phrases on foils as indication of topic |
CA2361341A1 (en) * | 2001-11-07 | 2003-05-07 | Idelix Software Inc. | Use of detail-in-context presentation on stereoscopically paired images |
US6920428B2 (en) * | 2001-11-19 | 2005-07-19 | The Friday Group Llc | Method of selling and distributing articles associated with live events |
AU2002351310A1 (en) * | 2001-12-06 | 2003-06-23 | The Trustees Of Columbia University In The City Of New York | System and method for extracting text captions from video and generating video summaries |
JP3780252B2 (en) * | 2001-12-25 | 2006-05-31 | キヤノン株式会社 | Recording / reproducing apparatus and recording / reproducing method |
CA2370752A1 (en) * | 2002-02-05 | 2003-08-05 | Idelix Software Inc. | Fast rendering of pyramid lens distorted raster images |
US20030200192A1 (en) * | 2002-04-18 | 2003-10-23 | Bell Brian L. | Method of organizing information into topical, temporal, and location associations for organizing, selecting, and distributing information |
WO2003091850A2 (en) * | 2002-04-26 | 2003-11-06 | The Trustees Of Columbia University In The City Of New York | Method and system for optimal video transcoding based on utility function descriptors |
CA2386560A1 (en) * | 2002-05-15 | 2003-11-15 | Idelix Software Inc. | Controlling optical hardware and dynamic data viewing systems with detail-in-context viewing tools |
AU2003249663A1 (en) * | 2002-05-28 | 2003-12-12 | Yesvideo, Inc. | Summarization of a visual recording |
US7260257B2 (en) * | 2002-06-19 | 2007-08-21 | Microsoft Corp. | System and method for whiteboard and audio capture |
US20040006628A1 (en) * | 2002-07-03 | 2004-01-08 | Scott Shepard | Systems and methods for providing real-time alerting |
US7337115B2 (en) * | 2002-07-03 | 2008-02-26 | Verizon Corporate Services Group Inc. | Systems and methods for providing acoustic classification |
US8120624B2 (en) * | 2002-07-16 | 2012-02-21 | Noregin Assets N.V. L.L.C. | Detail-in-context lenses for digital image cropping, measurement and online maps |
CA2393887A1 (en) | 2002-07-17 | 2004-01-17 | Idelix Software Inc. | Enhancements to user interface for detail-in-context data presentation |
CA2406131A1 (en) * | 2002-09-30 | 2004-03-30 | Idelix Software Inc. | A graphical user interface using detail-in-context folding |
US20070097109A1 (en) * | 2005-10-18 | 2007-05-03 | Idelix Software Inc. | Method and system for generating detail-in-context presentations in client/server systems |
CA2407383A1 (en) * | 2002-10-10 | 2004-04-10 | Idelix Software Inc. | Editing multiple layers of a presentation using detail-in-context lenses |
CA2449888A1 (en) | 2003-11-17 | 2005-05-17 | Idelix Software Inc. | Navigating large images using detail-in-context fisheye rendering techniques |
US20040138894A1 (en) * | 2002-10-17 | 2004-07-15 | Daniel Kiecza | Speech transcription tool for efficient speech transcription |
CA2411898A1 (en) | 2002-11-15 | 2004-05-15 | Idelix Software Inc. | A method and system for controlling access to detail-in-context presentations |
US7761505B2 (en) * | 2002-11-18 | 2010-07-20 | Openpeak Inc. | System, method and computer program product for concurrent performance of video teleconference and delivery of multimedia presentation and archiving of same |
US7349008B2 (en) * | 2002-11-30 | 2008-03-25 | Microsoft Corporation | Automated camera management system and method for capturing presentations using videography rules |
US7483618B1 (en) | 2003-12-04 | 2009-01-27 | Yesvideo, Inc. | Automatic editing of a visual recording to eliminate content of unacceptably low quality and/or very little or no interest |
PL364275A1 (en) * | 2003-12-30 | 2005-07-11 | Advanced Digital Broadcast Ltd. | Method and system for recording and tracing markers in a data flow |
US7394974B2 (en) | 2004-01-26 | 2008-07-01 | Sony Corporation | System and method for associating presented digital content within recorded digital stream and method for its playback from precise location |
US7486302B2 (en) * | 2004-04-14 | 2009-02-03 | Noregin Assets N.V., L.L.C. | Fisheye lens graphical user interfaces |
US20060031879A1 (en) * | 2004-04-30 | 2006-02-09 | Vulcan Inc. | Management and non-linear presentation of news-related broadcasted or streamed multimedia content |
US20060031885A1 (en) * | 2004-04-30 | 2006-02-09 | Vulcan Inc. | Management and non-linear presentation of music-related broadcasted or streamed multimedia content |
US20060031916A1 (en) * | 2004-04-30 | 2006-02-09 | Vulcan Inc. | Management and non-linear presentation of broadcasted or streamed multimedia content |
US8106927B2 (en) * | 2004-05-28 | 2012-01-31 | Noregin Assets N.V., L.L.C. | Graphical user interfaces and occlusion prevention for fisheye lenses with line segment foci |
US9317945B2 (en) * | 2004-06-23 | 2016-04-19 | Callahan Cellular L.L.C. | Detail-in-context lenses for navigation |
NO20043156A (en) * | 2004-07-23 | 2006-01-23 | Telenor Asa | Device and method for reproducing audiovisual content |
US7714859B2 (en) | 2004-09-03 | 2010-05-11 | Shoemaker Garth B D | Occlusion reduction and magnification for multidimensional data presentations |
US7995078B2 (en) | 2004-09-29 | 2011-08-09 | Noregin Assets, N.V., L.L.C. | Compound lenses for multi-source data presentation |
WO2006096612A2 (en) * | 2005-03-04 | 2006-09-14 | The Trustees Of Columbia University In The City Of New York | System and method for motion estimation and mode decision for low-complexity h.264 decoder |
US7580036B2 (en) | 2005-04-13 | 2009-08-25 | Catherine Montagnese | Detail-in-context terrain displacement algorithm with optimizations |
JP2007004896A (en) * | 2005-06-23 | 2007-01-11 | Toshiba Corp | Information storage medium, information transfer method, information reproducing method, and information recording method |
JP2007052564A (en) * | 2005-08-16 | 2007-03-01 | Fuji Xerox Co Ltd | Information processing system and information processing method |
US8977636B2 (en) * | 2005-08-19 | 2015-03-10 | International Business Machines Corporation | Synthesizing aggregate data of disparate data types into data of a uniform data type |
US20070074116A1 (en) * | 2005-09-29 | 2007-03-29 | Teleios, Inc. | Multi-pane navigation/synchronization in a multimedia presentation system |
US8031206B2 (en) * | 2005-10-12 | 2011-10-04 | Noregin Assets N.V., L.L.C. | Method and system for generating pyramid fisheye lens detail-in-context presentations |
US20070099683A1 (en) * | 2005-10-29 | 2007-05-03 | Similarity, Inc., A Delaware Corporation | Interactive pictures |
US8756057B2 (en) * | 2005-11-02 | 2014-06-17 | Nuance Communications, Inc. | System and method using feedback speech analysis for improving speaking ability |
US8694319B2 (en) * | 2005-11-03 | 2014-04-08 | International Business Machines Corporation | Dynamic prosody adjustment for voice-rendering synthesized data |
US9135339B2 (en) | 2006-02-13 | 2015-09-15 | International Business Machines Corporation | Invoking an audio hyperlink |
KR100856403B1 (en) * | 2006-03-03 | 2008-09-04 | 삼성전자주식회사 | Video conference recording method and video conference terminal for the same |
US7823056B1 (en) * | 2006-03-15 | 2010-10-26 | Adobe Systems Incorporated | Multiple-camera video recording |
US7983473B2 (en) | 2006-04-11 | 2011-07-19 | Noregin Assets, N.V., L.L.C. | Transparency adjustment of a presentation |
US7623755B2 (en) | 2006-08-17 | 2009-11-24 | Adobe Systems Incorporated | Techniques for positioning audio and video clips |
US8275243B2 (en) * | 2006-08-31 | 2012-09-25 | Georgia Tech Research Corporation | Method and computer program product for synchronizing, displaying, and providing access to data collected from various media |
US20080147692A1 (en) * | 2006-12-14 | 2008-06-19 | General Motors Corporation | Method for manipulating the contents of an xml-based message |
US9318100B2 (en) | 2007-01-03 | 2016-04-19 | International Business Machines Corporation | Supplementing audio recorded in a media file |
US7465241B2 (en) * | 2007-03-23 | 2008-12-16 | Acushnet Company | Functionalized, crosslinked, rubber nanoparticles for use in golf ball castable thermoset layers |
US20080258909A1 (en) * | 2007-04-18 | 2008-10-23 | Brian Nedward Meyer | Methods and systems for automated data management |
US20080313542A1 (en) * | 2007-06-15 | 2008-12-18 | Trial Technologies, Inc. | System and method for witness testimony collection |
US9026938B2 (en) * | 2007-07-26 | 2015-05-05 | Noregin Assets N.V., L.L.C. | Dynamic detail-in-context user interface for application access and content access on electronic displays |
US9666091B2 (en) | 2008-01-10 | 2017-05-30 | Lifelong Driver Llc | Driver training system |
US8601494B2 (en) * | 2008-01-14 | 2013-12-03 | International Business Machines Corporation | Multi-event type monitoring and searching |
WO2009126785A2 (en) * | 2008-04-10 | 2009-10-15 | The Trustees Of Columbia University In The City Of New York | Systems and methods for image archaeology |
WO2009155281A1 (en) * | 2008-06-17 | 2009-12-23 | The Trustees Of Columbia University In The City Of New York | System and method for dynamically and interactively searching media data |
US8208905B2 (en) | 2008-06-27 | 2012-06-26 | Microsoft Corporation | Discovering an event using a personal preference list and presenting matching events to a user on a display |
US8671069B2 (en) | 2008-12-22 | 2014-03-11 | The Trustees Of Columbia University, In The City Of New York | Rapid image annotation via brain state decoding and visual pattern mining |
US8099473B2 (en) | 2008-12-31 | 2012-01-17 | Apple Inc. | Variant streams for real-time or near real-time streaming |
US8260877B2 (en) | 2008-12-31 | 2012-09-04 | Apple Inc. | Variant streams for real-time or near real-time streaming to provide failover protection |
US8156089B2 (en) | 2008-12-31 | 2012-04-10 | Apple, Inc. | Real-time or near real-time streaming with compressed playlists |
US8578272B2 (en) | 2008-12-31 | 2013-11-05 | Apple Inc. | Real-time or near real-time streaming |
US8990083B1 (en) | 2009-09-30 | 2015-03-24 | Cisco Technology, Inc. | System and method for generating personal vocabulary from network data |
US8489390B2 (en) * | 2009-09-30 | 2013-07-16 | Cisco Technology, Inc. | System and method for generating vocabulary from network data |
US8468195B1 (en) | 2009-09-30 | 2013-06-18 | Cisco Technology, Inc. | System and method for controlling an exchange of information in a network environment |
US9201965B1 (en) | 2009-09-30 | 2015-12-01 | Cisco Technology, Inc. | System and method for providing speech recognition using personal vocabulary in a network environment |
GB201105502D0 (en) | 2010-04-01 | 2011-05-18 | Apple Inc | Real time or near real time streaming |
US8560642B2 (en) | 2010-04-01 | 2013-10-15 | Apple Inc. | Real-time or near real-time streaming |
US8805963B2 (en) | 2010-04-01 | 2014-08-12 | Apple Inc. | Real-time or near real-time streaming |
CN102238179B (en) | 2010-04-07 | 2014-12-10 | 苹果公司 | Real-time or near real-time streaming |
US8935274B1 (en) | 2010-05-12 | 2015-01-13 | Cisco Technology, Inc | System and method for deriving user expertise based on data propagating in a network environment |
US9465795B2 (en) | 2010-12-17 | 2016-10-11 | Cisco Technology, Inc. | System and method for providing feeds based on activity in a network environment |
US8667169B2 (en) | 2010-12-17 | 2014-03-04 | Cisco Technology, Inc. | System and method for providing argument maps based on activity in a network environment |
US8553065B2 (en) | 2011-04-18 | 2013-10-08 | Cisco Technology, Inc. | System and method for providing augmented data in a network environment |
US8528018B2 (en) | 2011-04-29 | 2013-09-03 | Cisco Technology, Inc. | System and method for evaluating visual worthiness of video data in a network environment |
US8620136B1 (en) * | 2011-04-30 | 2013-12-31 | Cisco Technology, Inc. | System and method for media intelligent recording in a network environment |
US8909624B2 (en) | 2011-05-31 | 2014-12-09 | Cisco Technology, Inc. | System and method for evaluating results of a search query in a network environment |
US8843586B2 (en) | 2011-06-03 | 2014-09-23 | Apple Inc. | Playlists for real-time or near real-time streaming |
US8856283B2 (en) | 2011-06-03 | 2014-10-07 | Apple Inc. | Playlists for real-time or near real-time streaming |
US8886797B2 (en) | 2011-07-14 | 2014-11-11 | Cisco Technology, Inc. | System and method for deriving user expertise based on data propagating in a network environment |
US9349300B2 (en) | 2011-10-31 | 2016-05-24 | Lifelong Driver Llc | Senior driver training |
US10013893B2 (en) | 2011-10-31 | 2018-07-03 | Lifelong Driver Llc | Driver training |
US8831403B2 (en) | 2012-02-01 | 2014-09-09 | Cisco Technology, Inc. | System and method for creating customized on-demand video reports in a network environment |
US9070409B1 (en) * | 2014-08-04 | 2015-06-30 | Nathan Robert Yntema | System and method for visually representing a recorded audio meeting |
US9307317B2 (en) | 2014-08-29 | 2016-04-05 | Coban Technologies, Inc. | Wireless programmable microphone apparatus and system for integrated surveillance system devices |
US9225527B1 (en) | 2014-08-29 | 2015-12-29 | Coban Technologies, Inc. | Hidden plug-in storage drive for data integrity |
US9342227B2 (en) | 2014-09-02 | 2016-05-17 | Microsoft Technology Licensing, Llc | Semantic card view |
US10614726B2 (en) | 2014-12-08 | 2020-04-07 | Life Long Driver, Llc | Behaviorally-based crash avoidance system |
US10165171B2 (en) | 2016-01-22 | 2018-12-25 | Coban Technologies, Inc. | Systems, apparatuses, and methods for controlling audiovisual apparatuses |
US10789840B2 (en) | 2016-05-09 | 2020-09-29 | Coban Technologies, Inc. | Systems, apparatuses and methods for detecting driving behavior and triggering actions based on detected driving behavior |
US10370102B2 (en) | 2016-05-09 | 2019-08-06 | Coban Technologies, Inc. | Systems, apparatuses and methods for unmanned aerial vehicle |
US10152858B2 (en) | 2016-05-09 | 2018-12-11 | Coban Technologies, Inc. | Systems, apparatuses and methods for triggering actions based on data capture and characterization |
US9916822B1 (en) * | 2016-10-07 | 2018-03-13 | Gopro, Inc. | Systems and methods for audio remixing using repeated segments |
US20190020854A1 (en) * | 2017-07-13 | 2019-01-17 | Bryce Jonathan Di Geronimo | Method and system for recording online tutoring |
CN111052732B (en) | 2017-09-27 | 2022-01-28 | 杜比实验室特许公司 | Method for processing sequence of video frames, computer readable storage medium and data processing system |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5045940A (en) * | 1989-12-22 | 1991-09-03 | Avid Technology, Inc. | Video/audio transmission systsem and method |
US5262875A (en) * | 1992-04-30 | 1993-11-16 | Instant Video Technologies, Inc. | Audio/video file server including decompression/playback means |
US5371532A (en) * | 1992-05-15 | 1994-12-06 | Bell Communications Research, Inc. | Communications architecture and method for distributing information services |
US5414455A (en) * | 1993-07-07 | 1995-05-09 | Digital Equipment Corporation | Segmented video on demand system |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4630110A (en) * | 1984-02-15 | 1986-12-16 | Supervision Control Systems, Inc. | Surveillance system |
US5130792A (en) * | 1990-02-01 | 1992-07-14 | Usa Video Inc. | Store and forward video system |
US5251909A (en) * | 1991-05-28 | 1993-10-12 | Reed Michael J | Secured high throughput data channel for public broadcast system |
US5359601A (en) * | 1992-10-30 | 1994-10-25 | Scientific-Atlanta, Inc. | Apparatus providing dynamic virtual service selection in a multi-service communications system |
US5442390A (en) * | 1993-07-07 | 1995-08-15 | Digital Equipment Corporation | Video on demand with memory accessing and or like functions |
US5491511A (en) * | 1994-02-04 | 1996-02-13 | Odle; James A. | Multimedia capture and audit system for a video surveillance network |
-
1994
- 1994-09-02 US US08/300,824 patent/US5613032A/en not_active Expired - Lifetime
-
1995
- 1995-08-07 EP EP95928757A patent/EP0779007A1/en not_active Withdrawn
- 1995-08-07 JP JP08509490A patent/JP3143125B2/en not_active Expired - Lifetime
- 1995-08-07 WO PCT/US1995/009959 patent/WO1996008107A1/en not_active Application Discontinuation
- 1995-08-07 CA CA002198699A patent/CA2198699C/en not_active Expired - Lifetime
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5045940A (en) * | 1989-12-22 | 1991-09-03 | Avid Technology, Inc. | Video/audio transmission systsem and method |
US5262875A (en) * | 1992-04-30 | 1993-11-16 | Instant Video Technologies, Inc. | Audio/video file server including decompression/playback means |
US5371532A (en) * | 1992-05-15 | 1994-12-06 | Bell Communications Research, Inc. | Communications architecture and method for distributing information services |
US5414455A (en) * | 1993-07-07 | 1995-05-09 | Digital Equipment Corporation | Segmented video on demand system |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH1055391A (en) * | 1996-08-12 | 1998-02-24 | Fuji Xerox Co Ltd | Information reproducing device and material display device |
US7831916B2 (en) | 2001-06-15 | 2010-11-09 | Fry-Altec, Llc | Method, system, and program for creating, recording, and distributing digital stream contents |
US8276082B2 (en) | 2001-06-15 | 2012-09-25 | Fry-Altec, Inc. | Method and computer readable media for organizing digital stream contents |
WO2008070105A2 (en) * | 2006-12-06 | 2008-06-12 | Carnegie Mellon University | System and method for capturing, editing, searching, and delivering multi-media content |
WO2008070105A3 (en) * | 2006-12-06 | 2009-04-09 | Univ Carnegie Mellon | System and method for capturing, editing, searching, and delivering multi-media content |
US8437409B2 (en) | 2006-12-06 | 2013-05-07 | Carnagie Mellon University | System and method for capturing, editing, searching, and delivering multi-media content |
US8910225B2 (en) | 2006-12-06 | 2014-12-09 | Carnegie Mellon University, Center for Technology Transfer | System and method for capturing, editing, searching, and delivering multi-media content with local and global time |
US9584571B2 (en) | 2006-12-06 | 2017-02-28 | Carnegie Mellon University | System and method for capturing, editing, searching, and delivering multi-media content with local and global time |
WO2008107686A2 (en) * | 2007-03-08 | 2008-09-12 | John Richard Mcintosh | Video imagery display system and method |
WO2008107686A3 (en) * | 2007-03-08 | 2008-11-06 | John Richard Mcintosh | Video imagery display system and method |
US20230154187A1 (en) * | 2021-11-12 | 2023-05-18 | Unisys Corporation | Method of decomposing a lecture into events |
Also Published As
Publication number | Publication date |
---|---|
JP3143125B2 (en) | 2001-03-07 |
US5613032A (en) | 1997-03-18 |
CA2198699C (en) | 1999-10-12 |
CA2198699A1 (en) | 1996-03-14 |
EP0779007A1 (en) | 1997-06-18 |
JPH09510852A (en) | 1997-10-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US5613032A (en) | System and method for recording, playing back and searching multimedia events wherein video, audio and text can be searched and retrieved | |
JP3733984B2 (en) | Information storage device and information storage method | |
US6088026A (en) | Method and apparatus for multimedia information association to an electronic calendar event | |
US20020091658A1 (en) | Multimedia electronic education system and method | |
US7085842B2 (en) | Line navigation conferencing system | |
US5956037A (en) | Video information providing/receiving system | |
EP0846395B1 (en) | Method and apparatus for recording and indexing an audio and multimedia conference | |
US6263507B1 (en) | Browser for use in navigating a body of information, with particular application to browsing information represented by audiovisual data | |
US5808662A (en) | Synchronized, interactive playback of digital movies across a network | |
US6789228B1 (en) | Method and system for the storage and retrieval of web-based education materials | |
RU2219677C2 (en) | Device for video signal access and control through computer network including image correction | |
US7352385B2 (en) | Video editing system, video editing method, recording/reproducing method of visual information, apparatus therefor, and communication system | |
US6697569B1 (en) | Automated conversion of a visual presentation into digital data format | |
US20110123972A1 (en) | System for automatic production of lectures and presentations for live or on-demand publishing and sharing | |
US20020036694A1 (en) | Method and system for the storage and retrieval of web-based educational materials | |
WO2001053922A2 (en) | System, method and computer program product for collection of opinion data | |
JP2004343756A5 (en) | ||
EP0469850A2 (en) | Method and apparatus for pre-recording, editing and playing back presentations on a computer system | |
US7831916B2 (en) | Method, system, and program for creating, recording, and distributing digital stream contents | |
JP2002176638A (en) | Data communication system and device, data communication method and recording medium | |
JPH07123273B2 (en) | Method of transmitting multimedia information | |
JP2003163911A (en) | Image reproducing method based on favorable public image information, image reproduction control system, server apparatus, client apparatus, image reproduction control program and its recording medium | |
CN1195445A (en) | Phone based dynamic image annotation | |
JP3757229B2 (en) | Lectures at academic conferences, editing systems for lectures, and knowledge content distribution systems | |
JP2002223264A (en) | Cooperative processing method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): CA JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE |
|
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
ENP | Entry into the national phase |
Ref document number: 2198699 Country of ref document: CA Ref country code: CA Ref document number: 2198699 Kind code of ref document: A Format of ref document f/p: F |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1995928757 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1995928757 Country of ref document: EP |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 1995928757 Country of ref document: EP |