WO2001069931A1 - Method and apparatus for video surveillance with defined zones - Google Patents
Method and apparatus for video surveillance with defined zones Download PDFInfo
- Publication number
- WO2001069931A1 WO2001069931A1 PCT/US2001/007330 US0107330W WO0169931A1 WO 2001069931 A1 WO2001069931 A1 WO 2001069931A1 US 0107330 W US0107330 W US 0107330W WO 0169931 A1 WO0169931 A1 WO 0169931A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- zone
- tracking
- control zone
- motion video
- field
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S11/00—Systems for determining distance or velocity not using reflection or reradiation
- G01S11/12—Systems for determining distance or velocity not using reflection or reradiation using electromagnetic waves other than radio waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S3/00—Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received
- G01S3/78—Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received using electromagnetic waves other than radio waves
- G01S3/782—Systems for determining direction or deviation from predetermined direction
- G01S3/785—Systems for determining direction or deviation from predetermined direction using adjustment of orientation of directivity characteristics of a detector or detector system to give a desired condition of signal derived from that detector or detector system
- G01S3/786—Systems for determining direction or deviation from predetermined direction using adjustment of orientation of directivity characteristics of a detector or detector system to give a desired condition of signal derived from that detector or detector system the desired condition being maintained automatically
- G01S3/7864—T.V. type tracking systems
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S5/00—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
- G01S5/16—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using electromagnetic waves other than radio waves
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19678—User interface
- G08B13/19686—Interfaces masking personal details for privacy, e.g. blurring faces, vehicle license plates
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/188—Capturing isolated or intermittent images triggered by the occurrence of a predetermined event, e.g. an object reaching a predetermined position
Definitions
- the present invention relates to the field of video surveillance systems using motion video cameras.
- U.S. Patent 5,473,369 (Abe) describes the use of a camera to detect and track a moving object without using conventional block matching. In the system described in Abe single object tracking is performed only after an object is placed within a frame on a screen; however, there is no user input device for manual target selection Moreover, Abe does not provide for camera movement to maintain line-of-site.
- U.S. Patent 5,629,988 (Burt) teaches electronic stabilization of a sequence of images with respect to one another but provides no tracking facility. Still other prior art solutions control camera movement to maintain line-of-sight between camera and object but lack arbitrary motion compensation or do not provide for automatic and user selected object tracking.
- U.S. Patent 5,434,621 (Yu) teaches a method for automatic zooming and automatic tracking of an object using a zoom lens but does not provide for reorienting the camera's line-of-sight.
- a method for defining a control zone in a field of view of a motion video camera comprising the steps of: displaying motion video data representative of the field of view of the motion video camera; receiving indication of a control zone type; and receiving indication of a control zone size within the field of view of the motion video camera.
- a system for defining control zones of different types in a field of view of a motion video camera comprising: a database containing a description for each of a plurality of control zone types; means for defining a control zone in a selected area of the field of view of the motion video camera, said control zone being of a type selected from one of said plurality of control zone types in said database; and means for displaying a received motion video signal from the motion video camera including an indication of said defined control zone.
- a computer readable medium having stored thereon computer-executable instructions for defining a control zone in a field of view of a motion video camera performing the steps comprising: displaying motion video data representative of the field of view of the motion video camera; receiving indication of a control zone type; and receiving indication of a control zone size within the field of view of the motion video camera.
- BRIEF DESCRIPTION OF THE DRAWINGS Figure 1 is an exemplary configuration of a video surveillance system having user- defined zones according to an embodiment of the present invention
- Figure 2 shows an exemplary field of view for a camera highlighting user-defined zones according to the present invention
- Figure 3 is a system diagram of a zone defining processing system the video surveillance system of Figure 1 according to an embodiment of the present invention
- Figure 4 is a flow chart illustrating a method of identifying user-defined zones according to the present invention
- Figure 5 is a flow chart illustrating a general method for video surveillance using the user-defined zones identified by the method in Figure 4.
- FIG. 1 illustrates a video surveillance system 100 according to an embodiment of the present invention.
- a motion video camera 108 has a field of view 118.
- a computer 102 receives and processes a video signal 112 from the motion video camera 108 and performs object tracking and detection to determine if there was movement in the field of view 118.
- the computer 102 contains a zone defining processing system ( Figure 3) for defining surveillance zones in the camera's 108 field of view.
- the computer 102 For a moveable motion video camera 108, the computer 102 generates a control signal 114 for servo controlled pan-tilt-zoom assembly 110.
- the control signal 114 is based on the current position of the servo controlled pan-tilt-zoom assembly 110 and information contained in the video signal 112. Such movement allows the motion video camera 108 to capture an object of interest in greater detail or improve the camera's 108 line of sight with the object of interest.
- the object detection and tracking techniques that are used can be, for example, those taught in Applicant's related applications entitled “Method and Apparatus for Object Tracking and Detection” and “Method and Apparatus for Remote Object Tracking and Detection", filed concurrently herewith, both of which are incorporated herein by reference.
- Object detection may be accomplished using any number of methods for image segmentation known in the art. For example, motion detection may be performed by frame differencing sequential pairs of video frames and applying thresholding techniques thereby yielding pixels within the processed image that reflect motion of objects with the field of view of the camera 108. Additional image processing techniques such as centroid analysis may then be applied to remove spurious motion. Kalman filtering may be applied over time to further remove random motion and to estimate motion of objects for the purpose of anticipating camera 108 repositioning and maintaining tracking when moving objects are temporarily occluded by stationary ones.
- the computer 102 uses the video signal and the current magnification of the camera 108 to create a control signal similar to the control signal of the moveable camera 114.
- the control signal for a fixed camera only uses a zoom function already on the camera 108 to capture the object of interest in greater detail.
- the video signal 112 received from video camera 108 is passed out from the computer 102 as video signal out 106 either directly to a display 104 or modified to include graphic information that may be used to set up response parameters of a tracking program, indicate an object that is actively being tracked or identify (or allow identification of) user- defined zones for tracking.
- a pointing device 116 may be a mouse or trackball and is the user input for modifying said response parameters or defining the tracking zones.
- the pointing device 116 may also be used by the user to select an object that appears within the field of view 118 such that tracking program residing on the computer 102 acknowledges user's selection and initiates tracking of the selected object.
- Figure 2 is an illustration of an exemplary field of view 200 for a camera highlighting user-defined zones according to the present invention.
- the environment that the motion video camera 108 is receiving within its field of view 200 is shown.
- the environment contains a house 220, a tree 230, a walkway 240, and a pond 250.
- the tracking zone 208 is shown on the display encompassing the house 220, the tree 230, the walkway 240, and the pond 250. Only for the purpose of illustration is the tracking zone 208 contained within a single field of view 118. For embodiments using a moveable camera 108, the tracking zone 208 may span several fields of view.
- An exclusion zone 250 is shown encompassing that portion of the pond 250 that is within the tracking zone 208.
- a black-out zone 202 is shown encompassing the leafy portion of the tree 230 which may generate tracking events when the leaves move in the wind. Since in this field of view 200, it is not anticipated that tracks will not enter into the perimeter, the entire region receives a black-out zone 202 rather than an exclusion zone 204.
- An entry zone 206 is shown encompassing the door of the house 220.
- a privacy zone 210 is shown encompassing the window of the house 220.
- FIG. 3 is a zone defining processing system 120 for defining surveillance zones in the camera's 108 field of view according to an embodiment of the present invention.
- the zone defining processing system 120 has a field of view (FOV) area definer 132 that connects interfaces to various devices (i.e. camera 108, input device 116, and display 104), a zone type database 128 and a tracking/monitoring controller 130.
- a camera interface 122 receives the video signal in from the motion video camera 108 and passes this signal to the FOV area definer 132.
- the FOV area definer 132 can pass this signal directly to a display interface 126 to be shown on the display 106 or the signal may be modified to include a graphic overlay containing information on zones within the field of view.
- the FOV area definer 132 may send a zone type selection menu to the display interface 126 to prompt a user to select a zone type using the input device 116.
- the zone type database 128 contains definitions for the different types of zones as well as corresponding actions for each zone type (i.e. do not initiate tracking objects but continue following previous tracks). Indication of a selected zone type is received at a input device interface 124. After a zone type has been selected the FOV area definer 132 provides the display interface 126 with a graphic overlay for the field of view to assist the user in drawing a zone of the selected type. After an indication of a drawn zone has been received through the input device interface 124, the FOV area definer 132 provides the display interface 126 with a graphic overlay indicating the defined zone.
- the tracking/monitoring controller 130 tracks and monitors defined objects within the field of view. Given a field of view the tracking/monitoring controller 130 consults the FOV area definer 132 to determine a mapping between defined zones and the field of view. The FOV area definer 132 provides the tracking/monitoring controller 130 with the mapping between defined zones and the field of view as well as a definition of the zone types and the corresponding actions for each zone type.
- FIG. 4 is a flow chart illustrating a method 300 of identifying user-defined zones according to the present invention. All of these zones are shown on the display 104 as graphic overlays to the video signal in 112 received from the camera 108 in step 302.
- An input device such as a mouse 116 is provided for selecting a zone type in step 304.
- the type of zone may be selected, for example, from a menu listing all zone types or applicable zone types given other settings.
- These zones may be selected and resized in step 306 by a user with a mouse 116 or similar pointing device by holding down a mouse button and moving the mouse 116 to resize a rectangular region over some portion of the display's image.
- the mouse 116 is moved to any of the four edges or corners of the display which results in the tracking program residing on the computer 102 sending control signal 114 to the pan-tilt-zoom assembly 110 resulting in an altered field of view.
- the mouse button is released when the desired region is encompassed by a rectangle.
- the method 300 is repeated for multiple instances of zones of a single type, as applicable, as for other types of zones. Multiple overlapping and non overlapping shapes may be used to define more complex zones. Once these regions are defined, the tracker performs in its environment according to the rules that are defined for each type of region.
- Fig. 5 is a flow chart illustrating a general method 400 for video surveillance using the user-defined zones identified by the method 300 in Figure 4.
- the video surveillance system continuously monitors a tracking zone (except for any black-out zones) in the field of view for movement 402.
- a tracking zone except for any black-out zones
- subsequent video frames are compared 406 to determine what zone the movement occurred in 408. If the movement is not in a black-out zone or exclusion zone the object is isolated 410 and tracking starts 412. If the movement is in a black-out zone or exclusion zone tracking is not initiated and the video surveillance system continues to monitor the tracking zone for movement 402.
- step 404 If movement is not detected in step 404 then the camera 108 is moved to the next field of view 403 to continue searching for movement 402. It is apparent to one skilled in the art that numerous modifications and departures from the specific embodiments described herein may be made without departing from the spirit and scope of the invention.
Abstract
In the field of motion video object tracking, environmental noise and other such sources of movement can lead to false tracking and tracking of unimportant objects. The present invention uses user-defined exclusion regions within a motion video camera's field of view to ignore the origination of object track data. Therefore, the user defines in advance, those regions that may generate false alarms within a motion video camera's field of view. The use of user-defined zones in conjunction with object detection and tracking techniques provides a video surveillance system that has greater object tracking accuracy than existing systems.
Description
METHOD AND APPARATUS FOR VIDEO SURVEILLANCE WITH DEFINED ZONES
CROSS REFERENCE TO RELATED APPLICATIONS This application claims the benefit of United States provisional patent application serial number 60/188,171 filed on March 10, 2000. United States Patent applications, also claiming the benefit of U.S. Provisional application no. 60/188,171, and entitled "Method and Apparatus for Object Tracking and Detection" and "Method and Apparatus for Object Surveillance with a Movable Camera" were filed concurrently herewith.
FIELD OF THE INVENTION
The present invention relates to the field of video surveillance systems using motion video cameras.
BACKGROUND OF THE INVENTION There are several shortcomings in current video surveillance systems that need to be overcome for widespread use of automatic detection and collection of relevant video data in response to scene stimulus without the need of a human operator present. When viewing a scene from a video camera a large amount of data is generated. The vast amount of data created produces a data reduction problem. Automatically detecting and accurately and reliably collecting image information of a moving object using a motion video camera is a difficult task. This task is made even more difficult when trying to detect, track and maintain camera line-of-sight using a single motion video camera without requiring human intervention.
U.S. Patent 5,473,369 (Abe) describes the use of a camera to detect and track a moving object without using conventional block matching. In the system described in Abe single object tracking is performed only after an object is placed within a frame on a screen; however, there is no user input device for manual target selection Moreover, Abe does not provide for camera movement to maintain line-of-site.
Other prior art solutions provide for image stabilization for a camera in arbitrary motion without object tracking functionality. U.S. Patent 5,629,988 (Burt) teaches electronic stabilization of a sequence of images with respect to one another but provides no tracking facility.
Still other prior art solutions control camera movement to maintain line-of-sight between camera and object but lack arbitrary motion compensation or do not provide for automatic and user selected object tracking. U.S. Patent 5,434,621 (Yu) teaches a method for automatic zooming and automatic tracking of an object using a zoom lens but does not provide for reorienting the camera's line-of-sight.
SUMMARY OF THE INVENTION It is an object of the present invention to provide a video surveillance method and system having increased accuracy. It is an object of the present invention to provide a method and system for reducing false alarms in motion video object detection and surveillance.
It is an object of the present invention to provide a method and system for identifying user-defined zones to control motion video tracking.
In accordance with one aspect of the present invention there is provided a method for defining a control zone in a field of view of a motion video camera, said method comprising the steps of: displaying motion video data representative of the field of view of the motion video camera; receiving indication of a control zone type; and receiving indication of a control zone size within the field of view of the motion video camera.
In accordance with another aspect of the present invention there is provided a system for defining control zones of different types in a field of view of a motion video camera, said system comprising: a database containing a description for each of a plurality of control zone types; means for defining a control zone in a selected area of the field of view of the motion video camera, said control zone being of a type selected from one of said plurality of control zone types in said database; and means for displaying a received motion video signal from the motion video camera including an indication of said defined control zone.
In accordance with a further aspect of the present invention there is provided a computer readable medium having stored thereon computer-executable instructions for defining a control zone in a field of view of a motion video camera performing the steps comprising: displaying motion video data representative of the field of view of the motion video camera; receiving indication of a control zone type; and receiving indication of a control zone size within the field of view of the motion video camera.
BRIEF DESCRIPTION OF THE DRAWINGS Figure 1 is an exemplary configuration of a video surveillance system having user- defined zones according to an embodiment of the present invention;
Figure 2 shows an exemplary field of view for a camera highlighting user-defined zones according to the present invention;
Figure 3 is a system diagram of a zone defining processing system the video surveillance system of Figure 1 according to an embodiment of the present invention; Figure 4 is a flow chart illustrating a method of identifying user-defined zones according to the present invention; and Figure 5 is a flow chart illustrating a general method for video surveillance using the user-defined zones identified by the method in Figure 4.
DETAILED DESCRIPTION Object tracks, derived from motion video data, are a very useful way to automatically detect alarm conditions within a camera's field of view for purposes such as security and surveillance. However, in the field of motion video object tracking, environmental noise and other such sources of movement can lead to false tracking and tracking of unimportant objects such as leaves falling from a tree.
The present invention uses user-defined regions within a motion video camera's field of view to ignore the origination of object track data but allow tracking of objects already being tracked before they entered that region. Therefore, the user defines in advance those regions that may generate false alarms within a motion video camera's field of view. The use of user-defined zones in conjunction with object detection and tracking techniques provides a video surveillance system that has greater object tracking accuracy than existing systems. Figure 1 illustrates a video surveillance system 100 according to an embodiment of the present invention. A motion video camera 108 has a field of view 118. A computer 102 receives and processes a video signal 112 from the motion video camera 108 and performs object tracking and detection to determine if there was movement in the field of view 118. The computer 102 contains a zone defining processing system (Figure 3) for defining surveillance zones in the camera's 108 field of view.
For a moveable motion video camera 108, the computer 102 generates a control signal 114 for servo controlled pan-tilt-zoom assembly 110. The control signal 114 is based on the current position of the servo controlled pan-tilt-zoom assembly 110 and information
contained in the video signal 112. Such movement allows the motion video camera 108 to capture an object of interest in greater detail or improve the camera's 108 line of sight with the object of interest. The object detection and tracking techniques that are used can be, for example, those taught in Applicant's related applications entitled "Method and Apparatus for Object Tracking and Detection" and "Method and Apparatus for Remote Object Tracking and Detection", filed concurrently herewith, both of which are incorporated herein by reference.
Object detection may be accomplished using any number of methods for image segmentation known in the art. For example, motion detection may be performed by frame differencing sequential pairs of video frames and applying thresholding techniques thereby yielding pixels within the processed image that reflect motion of objects with the field of view of the camera 108. Additional image processing techniques such as centroid analysis may then be applied to remove spurious motion. Kalman filtering may be applied over time to further remove random motion and to estimate motion of objects for the purpose of anticipating camera 108 repositioning and maintaining tracking when moving objects are temporarily occluded by stationary ones. Object tracking and detection is discussed in greater detail in applicant's co-pending related applications entitled "Method and Apparatus for Object Tracking and Detection" and "Method and Apparatus for Object Surveillance with a Movable Camera" filed concurrently herewith and incorporated herein by reference.
For a fixed motion video camera 108 (not shown) not having a servo controlled pan- tilt-zoom assembly 110, the computer 102 uses the video signal and the current magnification of the camera 108 to create a control signal similar to the control signal of the moveable camera 114. However, the control signal for a fixed camera only uses a zoom function already on the camera 108 to capture the object of interest in greater detail.
The video signal 112 received from video camera 108 is passed out from the computer 102 as video signal out 106 either directly to a display 104 or modified to include graphic information that may be used to set up response parameters of a tracking program, indicate an object that is actively being tracked or identify (or allow identification of) user- defined zones for tracking. A pointing device 116 may be a mouse or trackball and is the user input for modifying said response parameters or defining the tracking zones. The pointing device 116 may also be used by the user to select an object that appears within the field of view 118 such that tracking program residing on the computer 102 acknowledges user's selection and initiates tracking of the selected object.
Figure 2 is an illustration of an exemplary field of view 200 for a camera highlighting user-defined zones according to the present invention. In this illustration, the environment that the motion video camera 108 is receiving within its field of view 200 is shown. For the purpose of illustration, the environment contains a house 220, a tree 230, a walkway 240, and a pond 250. There are five different types of zones 202, 204, 206, 208, and 210 that control tracking behavior and subsequent video output. These zones are defined as follows:
The tracking zone 208 is shown on the display encompassing the house 220, the tree 230, the walkway 240, and the pond 250. Only for the purpose of illustration is the tracking zone 208 contained within a single field of view 118. For embodiments using a moveable camera 108, the tracking zone 208 may span several fields of view. An exclusion zone 250 is shown encompassing that portion of the pond 250 that is within the tracking zone 208. A black-out zone 202 is shown encompassing the leafy portion of the tree 230 which may generate tracking events when the leaves move in the wind. Since in this field of view 200, it is not anticipated that tracks will not enter into the perimeter, the entire region receives a black-out zone 202 rather than an exclusion zone 204. An entry zone 206 is shown encompassing the door of the house 220. A privacy zone 210 is shown encompassing the window of the house 220.
Figure 3 is a zone defining processing system 120 for defining surveillance zones in the camera's 108 field of view according to an embodiment of the present invention. The zone defining processing system 120 has a field of view (FOV) area definer 132 that connects interfaces to various devices (i.e. camera 108, input device 116, and display 104), a zone type database 128 and a tracking/monitoring controller 130. A camera interface 122 receives the video signal in from the motion video camera 108 and passes this signal to the FOV area definer 132. The FOV area definer 132 can pass this signal directly to a display interface 126 to be shown on the display 106 or the signal may be modified to include a graphic overlay containing information on zones within the field of view.
The FOV area definer 132 may send a zone type selection menu to the display interface 126 to prompt a user to select a zone type using the input device 116. The zone type database 128 contains definitions for the different types of zones as well as corresponding actions for each zone type (i.e. do not initiate tracking objects but continue following previous tracks). Indication of a selected zone type is received at a input device interface 124.
After a zone type has been selected the FOV area definer 132 provides the display interface 126 with a graphic overlay for the field of view to assist the user in drawing a zone of the selected type. After an indication of a drawn zone has been received through the input device interface 124, the FOV area definer 132 provides the display interface 126 with a graphic overlay indicating the defined zone.
The tracking/monitoring controller 130 tracks and monitors defined objects within the field of view. Given a field of view the tracking/monitoring controller 130 consults the FOV area definer 132 to determine a mapping between defined zones and the field of view. The FOV area definer 132 provides the tracking/monitoring controller 130 with the mapping between defined zones and the field of view as well as a definition of the zone types and the corresponding actions for each zone type.
Figure 4 is a flow chart illustrating a method 300 of identifying user-defined zones according to the present invention. All of these zones are shown on the display 104 as graphic overlays to the video signal in 112 received from the camera 108 in step 302. An input device such as a mouse 116 is provided for selecting a zone type in step 304. The type of zone may be selected, for example, from a menu listing all zone types or applicable zone types given other settings. These zones may be selected and resized in step 306 by a user with a mouse 116 or similar pointing device by holding down a mouse button and moving the mouse 116 to resize a rectangular region over some portion of the display's image. If the extent of a rectangular region exceeds a single field of view, the mouse 116 is moved to any of the four edges or corners of the display which results in the tracking program residing on the computer 102 sending control signal 114 to the pan-tilt-zoom assembly 110 resulting in an altered field of view. The mouse button is released when the desired region is encompassed by a rectangle. After the size of the zone has been drawn in step 306, the method 300 is repeated for multiple instances of zones of a single type, as applicable, as for other types of zones. Multiple overlapping and non overlapping shapes may be used to define more complex zones. Once these regions are defined, the tracker performs in its environment according to the rules that are defined for each type of region.
Fig. 5 is a flow chart illustrating a general method 400 for video surveillance using the user-defined zones identified by the method 300 in Figure 4. The video surveillance system continuously monitors a tracking zone (except for any black-out zones) in the field of view for movement 402. When movement is detected 404 subsequent video frames are compared 406 to determine what zone the movement occurred in 408. If the movement is not in a
black-out zone or exclusion zone the object is isolated 410 and tracking starts 412. If the movement is in a black-out zone or exclusion zone tracking is not initiated and the video surveillance system continues to monitor the tracking zone for movement 402. While an object is being tracked if it leaves the current field of view and enters a new field of view 414 the camera 108 may be moved 418 to center the tracked object in the camera's 108 field of view. Movement of the camera 108 to maintain a moving object in the field of view is discussed in more detail in applicant's co-pending application related applications entitled "Method and Apparatus for Object Surveillance with a Movable Camera" filed concurrently herewith and incorporated herein by reference. If the object enters a black-out zone or exits the tracking zone 416 then tracking stops and the system continues to monitor the tracking zone for movement 402. If the object has not entered a black-out zone or exited the tracking zone then the object continues to be tracked.
If movement is not detected in step 404 then the camera 108 is moved to the next field of view 403 to continue searching for movement 402. It is apparent to one skilled in the art that numerous modifications and departures from the specific embodiments described herein may be made without departing from the spirit and scope of the invention.
Claims
1. A method for defining a control zone in a field of view of a motion video camera, said method comprising the steps of: displaying motion video data representative of the field of view of the motion video camera; receiving indication of a control zone type; and receiving indication of a control zone size within the field of view of the motion video camera.
2. The method according to claim 1 further including the step of displaying graphics representative of the control zone size in association with the field of view and the motion video data.
3. The method according to claim 1 wherein the control zone type is selected from the group consisting of tracking, black-out, exclusion, entry and privacy.
4. A video surveillance method executed according to the control zone defined in claim 1 , said method comprising the steps of:
(a) detecting movement in a field of view of the motion video camera;
(b) determining if a moving object is in a tracking origination zone; (c) defining the moving object if the moving object is in a tracking origination zone; and
(d) tracking the defined moving object.
5. The video surveillance method according to claim 4 further comprising the steps of:
(e) determining if the defined moving object has entered a new control zone type;
(f) determining if the new control zone type is a tracking continuation zone; and
(g) repeating step (d) to (f) if the new control zone type is a tracking continuation zone.
6. The video surveillance method according to claim 4 further comprising the steps of:
(e) determining if the defined moving object has entered a new control zone type;
(f) determining if the new control zone type is a tracking continuation zone; and
(g) ceasing tracking of the defined moving object if the new control zone type is not a tracking continuation zone.
7. The video surveillance method according to claim 4 wherein the control zone type is selected from the group consisting of tracking, black-out, exclusion, entry and privacy.
8. The video surveillance method according to claim 7 wherein a tracking zone is a tracking origination zone and a tracking continuation zone and defines a region in which motion is monitored in the field of view of the motion video camera.
9. The video surveillance method according to claim 7 wherein a privacy zone only monitors movement.
10. The video surveillance method according to claim 7 wherein an exclusion zone is a tracking continuation zone.
11. The video surveillance method according to claim 7 wherein an entry zone is a tracking origination zone and a tracking continuation zone.
12. The video surveillance method according to claim 7 wherein a black-out zone is not monitored for movement.
13. A system for defining control zones of different types in a field of view of a motion video camera, said system comprising: a database containing a description for each of a plurality of control zone types; means for defining a control zone in a selected area of the field of view of the motion video camera, said control zone being of a type selected from one of said plurality of control zone types in said database; and means for displaying a received motion video signal from the motion video camera including an indication of said defined control zone.
14. The system according to claim 13 wherein said means for displaying includes means for providing a graphical representation of a size of said selected area of the field of view with the received motion video signal.
15. A video surveillance system using a motion video camera having control zones in the field of view thereof as defined in claim 13, said system comprising: means for detecting movement in the field of view of the motion video camera; means for determining a current control zone of the moving object; means for defining the moving object dependent on the current control zone of the moving object; and means for performing a tracking operation the defined moving object dependent on a control zone type of the current control zone.
16. A computer readable medium having stored thereon computer-executable instructions for defining a control zone in a field of view of a motion video camera performing the steps comprising: displaying motion video data representative of the field of view of the motion video camera; receiving indication of a control zone type; and receiving indication of a control zone size within the field of view of the motion video camera.
17. The computer readable medium according to claim 16 further including the step of displaying graphics representative of the control zone size in association with the motion video data.
18. The computer readable medium according to claim 16 wherein the control zone type is selected from the group consisting of tracking, black-out, exclusion, entry and privacy.
19. A computer readable medium having stored thereon computer-executable instructions for executing motion video camera surveillance according to the control zone defined in claim 16 performing the steps comprising:
(a) detecting movement in a field of view of a motion video camera;
(b) determining if a moving object is in a tracking origination zone;
(c) defining the moving object if the moving object is in a tracking origination zone; and
(d) tracking the defined moving object.
20. The computer readable medium according to claim 19 further comprising the steps of:
(e) determining if the defined moving object has entered a new control zone type;
(f) determining if the new control zone type is a tracking continuation zone; and
(g) repeating step (d) to (f) if the new control zone type is a tracking continuation zone.
21. The computer readable medium according to claim 19 further comprising the steps of:
(e) determining if the defined moving object has entered a new control zone type;
(f) determining if the new control zone type is a tracking continuation zone; and
(g) ceasing tracking of the defined moving object if the new control zone type is not a tracking continuation zone.
22. The computer readable medium according to claim 19 wherein the control zone type is selected from the group consisting of tracking, black-out, exclusion, entry and privacy.
23. The computer readable medium according to claim 22 wherein a tracking zone is a tracking origination zone and a tracking continuation zone and defines a region in which motion is monitored in the field of view of the motion video camera.
24. The computer readable medium according to claim 22 wherein a privacy zone only monitors movement.
25. The computer readable medium according to claim 22 wherein an exclusion zone is a tracking continuation zone.
26. The computer readable medium according to claim 22 wherein an entry zone is a tracking origination zone and a tracking continuation zone.
27. The computer readable medium according to claim 22 wherein a black-out zone is not monitored for movement.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU2001240100A AU2001240100A1 (en) | 2000-03-10 | 2001-03-07 | Method and apparatus for video surveillance with defined zones |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18817100P | 2000-03-10 | 2000-03-10 | |
US60/188,171 | 2000-03-10 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2001069931A1 true WO2001069931A1 (en) | 2001-09-20 |
Family
ID=22692026
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2001/007361 WO2001069932A1 (en) | 2000-03-10 | 2001-03-07 | Method and apparatus for object tracking and detection |
PCT/US2001/007228 WO2001069930A1 (en) | 2000-03-10 | 2001-03-07 | Method and apparatus for object surveillance with a movable camera |
PCT/US2001/007330 WO2001069931A1 (en) | 2000-03-10 | 2001-03-07 | Method and apparatus for video surveillance with defined zones |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2001/007361 WO2001069932A1 (en) | 2000-03-10 | 2001-03-07 | Method and apparatus for object tracking and detection |
PCT/US2001/007228 WO2001069930A1 (en) | 2000-03-10 | 2001-03-07 | Method and apparatus for object surveillance with a movable camera |
Country Status (3)
Country | Link |
---|---|
US (3) | US6867799B2 (en) |
AU (3) | AU2001247302A1 (en) |
WO (3) | WO2001069932A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019089441A1 (en) * | 2017-10-30 | 2019-05-09 | Qualcomm Incorporated | Exclusion zone in video analytics |
Families Citing this family (175)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AU2001272980A1 (en) * | 2000-06-23 | 2002-01-08 | Sportvision, Inc. | Gps based tracking system |
US7082209B2 (en) * | 2000-08-31 | 2006-07-25 | Hitachi Kokusai Electric, Inc. | Object detecting method and object detecting apparatus and intruding object monitoring apparatus employing the object detecting method |
US8564661B2 (en) * | 2000-10-24 | 2013-10-22 | Objectvideo, Inc. | Video analytic rule detection system and method |
US7868912B2 (en) | 2000-10-24 | 2011-01-11 | Objectvideo, Inc. | Video surveillance system employing video primitives |
US9892606B2 (en) * | 2001-11-15 | 2018-02-13 | Avigilon Fortress Corporation | Video surveillance system employing video primitives |
US20050162515A1 (en) * | 2000-10-24 | 2005-07-28 | Objectvideo, Inc. | Video surveillance system |
US20050146605A1 (en) * | 2000-10-24 | 2005-07-07 | Lipton Alan J. | Video surveillance system employing video primitives |
US8711217B2 (en) | 2000-10-24 | 2014-04-29 | Objectvideo, Inc. | Video surveillance system employing video primitives |
US7256817B2 (en) * | 2000-10-26 | 2007-08-14 | Fujinon Corporation | Following device |
US20020054211A1 (en) * | 2000-11-06 | 2002-05-09 | Edelson Steven D. | Surveillance video camera enhancement system |
US7102666B2 (en) * | 2001-02-12 | 2006-09-05 | Carnegie Mellon University | System and method for stabilizing rotational images |
US7027083B2 (en) * | 2001-02-12 | 2006-04-11 | Carnegie Mellon University | System and method for servoing on a moving fixation point within a dynamic scene |
US7424175B2 (en) | 2001-03-23 | 2008-09-09 | Objectvideo, Inc. | Video segmentation using statistical pixel modeling |
US20020171742A1 (en) * | 2001-03-30 | 2002-11-21 | Wataru Ito | Method and apparatus for controlling a view field of an image picking-up apparatus and computer program therefor |
US7728870B2 (en) * | 2001-09-06 | 2010-06-01 | Nice Systems Ltd | Advanced quality management and recording solutions for walk-in environments |
JP2003083712A (en) * | 2001-09-10 | 2003-03-19 | Denso Corp | Position recognition method for honeycomb structure |
US6970083B2 (en) * | 2001-10-09 | 2005-11-29 | Objectvideo, Inc. | Video tripwire |
US20030076413A1 (en) * | 2001-10-23 | 2003-04-24 | Takeo Kanade | System and method for obtaining video of multiple moving fixation points within a dynamic scene |
US7650058B1 (en) | 2001-11-08 | 2010-01-19 | Cernium Corporation | Object selective video recording |
US7113616B2 (en) * | 2001-12-05 | 2006-09-26 | Hitachi Kokusai Electric Inc. | Object tracking method and apparatus using template matching |
US7583416B2 (en) * | 2001-12-20 | 2009-09-01 | Eastman Kodak Company | Document scanning system with tethered platen element providing sheet-fed and platen scanning functions |
US7062511B1 (en) | 2001-12-31 | 2006-06-13 | Oracle International Corporation | Method and system for portal web site generation |
JP4383891B2 (en) * | 2002-02-22 | 2009-12-16 | ローベルト ボツシユ ゲゼルシヤフト ミツト ベシユレンクテル ハフツング | Measurement data transmission device |
US7548957B1 (en) | 2002-05-07 | 2009-06-16 | Oracle International Corporation | Method and mechanism for a portal website architecture |
US7277924B1 (en) | 2002-05-07 | 2007-10-02 | Oracle International Corporation | Method and mechanism for a portal website architecture |
DE10222203B4 (en) * | 2002-05-18 | 2004-03-25 | Robert Bosch Gmbh | Image monitoring system |
EP1376502A1 (en) * | 2002-06-10 | 2004-01-02 | Siemens Building Technologies AG | Surveillance system |
US7385626B2 (en) * | 2002-10-21 | 2008-06-10 | Sarnoff Corporation | Method and system for performing surveillance |
US20040100563A1 (en) | 2002-11-27 | 2004-05-27 | Sezai Sablak | Video tracking system and method |
DE602004015173D1 (en) * | 2003-01-21 | 2008-09-04 | Canon Kk | Imaging device and image acquisition system |
US7421727B2 (en) * | 2003-02-14 | 2008-09-02 | Canon Kabushiki Kaisha | Motion detecting system, motion detecting method, motion detecting apparatus, and program for implementing the method |
US20050134685A1 (en) * | 2003-12-22 | 2005-06-23 | Objectvideo, Inc. | Master-slave automated video-based surveillance system |
US20040196369A1 (en) * | 2003-03-07 | 2004-10-07 | Canon Kabushiki Kaisha | Monitoring system |
US8004565B2 (en) * | 2003-06-19 | 2011-08-23 | Nvidia Corporation | System and method for using motion vectors for object tracking |
US7310442B2 (en) | 2003-07-02 | 2007-12-18 | Lockheed Martin Corporation | Scene analysis surveillance system |
GB2404247B (en) * | 2003-07-22 | 2005-07-20 | Hitachi Int Electric Inc | Object tracing method and object tracking apparatus |
US20100013917A1 (en) * | 2003-08-12 | 2010-01-21 | Keith Hanna | Method and system for performing surveillance |
US20050052533A1 (en) * | 2003-09-05 | 2005-03-10 | Hitachi Kokusai Electric Inc. | Object tracking method and object tracking apparatus |
CN100458910C (en) * | 2003-10-28 | 2009-02-04 | 松下电器产业株式会社 | Image display device and image display method |
US20050104958A1 (en) * | 2003-11-13 | 2005-05-19 | Geoffrey Egnal | Active camera video-based surveillance systems and methods |
KR100601933B1 (en) * | 2003-11-18 | 2006-07-14 | 삼성전자주식회사 | Method and apparatus of human detection and privacy protection method and system employing the same |
US7428314B2 (en) * | 2003-12-03 | 2008-09-23 | Safehouse International Inc. | Monitoring an environment |
GB2408881B (en) | 2003-12-03 | 2009-04-01 | Safehouse Internat Inc | Monitoring an environment to produce graphical output data representing events of interest |
US8558892B2 (en) * | 2004-01-20 | 2013-10-15 | Honeywell International Inc. | Object blocking zones to reduce false alarms in video surveillance systems |
US7742077B2 (en) * | 2004-02-19 | 2010-06-22 | Robert Bosch Gmbh | Image stabilization system and method for a video camera |
US7382400B2 (en) * | 2004-02-19 | 2008-06-03 | Robert Bosch Gmbh | Image stabilization system and method for a video camera |
US7643066B2 (en) * | 2004-02-19 | 2010-01-05 | Robert Bosch Gmbh | Method and apparatus for producing frame accurate position data in a PTZ dome camera with open loop control |
US7447331B2 (en) * | 2004-02-24 | 2008-11-04 | International Business Machines Corporation | System and method for generating a viewable video index for low bandwidth applications |
JP2005295004A (en) * | 2004-03-31 | 2005-10-20 | Sanyo Electric Co Ltd | Stereoscopic image processing method and apparatus thereof |
EP1584944B1 (en) * | 2004-04-08 | 2008-07-16 | Festo AG & Co. KG | Imaging device and process for detecting moving objects |
US7542588B2 (en) * | 2004-04-30 | 2009-06-02 | International Business Machines Corporation | System and method for assuring high resolution imaging of distinctive characteristics of a moving object |
US20050270372A1 (en) * | 2004-06-02 | 2005-12-08 | Henninger Paul E Iii | On-screen display and privacy masking apparatus and method |
US8212872B2 (en) * | 2004-06-02 | 2012-07-03 | Robert Bosch Gmbh | Transformable privacy mask for video camera images |
US9210312B2 (en) * | 2004-06-02 | 2015-12-08 | Bosch Security Systems, Inc. | Virtual mask for use in autotracking video camera images |
JP4140567B2 (en) * | 2004-07-14 | 2008-08-27 | 松下電器産業株式会社 | Object tracking device and object tracking method |
US7391907B1 (en) * | 2004-10-01 | 2008-06-24 | Objectvideo, Inc. | Spurious object detection in a video surveillance system |
US20060126737A1 (en) * | 2004-12-15 | 2006-06-15 | International Business Machines Corporation | Method, system and program product for a camera to track an object using motion vector data |
US20060126738A1 (en) * | 2004-12-15 | 2006-06-15 | International Business Machines Corporation | Method, system and program product for a plurality of cameras to track an object using motion vector data |
US8493443B2 (en) * | 2005-01-04 | 2013-07-23 | Hewlett-Packard Development Company, L.P. | Methods and apparatus for location determination and asserting and maintaining privacy |
TW200634674A (en) * | 2005-03-28 | 2006-10-01 | Avermedia Tech Inc | Surveillance system having multi-area motion-detection function |
US11232768B2 (en) | 2005-04-12 | 2022-01-25 | Douglas G. Richardson | Embedding animation in electronic mail, text messages and websites |
IL168210A (en) * | 2005-04-21 | 2012-03-29 | Rafael Advanced Defense Sys | Method for assisting a user to designate a target |
US8964029B2 (en) * | 2005-04-29 | 2015-02-24 | Chubb Protection Corporation | Method and device for consistent region of interest |
WO2007014216A2 (en) | 2005-07-22 | 2007-02-01 | Cernium Corporation | Directed attention digital video recordation |
US20070058717A1 (en) * | 2005-09-09 | 2007-03-15 | Objectvideo, Inc. | Enhanced processing for scanning video |
US20070152157A1 (en) * | 2005-11-04 | 2007-07-05 | Raydon Corporation | Simulation arena entity tracking system |
US8942426B2 (en) * | 2006-03-02 | 2015-01-27 | Michael Bar-Am | On-train rail track monitoring system |
JP2007249722A (en) * | 2006-03-17 | 2007-09-27 | Hitachi Ltd | Object detector |
JP2009533778A (en) | 2006-04-17 | 2009-09-17 | オブジェクトビデオ インコーポレイテッド | Video segmentation using statistical pixel modeling |
US7907750B2 (en) * | 2006-06-12 | 2011-03-15 | Honeywell International Inc. | System and method for autonomous object tracking |
TW200801933A (en) * | 2006-06-22 | 2008-01-01 | Univ Nat Central | Positioning apparatus and method for a remote control camera combined with electronic map |
JP4940820B2 (en) * | 2006-08-09 | 2012-05-30 | パナソニック株式会社 | Network camera |
NZ551762A (en) * | 2006-11-30 | 2008-03-28 | Lincoln Ventures Ltd | Player position validation interface |
US9008354B2 (en) * | 2007-02-07 | 2015-04-14 | The Directv Group, Inc. | Video camera tracking system based on geoposition data feedback |
US8451331B2 (en) * | 2007-02-26 | 2013-05-28 | Christopher L. Hughes | Automotive surveillance system |
US20080219504A1 (en) * | 2007-03-05 | 2008-09-11 | Adams Henry W | Automatic measurement of advertising effectiveness |
TWI355615B (en) * | 2007-05-11 | 2012-01-01 | Ind Tech Res Inst | Moving object detection apparatus and method by us |
US20100157056A1 (en) * | 2007-05-20 | 2010-06-24 | Rafael Advanced Defense Systems Ltd. | Tracking and imaging data fusion |
US8159519B2 (en) * | 2007-05-31 | 2012-04-17 | Eastman Kodak Company | Personal controls for personal video communications |
US8154578B2 (en) * | 2007-05-31 | 2012-04-10 | Eastman Kodak Company | Multi-camera residential communication system |
US8253770B2 (en) * | 2007-05-31 | 2012-08-28 | Eastman Kodak Company | Residential video communication system |
US8154583B2 (en) * | 2007-05-31 | 2012-04-10 | Eastman Kodak Company | Eye gazing imaging for video communications |
US20080306708A1 (en) * | 2007-06-05 | 2008-12-11 | Raydon Corporation | System and method for orientation and location calibration for image sensors |
WO2009006605A2 (en) * | 2007-07-03 | 2009-01-08 | Pivotal Vision, Llc | Motion-validating remote monitoring system |
JP4872834B2 (en) | 2007-07-04 | 2012-02-08 | 株式会社ニコン | Image recognition device, focus adjustment device, and imaging device |
US8587667B2 (en) | 2007-07-09 | 2013-11-19 | International Business Machines Corporation | Beyond field-of-view tracked object positional indicators for television event directors and camera operators |
JP4998156B2 (en) * | 2007-08-30 | 2012-08-15 | ソニー株式会社 | Information presenting system, information presenting apparatus, information presenting method, program, and recording medium recording the program |
US20100114704A1 (en) * | 2007-09-07 | 2010-05-06 | Ryan Steelberg | System and method for brand affinity content distribution and optimization |
KR101187909B1 (en) * | 2007-10-04 | 2012-10-05 | 삼성테크윈 주식회사 | Surveillance camera system |
DE102007052307A1 (en) * | 2007-10-31 | 2009-05-07 | Deutsche Industrie Video System Gmbh | Object's e.g. package, position determining and tracking system, has memory unit storing image information linked with numerical values or vector information of computer unit that analyses camera information |
US8315430B2 (en) * | 2007-11-07 | 2012-11-20 | Viewdle Inc. | Object recognition and database population for video indexing |
US8064641B2 (en) * | 2007-11-07 | 2011-11-22 | Viewdle Inc. | System and method for identifying objects in video |
CN101448142A (en) * | 2007-11-27 | 2009-06-03 | 鸿富锦精密工业(深圳)有限公司 | Image tracking device and image tracking method thereof |
DE102007058959A1 (en) * | 2007-12-07 | 2009-06-10 | Robert Bosch Gmbh | Configuration module for a monitoring system, monitoring system, method for configuring the monitoring system and computer program |
TWI353778B (en) * | 2007-12-21 | 2011-12-01 | Ind Tech Res Inst | Moving object detection apparatus and method |
TWI381719B (en) * | 2008-02-18 | 2013-01-01 | Univ Nat Taiwan | Full-frame video stabilization with a polyline-fitted camcorder path |
US8320615B2 (en) * | 2008-02-27 | 2012-11-27 | Honeywell International Inc. | Systems and methods for recognizing a target from a moving platform |
US8330810B2 (en) * | 2008-02-27 | 2012-12-11 | Honeywell International Inc. | Systems and method for dynamic stabilization of target data detected from a moving platform |
JP5456022B2 (en) * | 2008-04-14 | 2014-03-26 | ジーブイビービー ホールディングス エス.エイ.アール.エル. | Technology that automatically tracks objects |
US8169481B2 (en) * | 2008-05-05 | 2012-05-01 | Panasonic Corporation | System architecture and process for assessing multi-perspective multi-context abnormal behavior |
US8311275B1 (en) * | 2008-06-10 | 2012-11-13 | Mindmancer AB | Selective viewing of a scene |
US9288449B2 (en) * | 2008-08-05 | 2016-03-15 | University Of Florida Research Foundation, Inc. | Systems and methods for maintaining multiple objects within a camera field-of-view |
FR2935214B1 (en) * | 2008-08-20 | 2010-10-08 | Eads Europ Aeronautic Defence | METHOD AND DEVICE FOR REMOTELY CONTROLLING AN INBOARD CAMERA IN A MOBILE STATION |
US8150169B2 (en) * | 2008-09-16 | 2012-04-03 | Viewdle Inc. | System and method for object clustering and identification in video |
US9215467B2 (en) | 2008-11-17 | 2015-12-15 | Checkvideo Llc | Analytics-modulated coding of surveillance video |
KR101627185B1 (en) * | 2009-04-24 | 2016-06-03 | 삼성전자 주식회사 | Control method of image photographing apparatus |
TW201040581A (en) * | 2009-05-06 | 2010-11-16 | J Touch Corp | Digital image capturing device with stereo image display and touch functions |
TWI423146B (en) * | 2009-06-05 | 2014-01-11 | Univ Nat Taiwan Science Tech | Method and system for actively detecting and recognizing placards |
WO2011038465A1 (en) * | 2009-09-30 | 2011-04-07 | National Ict Australia Limited | Object tracking for artificial vision |
JP5570176B2 (en) * | 2009-10-19 | 2014-08-13 | キヤノン株式会社 | Image processing system and information processing method |
US10375287B2 (en) * | 2009-10-21 | 2019-08-06 | Disney Enterprises, Inc. | Object trail-based analysis and control of video |
US20110150272A1 (en) * | 2009-12-21 | 2011-06-23 | Honeywell International Inc. | Systems and methods of tracking object paths |
US20110187895A1 (en) * | 2010-02-03 | 2011-08-04 | Fred Cheng | Intelligent video compacting agent |
US9176152B2 (en) | 2010-05-25 | 2015-11-03 | Arryx, Inc | Methods and apparatuses for detection of positional freedom of particles in biological and chemical analyses and applications in immunodiagnostics |
US9794518B2 (en) | 2010-10-21 | 2017-10-17 | Sensormatic Electronics, LLC | Method and system for converting privacy zone planar images to their corresponding pan/tilt coordinates |
JP5791256B2 (en) * | 2010-10-21 | 2015-10-07 | キヤノン株式会社 | Display control apparatus and display control method |
GB2489675A (en) * | 2011-03-29 | 2012-10-10 | Sony Corp | Generating and viewing video highlights with field of view (FOV) information |
US9565403B1 (en) * | 2011-05-05 | 2017-02-07 | The Boeing Company | Video processing system |
US8790269B2 (en) * | 2011-05-09 | 2014-07-29 | Xerox Corporation | Monitoring respiration with a thermal imaging system |
US8548848B1 (en) * | 2011-06-21 | 2013-10-01 | Google Inc. | Mobile interstitial ads |
US8996234B1 (en) * | 2011-10-11 | 2015-03-31 | Lytx, Inc. | Driver performance determination based on geolocation |
US9298575B2 (en) | 2011-10-12 | 2016-03-29 | Lytx, Inc. | Drive event capturing based on geolocation |
JP6004371B2 (en) * | 2011-11-16 | 2016-10-05 | パナソニックIpマネジメント株式会社 | Imaging device |
US9210300B2 (en) * | 2011-12-19 | 2015-12-08 | Nec Corporation | Time synchronization information computation device for synchronizing a plurality of videos, time synchronization information computation method for synchronizing a plurality of videos and time synchronization information computation program for synchronizing a plurality of videos |
US8704904B2 (en) | 2011-12-23 | 2014-04-22 | H4 Engineering, Inc. | Portable system for high quality video recording |
WO2013116810A1 (en) | 2012-02-03 | 2013-08-08 | H4 Engineering, Inc. | Apparatus and method for securing a portable electronic device |
CA2866102A1 (en) | 2012-03-01 | 2013-09-06 | H4 Engineering, Inc. | Apparatus and method for automatic video recording |
US9723192B1 (en) | 2012-03-02 | 2017-08-01 | H4 Engineering, Inc. | Application dependent video recording device architecture |
WO2013131100A1 (en) | 2012-03-02 | 2013-09-06 | H4 Engineering, Inc. | Multifunction automatic video recording device |
IL219639A (en) | 2012-05-08 | 2016-04-21 | Israel Aerospace Ind Ltd | Remote tracking of objects |
JP6057562B2 (en) * | 2012-06-28 | 2017-01-11 | キヤノン株式会社 | Information processing apparatus and control method thereof |
AU2013286547B2 (en) | 2012-07-06 | 2017-03-09 | H4 Engineering, Inc. | A remotely controlled automatic camera tracking system |
US10049542B2 (en) | 2012-11-04 | 2018-08-14 | Magal Security Systems Ltd. | Self-propelled security system on an adjacent to a fence track |
US11823499B2 (en) * | 2012-11-14 | 2023-11-21 | Golan Weiss | Methods and systems for enrollment and authentication |
US9344683B1 (en) | 2012-11-28 | 2016-05-17 | Lytx, Inc. | Capturing driving risk based on vehicle state and automatic detection of a state of a location |
EP2946361B1 (en) * | 2013-01-15 | 2018-01-03 | Israel Aerospace Industries Ltd. | Remote tracking of objects |
IL224273B (en) | 2013-01-17 | 2018-05-31 | Cohen Yossi | Delay compensation while controlling a remote sensor |
MY168266A (en) | 2013-01-29 | 2018-10-16 | Syusei Co Ltd | Surveillance system |
US8929603B1 (en) * | 2013-03-15 | 2015-01-06 | Puretech Systems, Inc. | Autonomous lock-on target tracking with geospatial-aware PTZ cameras |
CA2906997A1 (en) | 2013-03-15 | 2014-09-18 | Freefly Systems Inc. | Method and system for enabling pointing control of an actively stabilized camera |
US9374532B2 (en) | 2013-03-15 | 2016-06-21 | Google Inc. | Cascaded camera motion estimation, rolling shutter detection, and camera shake detection for video stabilization |
RU2543527C1 (en) * | 2013-08-08 | 2015-03-10 | Открытое акционерное общество "Ракетно-космическая корпорация "Энергия" имени С.П. Королева" | Method of measuring speed of objects from television images thereof |
WO2015102731A2 (en) * | 2013-10-18 | 2015-07-09 | Aerovironment, Inc. | Privacy shield for unmanned aerial systems |
US9684435B2 (en) * | 2014-01-13 | 2017-06-20 | Disney Enterprises, Inc. | Camera selection interface for producing a media presentation |
WO2015142919A1 (en) * | 2014-03-17 | 2015-09-24 | BARNES, Megan, D. | Airplane collision avoidance |
CN104036490B (en) * | 2014-05-13 | 2017-03-29 | 重庆大学 | Foreground segmentation method suitable for mobile communications network transmission |
US10334150B2 (en) * | 2014-05-14 | 2019-06-25 | Hanwha Aerospace Co., Ltd. | Camera system and method of tracking object using the same |
US10070077B2 (en) | 2014-09-26 | 2018-09-04 | Sensormatic Electronics, LLC | System and method for automated camera guard tour operation |
US9900511B2 (en) | 2015-04-10 | 2018-02-20 | Freefly Systems, Inc. | Method, system, and device for controlling a stabilized camera remotely |
US10043307B2 (en) | 2015-04-17 | 2018-08-07 | General Electric Company | Monitoring parking rule violations |
US10872241B2 (en) | 2015-04-17 | 2020-12-22 | Ubicquia Iq Llc | Determining overlap of a parking space by a vehicle |
US20160378268A1 (en) * | 2015-06-23 | 2016-12-29 | Honeywell International Inc. | System and method of smart incident analysis in control system using floor maps |
GB2540129A (en) * | 2015-06-29 | 2017-01-11 | Sony Corp | Apparatus, method and computer program |
US10445885B1 (en) | 2015-10-01 | 2019-10-15 | Intellivision Technologies Corp | Methods and systems for tracking objects in videos and images using a cost matrix |
US10325625B2 (en) | 2015-12-04 | 2019-06-18 | Amazon Technologies, Inc. | Motion detection for A/V recording and communication devices |
US10139281B2 (en) | 2015-12-04 | 2018-11-27 | Amazon Technologies, Inc. | Motion detection for A/V recording and communication devices |
WO2017127842A1 (en) * | 2016-01-21 | 2017-07-27 | Wizr Llc | Cloud platform with multi camera synchronization |
CN107105193B (en) * | 2016-02-23 | 2020-03-20 | 芋头科技(杭州)有限公司 | Robot monitoring system based on human body information |
US10055652B2 (en) * | 2016-03-21 | 2018-08-21 | Ford Global Technologies, Llc | Pedestrian detection and motion prediction with rear-facing camera |
US10650650B2 (en) * | 2016-08-12 | 2020-05-12 | A9.Com, Inc. | Parcel theft deterrence for A/V recording and communication devices |
WO2018081328A1 (en) * | 2016-10-26 | 2018-05-03 | Ring Inc. | Customizable intrusion zones for audio/video recording and communication devices |
US11106274B2 (en) * | 2017-04-10 | 2021-08-31 | Intel Corporation | Adjusting graphics rendering based on facial expression |
EP3419283B1 (en) * | 2017-06-21 | 2022-02-16 | Axis AB | System and method for tracking moving objects in a scene |
US10062414B1 (en) * | 2017-08-22 | 2018-08-28 | Futurewei Technologies, Inc. | Determining a future field of view (FOV) for a particular user viewing a 360 degree video stream in a network |
US10819645B2 (en) | 2017-09-20 | 2020-10-27 | Futurewei Technologies, Inc. | Combined method for data rate and field of view size adaptation for virtual reality and 360 degree video streaming |
US10759051B2 (en) * | 2018-04-23 | 2020-09-01 | General Electric Company | Architecture and methods for robotic mobile manipulation system |
US11399137B2 (en) * | 2018-08-10 | 2022-07-26 | Aurora Flight Sciences Corporation | Object-tracking system |
EP3640903B1 (en) * | 2018-10-18 | 2023-12-27 | IDEMIA Identity & Security Germany AG | Signal dependent video surveillance |
TWI673653B (en) | 2018-11-16 | 2019-10-01 | 財團法人國家實驗研究院 | Moving object detection system and method |
CN111291585B (en) * | 2018-12-06 | 2023-12-08 | 杭州海康威视数字技术股份有限公司 | GPS-based target tracking system, method and device and ball machine |
US20230044707A1 (en) * | 2020-01-14 | 2023-02-09 | Sony Group Corporation | Information processor, information processing method, and program |
US11373511B2 (en) | 2020-09-14 | 2022-06-28 | PureTech Systems Inc. | Alarm processing and classification system and method |
US11610224B2 (en) * | 2021-03-16 | 2023-03-21 | International Business Machines Corporation | Electronic display systems |
DE102021207642A1 (en) | 2021-07-16 | 2023-01-19 | Robert Bosch Gesellschaft mit beschränkter Haftung | Surveillance device with at least two cameras, surveillance method, computer program and storage medium |
DE102022202620A1 (en) | 2022-03-17 | 2023-09-21 | Robert Bosch Gesellschaft mit beschränkter Haftung | Monitoring arrangement for displaying moving objects monitoring area, method for displaying a moving object in a monitoring area, computer program and storage medium |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5387768A (en) * | 1993-09-27 | 1995-02-07 | Otis Elevator Company | Elevator passenger detector and door control system which masks portions of a hall image to determine motion and court passengers |
US5798787A (en) * | 1995-08-11 | 1998-08-25 | Kabushiki Kaisha Toshiba | Method and apparatus for detecting an approaching object within a monitoring zone |
Family Cites Families (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2644844B2 (en) * | 1988-09-20 | 1997-08-25 | 株式会社日立製作所 | Distributed image recognition system |
GB2279198B (en) | 1990-08-02 | 1995-05-10 | Hughes Aircraft Co | Image tracking system and technique |
JP2873338B2 (en) * | 1991-09-17 | 1999-03-24 | 富士通株式会社 | Moving object recognition device |
EP0557007A2 (en) * | 1992-02-15 | 1993-08-25 | Sony Corporation | Picture processing apparatus |
KR0147572B1 (en) * | 1992-10-09 | 1998-09-15 | 김광호 | Method & apparatus for object tracing |
CA2148231C (en) * | 1993-01-29 | 1999-01-12 | Michael Haysom Bianchi | Automatic tracking camera control system |
US5473369A (en) | 1993-02-25 | 1995-12-05 | Sony Corporation | Object tracking apparatus |
EP0701758B1 (en) | 1993-06-04 | 2000-11-02 | Sarnoff Corporation | System and method for electronic image stabilization |
US5739856A (en) * | 1993-10-07 | 1998-04-14 | Nikon Corporation | Photographic subject position predicting apparatus |
EP0660131A1 (en) * | 1993-12-23 | 1995-06-28 | Karl Osen | Camera guidance system |
US5694625A (en) * | 1994-09-07 | 1997-12-02 | Nikon Corporation | Camera provided with an automatic focus adjusting apparatus |
JP3689946B2 (en) * | 1995-10-05 | 2005-08-31 | ソニー株式会社 | Data processing apparatus and method |
JP3693130B2 (en) * | 1996-06-28 | 2005-09-07 | ソニー株式会社 | Control device |
DE69712676T2 (en) * | 1996-07-08 | 2003-01-02 | Hyundai Curitel Inc | Video encoding method |
US5677224A (en) * | 1996-09-03 | 1997-10-14 | Advanced Micro Devices, Inc. | Method of making asymmetrical N-channel and P-channel devices |
JPH1091795A (en) * | 1996-09-12 | 1998-04-10 | Toshiba Corp | Device for detecting mobile object and method therefor |
US5878151A (en) * | 1996-10-31 | 1999-03-02 | Combustion Engineering, Inc. | Moving object tracking |
US6061055A (en) * | 1997-03-21 | 2000-05-09 | Autodesk, Inc. | Method of tracking objects with an imaging device |
US6760061B1 (en) * | 1997-04-14 | 2004-07-06 | Nestor Traffic Systems, Inc. | Traffic sensor |
US6727938B1 (en) * | 1997-04-14 | 2004-04-27 | Robert Bosch Gmbh | Security system with maskable motion detection and camera with an adjustable field of view |
US6445409B1 (en) * | 1997-05-14 | 2002-09-03 | Hitachi Denshi Kabushiki Kaisha | Method of distinguishing a moving object and apparatus of tracking and monitoring a moving object |
US6263088B1 (en) * | 1997-06-19 | 2001-07-17 | Ncr Corporation | System and method for tracking movement of objects in a scene |
US6335985B1 (en) * | 1998-01-07 | 2002-01-01 | Kabushiki Kaisha Toshiba | Object extraction apparatus |
KR100281463B1 (en) * | 1998-03-14 | 2001-02-01 | 전주범 | Sub-data encoding apparatus in object based encoding system |
US6545705B1 (en) * | 1998-04-10 | 2003-04-08 | Lynx System Developers, Inc. | Camera with object recognition/data output |
JP2001169281A (en) * | 1999-12-13 | 2001-06-22 | Matsushita Electric Ind Co Ltd | Device and method for encoding moving image |
US6509926B1 (en) * | 2000-02-17 | 2003-01-21 | Sensormatic Electronics Corporation | Surveillance apparatus for camera surveillance system |
US20020104094A1 (en) * | 2000-12-01 | 2002-08-01 | Bruce Alexander | System and method for processing video data utilizing motion detection and subdivided video fields |
-
2001
- 2001-03-07 WO PCT/US2001/007361 patent/WO2001069932A1/en active Application Filing
- 2001-03-07 US US09/801,483 patent/US6867799B2/en not_active Expired - Lifetime
- 2001-03-07 AU AU2001247302A patent/AU2001247302A1/en not_active Abandoned
- 2001-03-07 AU AU2001240100A patent/AU2001240100A1/en not_active Abandoned
- 2001-03-07 WO PCT/US2001/007228 patent/WO2001069930A1/en active Application Filing
- 2001-03-07 AU AU2001247315A patent/AU2001247315A1/en not_active Abandoned
- 2001-03-07 US US09/801,441 patent/US20020008758A1/en not_active Abandoned
- 2001-03-07 WO PCT/US2001/007330 patent/WO2001069931A1/en active Application Filing
- 2001-03-07 US US09/801,484 patent/US7307652B2/en not_active Expired - Lifetime
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5387768A (en) * | 1993-09-27 | 1995-02-07 | Otis Elevator Company | Elevator passenger detector and door control system which masks portions of a hall image to determine motion and court passengers |
US5798787A (en) * | 1995-08-11 | 1998-08-25 | Kabushiki Kaisha Toshiba | Method and apparatus for detecting an approaching object within a monitoring zone |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019089441A1 (en) * | 2017-10-30 | 2019-05-09 | Qualcomm Incorporated | Exclusion zone in video analytics |
US10878578B2 (en) | 2017-10-30 | 2020-12-29 | Qualcomm Incorporated | Exclusion zone in video analytics |
Also Published As
Publication number | Publication date |
---|---|
US20020030741A1 (en) | 2002-03-14 |
US6867799B2 (en) | 2005-03-15 |
US20020008758A1 (en) | 2002-01-24 |
US20010035907A1 (en) | 2001-11-01 |
AU2001240100A1 (en) | 2001-09-24 |
US7307652B2 (en) | 2007-12-11 |
WO2001069930A1 (en) | 2001-09-20 |
WO2001069932A1 (en) | 2001-09-20 |
AU2001247315A1 (en) | 2001-09-24 |
AU2001247302A1 (en) | 2001-09-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2001069931A1 (en) | Method and apparatus for video surveillance with defined zones | |
KR100885465B1 (en) | Video tracking system, and method and apparatus for selecting a target in an automated video tracking system | |
CN110049206B (en) | Image processing method, image processing apparatus, and computer-readable storage medium | |
US20130208123A1 (en) | Method and System for Collecting Evidence in a Security System | |
EP2954499B1 (en) | Information processing apparatus, information processing method, program, and information processing system | |
US6437819B1 (en) | Automated video person tracking system | |
US7385626B2 (en) | Method and system for performing surveillance | |
KR101533686B1 (en) | Apparatus and method for tracking gaze, recording medium for performing the method | |
RU2670429C1 (en) | Systems and methods of tracking moving objects on video image | |
US20020140814A1 (en) | Method for assisting an automated video tracking system in reaquiring a target | |
US20100013917A1 (en) | Method and system for performing surveillance | |
US20040119848A1 (en) | Method and apparatus for computerized image background analysis | |
US20110169867A1 (en) | Moving object detection, tracking, and displaying systems | |
DE112016002373T5 (en) | TRACKING SUPPORT DEVICE, TRACKING SUPPORT SYSTEM AND TRACKING SUPPORT PROCESS | |
KR20050011698A (en) | Object tracking method and object tracking apparatus | |
US20110002548A1 (en) | Systems and methods of video navigation | |
WO2007067721A2 (en) | System and method for counting people near objects | |
GB2443739A (en) | Detecting image regions of salient motion | |
KR20190046351A (en) | Method and Apparatus for Detecting Intruder | |
JP4309900B2 (en) | Video surveillance system | |
CN110047092B (en) | multi-target real-time tracking method in complex environment | |
KR20180086048A (en) | Camera and imgae processing method thereof | |
KR101832274B1 (en) | System for crime prevention of intelligent type by video photographing and method for acting thereof | |
KR20160093253A (en) | Video based abnormal flow detection method and system | |
JPH05334572A (en) | Room monitoring device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AU BR CA JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
122 | Ep: pct application non-entry in european phase | ||
NENP | Non-entry into the national phase |
Ref country code: JP |