US20110211070A1 - Video Analysis, Archiving and Alerting Methods and Appartus for a Distributed, Modular and Extensible Video Surveillance System - Google Patents
- ️Thu Sep 01 2011
Info
-
Publication number
- US20110211070A1 US20110211070A1 US13/076,656 US201113076656A US2011211070A1 US 20110211070 A1 US20110211070 A1 US 20110211070A1 US 201113076656 A US201113076656 A US 201113076656A US 2011211070 A1 US2011211070 A1 US 2011211070A1 Authority
- US
- United States Prior art keywords
- video
- video surveillance
- track
- information
- database Prior art date
- 2004-10-12 Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 47
- 238000004458 analytical method Methods 0.000 title claims description 121
- 230000000694 effects Effects 0.000 claims description 83
- 230000009471 action Effects 0.000 claims description 11
- 230000009466 transformation Effects 0.000 claims description 11
- 230000008569 process Effects 0.000 claims description 7
- 239000000203 mixture Substances 0.000 claims description 2
- 230000001131 transforming effect Effects 0.000 claims description 2
- 238000009472 formulation Methods 0.000 claims 1
- 238000007726 management method Methods 0.000 description 30
- 238000010586 diagram Methods 0.000 description 18
- 230000002123 temporal effect Effects 0.000 description 15
- 230000008901 benefit Effects 0.000 description 14
- 230000004044 response Effects 0.000 description 13
- 241000282414 Homo sapiens Species 0.000 description 8
- 238000001514 detection method Methods 0.000 description 6
- 230000033001 locomotion Effects 0.000 description 6
- 238000009826 distribution Methods 0.000 description 4
- 238000013507 mapping Methods 0.000 description 4
- 238000012544 monitoring process Methods 0.000 description 4
- 238000013499 data model Methods 0.000 description 3
- 230000037406 food intake Effects 0.000 description 3
- 238000012384 transportation and delivery Methods 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 238000013480 data collection Methods 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000004806 packaging method and process Methods 0.000 description 2
- 241000238876 Acari Species 0.000 description 1
- 206010045178 Tunnel vision Diseases 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 238000013523 data management Methods 0.000 description 1
- 238000013501 data transformation Methods 0.000 description 1
- 238000000354 decomposition reaction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000012552 review Methods 0.000 description 1
- 230000009897 systematic effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 239000002699 waste material Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/23418—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/73—Querying
- G06F16/738—Presentation of query results
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7837—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7847—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content
- G06F16/786—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content using motion, e.g. object motion or camera motion
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19602—Image analysis to detect motion of the intruder, e.g. by frame subtraction
- G08B13/19608—Tracking movement of a target, e.g. by detecting an object predefined as a target, using target direction and or velocity to predict its new position
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19602—Image analysis to detect motion of the intruder, e.g. by frame subtraction
- G08B13/19613—Recognition of a predetermined image pattern or behaviour pattern indicating theft or intrusion
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19654—Details concerning communication with a camera
- G08B13/19656—Network used to communicate with a camera, e.g. WAN, LAN, Internet
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19665—Details related to the storage of video surveillance data
- G08B13/19671—Addition of non-video data, i.e. metadata, to video stream
- G08B13/19673—Addition of time stamp, i.e. time metadata, to video stream
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19678—User interface
- G08B13/19682—Graphic User Interface [GUI] presenting system data to the user, e.g. information on a screen helping a user interacting with an alarm system
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19678—User interface
- G08B13/19691—Signalling events for better perception by user, e.g. indicating alarms by making display brighter, adding text, creating a sound
- G08B13/19693—Signalling events for better perception by user, e.g. indicating alarms by making display brighter, adding text, creating a sound using multiple video sources viewed on a single or compound screen
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/2187—Live feed
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/231—Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion
- H04N21/23109—Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion by placing content in organized collections, e.g. EPG data repository
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/181—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
Definitions
- the present invention generally concerns video surveillance systems and more particularly concerns methods and systems for analyzing, managing and cataloguing of video tracks.
- the present invention also more particularly concerns systems and methods for organizing video tracks and associated event and statistical information in well-indexed databases whereby video tracks containing certain types of, for example, objects; events; and activities, are easily managed and accessed through web services and browsers.
- video surveillance system architectures that adopt a modular approach to video analysis applications and operations.
- those skilled in the art desire a software and middleware framework that accommodates the rapid addition of video analysis applications to a suite of pre-existing video analysis applications.
- Such a framework would make it far easier to tailor video surveillance system assets to evolving threat environments.
- those skilled in the art desire a modular, extensible and distributed video surveillance system architecture that easily accommodates the addition and management of video analysis applications.
- Those skilled in the art also desire a system for analyzing, managing, cataloguing and presenting video tracks.
- those skilled in the art desire a system for assigning meaningful track identifying tags to simplify and ease the cataloguing of such tracks.
- those skilled in the art desire a suite of video analysis applications capable of generating a wide degree of track information from key frames suitable for improving the situational awareness of surveillance analysts, to statistical information summarizing monitored object activities and events.
- the present invention comprises methods and apparatus for performing and managing video surveillance activities.
- a first alternate embodiment of the present invention comprises a video surveillance system comprising: at least one video surveillance system module comprising: a video surveillance camera, wherein the video surveillance camera is remotely programmable; a video encoder to encode video gathered by the video surveillance camera; and a video analysis engine coupled to the video surveillance camera to analyze video gathered by the video surveillance camera and to create data derived from the video; at least one video surveillance system network coupled to the at least one video surveillance system module; at least one database and application system coupled to the at least one video surveillance system network to store video gathered by the video surveillance camera and data created by the video analysis engine, and to receive and process queries for the video and data; a video surveillance system control interface coupled to the at least one video surveillance system network to control the operations of the video surveillance system; and a video surveillance application and control interface coupled to the at least one video surveillance system network for controlling video surveillance activities, the video surveillance application and control interface comprising: a program for controlling video surveillance activities; a display for displaying video gathered by the video surveillance camera and data created by the video analysis
- a second alternate embodiment of the present invention comprises: a signal-bearing medium tangibly embodying a program of machine-readable instructions executable by a digital processing apparatus of a computer system to perform operations comprising management of video surveillance activities using an electronic video surveillance system, the operations comprising: applying a video analysis application tool to video gathered by a video surveillance camera in real time, wherein the video analysis application tool comprises at least one user-specified video analysis criterion, and where application of the video analysis application tool to the video generates a data stream related to the at least one user-specified video analysis criterion; transmitting the data stream to an electronic database and application system; storing information comprising the data stream in the electronic database and application system; selecting at least a portion of the information comprising the data stream for forwarding to a video surveillance analyst performing video surveillance activities, wherein the selection is made based on at least one user-specified information forwarding criterion; and forwarding the selected information to the video surveillance analyst for display on an interface device of the electronic video surveillance system.
- embodiments of the present invention overcome the limitations of the prior art.
- the methods and apparatus of the present invention provide a highly flexible middleware architecture for adding new video analysis applications in order to accommodate the availability of new technology or to respond to new threat environments.
- methods and apparatus of the present invention achieve continuous situational awareness through around-the-clock automated object and event detection; object tracking and object classification.
- prior video surveillance systems the systems often depended on the constant vigilance of a video surveillance analyst, a situation fraught with danger due to the inherent limitations of human beings, for example the inability to remain at a high degree of attentiveness.
- computer-controlled automated systems perform most of the object and event detection, object tracking and object classification, meaning that the lack of attentiveness of a video surveillance analyst will be much less of an issue.
- the methods and apparatus of the present invention heighten the situational awareness of video surveillance analysts by performing video surveillance in a more structured manner.
- video surveillance often consisted of setting up a bank of video cameras and associated video displays and monitoring them. Such a low-level approach did not take into consideration likely threats; instead the video monitors are simply observed for “anything which might happen”.
- the methods and apparatus of the present invention institute highly regimented video surveillance procedures and activities which take into consideration likely threats and desired responses. For example, video analytics are selected based on the threat environment.
- alarm events are pre-instituted, meaning that the potential negative effect of human judgment is reduced.
- the methods and apparatus of the present invention apply video analytics continuously, and sometimes supply the results in real time, thereby providing the video surveillance analyst with a wealth of focused information that simply was not available in prior-art systems.
- the systematic way in which video analytics is performed and the resulting data saved means that information is available both for real-time alarm situations and near- and far-term investigative situations. This is particularly important as video surveillance analysts simply cannot observe patterns which only become visible over many months or years.
- just-occurred-event information is saved continuously and is immediately available in response to simple queries to assist a decision maker in deciding how to respond to a threat situation.
- FIG. 1A depicts a block diagram of a front end of an electronic video surveillance camera system organized in relation to video surveillance cameras comprising the video surveillance camera system and made in accordance with the present invention
- FIG. 1B depicts a block diagram of a front end of a video surveillance camera system organized in relation to programmed views of video surveillance cameras comprising the video surveillance camera system and made in accordance with the present invention
- FIG. 2 depicts the data flow architecture of a distributed video surveillance camera system made in accordance with the present invention
- FIG. 3 depicts the control flow architecture of a distributed video surveillance camera system made in accordance with the present invention
- FIG. 4 depicts the data flow architecture of an alternate distributed video surveillance camera system made in accordance with the present invention
- FIG. 5 depicts an alternate video surveillance system architecture made in accordance with the present invention.
- FIG. 6 depicts the steps of a method for automating video analysis operations and data collection and distribution activities of a distributed video surveillance camera system
- FIGS. 7A and 7B depict video analysis criteria and information forwarding criteria available for automating video analysis and data collection and distribution activities of a distributed video surveillance camera system
- FIG. 8 depicts a schematic representation of a track data model in accordance with the present invention.
- FIG. 9 depicts a track summary document in accordance with one embodiment of the present invention.
- FIG. 10 depicts a track data document in accordance with one embodiment of the present invention.
- FIG. 11 depicts a background image document in accordance with one embodiment of the present invention.
- FIG. 12 depicts a track identification format in accordance with the present invention.
- FIG. 13 depicts a temporal-based hierarchical data structure for organizing video track data in accordance with the present invention
- FIG. 14 depicts a conceptual block diagram illustrating the control flow in establishing the schematic architecture of an electronic database system for use in a distributed video surveillance camera system made in accordance with the present invention
- FIG. 15 depicts the steps of a method in establishing the architecture of an electronic database system for use in a distributed video surveillance camera system operating in accordance with the present invention
- FIG. 16 depicts additional operations available for managing the schematic architecture of an electronic database system for use in a distributed video surveillance camera system operating in accordance with the present invention
- FIG. 17 depicts a schematic representation of the data organization of an electronic database system operating in accordance with the present invention.
- FIG. 18 depicts in schematic representation of data flow from a video analysis engine to an electronic database system in an electronic video surveillance system operating in accordance with one embodiment of the present invention
- FIG. 19 depicts the flow of data from a video analysis engine to an electronic database system that in certain embodiments may occur in real- or near-real-time in a video surveillance camera system operating in accordance with the present invention
- FIG. 20 depicts the flow of data from a video analysis engine to an electronic database system that in certain embodiments may occur in background mode or off-peak hours in a video surveillance camera system operating in accordance with the present invention
- FIG. 21 depicts a conceptual block diagram of the control flow in an event retrieval module incorporated in a distributed video surveillance camera system operating in accordance with the present invention
- FIG. 22 depicts steps of a method for retrieving event information from an electronic database system of a video surveillance camera system operating in accordance with the present invention
- FIG. 23 depicts additional operations available for managing the retrieval of event information from an electronic database system of a video surveillance camera system operating in accordance with the present invention
- FIG. 24 depicts a conceptual block diagram of the control and data flow in a track information retrieval service incorporated in a distributed video surveillance camera system operating in accordance with the present invention
- FIG. 25 depicts the steps of a method for managing track information retrieval in a distributed video surveillance camera system operating in accordance with the present invention
- FIG. 26 depicts additional operations available for managing track information retrieval in a distributed video surveillance camera system operating in accordance with the present invention
- FIG. 27 depicts the control and data flow of a data transformation service incorporated in a distributed video surveillance camera system operating in accordance with the present invention
- FIG. 28 depicts the steps of a method for transforming the format of data retrieved from the electronic database system in a distributed video surveillance camera system operating in accordance with the present invention
- FIG. 29 depicts additional operations available in the transformation service of the present invention.
- FIG. 30 is a block diagram depicting the control and data flow of an instant alert management service incorporated in a video surveillance system operating in accordance with the present invention
- FIG. 31 is a block diagram depicting the track data flow in an instant alert management service operating in accordance with the present invention.
- FIG. 32 depicts the steps of a method performing the instant alert management service of the present invention
- FIG. 33 depicts additional operations available in the instant alert management service of the present invention.
- FIG. 34 depicts the establishment of a speeding alert definition for use by the instant alert management service of the present invention
- FIG. 35 depicts the establishment of a loitering alert definition for use by the instant alert management service of the present invention
- FIG. 36 depicts a browser page comprising an instant alert summary generated by the instant alert management service of the present invention
- FIG. 37 is a block diagram depicting the control and data flow of a summary management service operating in accordance with the present invention.
- FIG. 38 depicts the steps of a method performing the summary management service of the present invention.
- FIG. 39 depicts additional operations available in the summary management service of the present invention.
- FIG. 40 depicts a chart and browser page of keyframes generated by the summary management service of the present invention summarizing human activity
- FIG. 41 depicts a chart and browser page of keyframes generated by the summary management service of the present invention summarizing hourly automobile activity
- FIG. 42 depicts charts generated by the summary management service of the present invention summarizing daily object activity
- FIG. 43 depicts a chart generated by the summary management service of the present invention summarizing arrival and departure times
- FIG. 44-46 are background images depicting object movement tracks generated by the summary management service of the present invention.
- FIG. 47 is a browser page comprised of keyframes associated with human activity generated by the summary management service of the present invention.
- FIG. 48 is a browser page comprised of keyframes associated with automobile activity generated by the summary management service of the present invention.
- FIG. 1A depicts a portion of a video surveillance system front end 100 comprised of a root video surveillance cluster 102 , which is in turn comprised of individual video surveillance camera clusters 112 , 114 .
- Each cluster 112 , 114 is in turn comprised of at least one video surveillance camera; for example, cluster 1 comprises four video surveillance cameras 122 , 124 , 126 , 128 , while cluster 114 comprises one video surveillance camera 130 .
- Individual cameras comprising the videos surveillance camera clusters are preferably highly programmable with programmable pan, tilt, zoom and digital zoom features.
- the individual cameras also have view-programming capability, wherein a view comprised of a specific set of operations or movements selected from at least one of the pan, tilt, zoom or digital zoom options are programmed into a sequence of operations or movements.
- each camera may have multiple programmed views as illustrated by camera 124 in FIG. 1A , which has views 142 , 144 , and 146 , camera 128 which has views 148 and 150 ; and camera 130 which has views 152 and 154 .
- the video corresponding to a particular view 146 gathered by a video surveillance camera 124 is analyzed by a video analysis engine 214 according to one or more user-specified criteria.
- the video analysis 214 engine creates track data 162 , 164 by analyzing the video gathered by the video surveillance cameras, for example, video surveillance camera 124 .
- the video analysis occurs on a specific view 146 of video surveillance camera 124 .
- the track data comprises at least a pointer to video satisfying the user-specified criteria. For example, if the criteria correspond to the event of an automobile entering the view of the camera, the video analysis engine will create track data comprising at least a pointer to each segment of video recording the occurrence of an automobile entering the field of view.
- the embodiment depicted in FIG. 1B is organized in clusters of programmed camera views.
- the video surveillance camera front end 170 comprises a root cluster 171 further comprised of clusters 172 and 173 .
- Each of the clusters 172 , 173 is comprised of programmed camera views selected from different cameras.
- cluster 172 is comprised of a view 1 ( 174 ) selected from camera 1 ; view 3 ( 175 ) selected from a camera 2 ; and a view 2 ( 176 ) selected from a camera 5 ; and cluster 173 is comprised of a view 4 ( 177 ) from a camera 7 and a view 3 ( 178 ) from a camera 9 .
- video analysis operations can be applied to a cluster of views, or to individual views comprising the cluster. Further, data and alarm conditions derived from video analysis applied to a cluster of views can be reported from a cluster-centric perspective.
- This feature illustrates a particular advantage of the present invention in that video surveillance and analysis activities can be easily and continually customized to a particular threat environment.
- a video surveillance analyst is not confronted with a situation as in the case of the prior art where video surveillance options were predetermined at the establishment of the video surveillance system; rather, when practicing the methods and apparatus of the present invention the video surveillance analyst can continually adjust the system to evolving threat situations.
- FIG. 2 depicts data management details and FIG. 3 depicts control management details of a video surveillance system 200 made in accordance with the present invention.
- video surveillance system 200 comprises, in part, multiple video surveillance system modules 210 comprising video cameras 212 ; video analysis engines 214 ; video encoders 216 ; and camera controllers 218 .
- At least one video surveillance system module comprises a video surveillance camera cluster in the embodiment depicted in FIGS. 2 and 3 .
- the video surveillance system modules 210 are a particular advantage of the present invention because they provide a high degree of programmability, modularity, flexibility and extensibility.
- Each video surveillance camera 212 in a video surveillance system module is remotely programmable through camera controller 218 .
- the video analysis engine 214 is remotely programmable, permitting a user to specify video analysis criteria remotely.
- data derived from the analysis of video gathered by the video surveillance cameras 212 is transmitted by the video analysis engine to the electronic database system 220 over a video surveillance system network 215 .
- video surveillance system module is depicted as incorporating both a video analysis engine and video encoder, in other embodiments video analysis engines and video encoders may be distributed in other ways; for example, a video encoder may perform encoding operations for multiple cameras.
- the video surveillance system module further comprises a keyframe generation system to create a keyframe of a specific field of view viewed by the video surveillance camera. Keyframes are generated whenever a user-specified video analysis criterion is satisfied, and generally correspond to a still image of a field of view best capturing an object or event being observed.
- a keyframe may be a headshot of a person entering a facility. The headshot may be created by momentarily zooming the video surveillance camera on the person's face.
- a unitary system network 215 is depicted in FIGS. 2 and 3 , but in various alternate embodiments of the present invention the video surveillance system network can be segmented into various component networks operating in accordance with various protocols. For example, portions of the video surveillance system network can be hard-wired while other portions can be wireless. In addition, different wired and wireless network protocols can be used at various levels of the system.
- the electronic database system 220 comprises a database server 222 for receiving and cataloguing data generated by the video analysis engines 214 ; a video manager for receiving video encoded by video encoders 216 ; and an application server 226 for managing video surveillance applications.
- a database server 222 for receiving and cataloguing data generated by the video analysis engines 214 ; a video manager for receiving video encoded by video encoders 216 ; and an application server 226 for managing video surveillance applications.
- FIGS. 2 and 3 both implementation and management of the video surveillance system 200 , and video surveillance activities are managed through the electronic database system 220 of the video surveillance system 200 . In alternate embodiments, these activities can be managed through separate instrumentalities.
- FIGS. 2 and 3 the database system is depicted in FIGS. 2 and 3 as integrating both a database server 222 and video manager 224 . In alternate embodiments, these can be separated; and in still other embodiments there may be multiple database servers 222 and video managers 224 in a large-scale distributed video surveillance system operating in accordance with the present invention.
- the use of globally-unique identifiers for identifying video surveillance camera clusters; video surveillance cameras; video surveillance camera views and track data greatly reduces database management overhead since there is no need to centrally manage identification activities, and makes it possible to implement a distributed video surveillance system.
- two applications 228 and 230 interact with the video surveillance system 200 through the electronic database system 220 .
- application 228 is a control application for implementing and managing the video surveillance system
- application 230 is a surveillance application for controlling surveillance activities accomplished using the video surveillance system 200 .
- the surveillance application 230 would automatically disseminate video and data gathered by the video surveillance system 200 to video surveillance analysts performing surveillance activities.
- the surveillance application would receive queries of the electronic database system 200 formulated by the video surveillance analysts; locate the data and video satisfying the queries, and display the data and video in browser format 240 as shown in FIGS. 2 and 3 .
- the surveillance application 230 would generate alarm conditions for alerting the video surveillance analysts performing video surveillance activities.
- FIGS. 2 and 3 Particular advantages of the methods and apparatus of the present invention are apparent in FIGS. 2 and 3 .
- video analysis can be performed continuously by the video analysis engines 214 , thereby creating a rich database of object and event data that can be accessed either in near-real-time in aid of, for example, policing activities, or much later in aid of investigative activities.
- FIGS. 2 and 3 it is apparent the simplicity with which a video surveillance system having the architecture depicted in FIGS. 2 and 3 can be extended through the addition of modular video surveillance system modules 210 . This is further eased by the assigning of unique identifiers to video surveillance camera clusters; video surveillance cameras; and video surveillance camera views.
- the assignment of unique identifiers aids not only in the cataloguing and archiving of data and video in the electronic database system 220 , but also in the development of situational awareness by video surveillance analysts whose awareness is increased through the association of video surveillance cameras and views with unique identifiers.
- the unique identifiers assigned to the clusters, cameras and views can also function as web address in a web-based system, greatly simplifying interactions with the electronic video surveillance system. For example, if a video surveillance analyst desired to see a live feed associated with a particular camera view, the analyst need merely enter the address of the camera view using the unique identifier to form a portion of the address to access the camera view.
- FIG. 4 An alternate video surveillance system operating in accordance with the methods and apparatus of the present invention is depicted in FIG. 4 .
- the video analysis engine 214 is programmed with an object detection application 250 to perform object detection; an object tracking application 252 to perform object tracking; an object classification application 254 to perform object classification; and an activity analysis application 256 to perform activity analysis.
- the analysis activities are applied to video gathered from a single video surveillance camera 212 , illustrating that multiple video analysis activities can be simultaneously applied to the video.
- the video encoder 216 comprises multiple video encoding applications, providing a wide range of options for video surveillance analysts using the video surveillance system.
- video encoder 216 comprises an MPEG4 encoder 260 ; a Windows Media encoder 262 ; and a Verint Encoder 264 .
- the database server 222 depicted in FIG. 4 is programmed to perform various alerts in response to data received from the video analysis engine 214 .
- the programming of the database server comprises a speeding alert application 270 to issue speeding alerts; a vehicle alert application 272 to issue vehicle alerts; a loitering application 274 to issue loitering alerts; and another application 276 to issue other alerts.
- alarms can also be implemented locally to be performed by at least one of the video analysis engines 214 .
- the video surveillance system 200 includes at least one data processor 301 coupled to at least one system bus 302 through which the data processor 301 may address a memory sub-system 303 , also referred to herein simply as the memory 303 .
- the memory 303 may include RAM, ROM and fixed and removable disks and/or tape.
- the memory 303 is assumed to store at least one program comprising instructions for causing the data processor 301 to execute methods in accordance with the teachings of the invention.
- the data processor 301 is also coupled through the bus 302 to a user interface, preferably a graphical user interface (“GUI”) 305 that includes a user input device 306 , such as one or more of a keyboard, a mouse, a trackball, a voice recognition interface, as well as a user display device 307 , such as a high resolution graphical CRT display terminal, an LCD display terminal, or any suitable display device.
- GUI graphical user interface
- the data processor 301 may also be coupled through the bus 302 to a network interface 308 that provides bidirectional access to the video surveillance system network 215 , such as an intranet and/or the internet.
- a network interface 308 that provides bidirectional access to the video surveillance system network 215 , such as an intranet and/or the internet.
- both system implementation and management, as well as surveillance activities can be performed through the display 307 and input devices 306 .
- these teachings may be implemented using at least one software program running on a personal computer, a server, a microcomputer, a mainframe computer, a portable computer, an embedded computer, or by any suitable type of programmable data processor 301 operating in combination with video surveillance system management application 228 and video surveillance application 230 residing on the electronic database system 200 .
- a video surveillance system technician can also manage the video surveillance system modules 210 comprising at least one video surveillance camera cluster 211 through the display 305 and input device 306 .
- One of the advantages of the present invention is the ease in which a distributed video surveillance system can be quickly and easily constructed. This is the result of the modular concept embodied in the system, reflected in FIGS. 3-5 , wherein the firmware and software for establishing and managing the video surveillance system are designed to operate with video surveillance cameras having a high degree of programmability and video analysis engines providing a wide range of video analytics.
- the foregoing description generally concerned the architecture of a video surveillance system operating in accordance with the methods and apparatus of the present invention.
- the following description will concern establishment of video surveillance activities—particularly the establishment of automatic video analysis operations—and the handling of data created during video surveillance operations.
- FIG. 6 depicts a block diagram illustrating the steps of a method for establishing video analysis operations and data distribution in a distributed video surveillance system operating in accordance with the present invention.
- the first step 410 comprises selecting a view of a video surveillance camera.
- the video surveillance analyst selects a video analysis criterion to be applied to video gathered by the view of the video surveillance camera.
- the video surveillance analyst selects an information forwarding criterion, related to the video analysis criterion selected at step 420 , which governs when and what categories of data generated by the video analysis engine 214 and, in some instances the database server 220 , will be forwarded to the video surveillance analyst.
- the video analysis engine 214 performs video analysis on video gathered by the video surveillance camera at step 440 , and transmits a data stream created by the video analysis operations to the database server 222 of the electronic database system 220 for storage at step 450 .
- the database server 220 applies the information forwarding criterion selected by the video surveillance analyst to the data stream received from the video analysis engine 214 , and selects information meeting the information forwarding criterion established by the video surveillance analyst. Information meeting the criterion is then forwarded to the video surveillance analyst at step 470 .
- FIGS. 7A and 7B depict various video analysis criteria and information forwarding criteria available for selection by video surveillance analysts performing video surveillance activities aided by a video surveillance system operating in accordance with the present invention.
- available video analysis criteria listed in FIG. 7A are object detection 481 ; object tracking 482 ; object classification 483 ; activity detection 484 ; activity monitoring 485 ; and activity analysis 486 .
- Available information forwarding criteria are depicted in FIG. 7B and comprise a speeding alert 491 ; an accident alert 492 ; a vehicle appearance alert 493 ; a loitering alert 494 ; and a view monitor 495 for forwarding general information derived from monitoring a particular view of a video surveillance camera.
- FIGS. 7A and 7B depict the modularity of the present invention at the video analysis and alert levels.
- a video surveillance analyst can easily reprogram a video surveillance system to accommodate new threat environments by adding new video analysis applications and alert criteria.
- video gathered by a particular programmed view of a video surveillance camera is continuously analyzed by a video analysis engine 214 of the video surveillance system to generate tracks, and the data derived from the analysis is then transmitted to the electronic database system 220 for storage in the database server 222 .
- Each track is comprised of track data and captures the activity of an object inside the field of view of the video surveillance camera.
- the track data comprises at least a pointer to a video segment containing the object, and can further comprise in various embodiments location data, trajectory data, motion data, visual features (for example, color texture and shape) and object class or identity.
- FIG. 8 depicts one possible data model for an electronic video surveillance system operating in accordance with the present invention.
- track data 510 can be comprised of four parts: video media 512 ; background information 514 ; object information 516 and track summary information 518 .
- primitive tracks output from a video analysis engine 214 are represented as XML documents.
- XML documents there are three types of XML documents: track summary documents, track data documents and background image documents.
- FIG. 9 depicts an example of a track summary document 520 ;
- FIG. 10 depicts an example of a track data document 530 ;
- FIG. 11 depicts an example of a background image document 540 .
- the track summary document 520 depicted in FIG. 9 is structured as follows:
- Each track summary document 520 is associated with a particular video surveillance camera view that is identified through use of a globally-unique identifier.
- Track data documents 530 each contain a root tag ⁇ TrackData> . . . ⁇ /TrackData> and are structured as follows:
- the track data document type allows multiple occurrences of ⁇ TrackDataFragment> so that the video analysis engine 214 can send the optimal amount of ⁇ TrackDataFragment> to the database server 222 of the electronic database system 220 in dependence on load conditions and the urgency of real time data delivery.
- Each ⁇ TrackDataFragment> contains all the meta-data of a primitive track.
- the background image XML document 540 contains a root tag ⁇ BGImages> . . . ⁇ /BGImages> and is structured as follows:
- the background image document type allows multiple occurrences of ⁇ BGImage> so that the video analysis engine 214 can send the optimal amount of ⁇ BGImage> to the database server 222 of the electronic database system 220 in dependence on load conditions and the urgency of real time data delivery.
- the track summary 520 , track data 530 and background image 540 XML documents created during video analysis activities of a video analysis engine operating on video gathered by a particular view of a video surveillance camera each contain at least a sub-tag ⁇ ViewID> nnnnn ⁇ /ViewID> to uniquely identify in a global manner the view under analysis by the video analysis engine 214 .
- ⁇ TrackSummary> and ⁇ TrackDataFragment> also contain at least a sub-tag ⁇ TrackID> nnnnn ⁇ /TrackID> to identify the primitive track extracted by the video analysis engine 214 .
- This identifier is globally unique across the electronic database system 220 .
- the electronic database system can use the unique identifier to access all data associated with a primitive track.
- the electronic database system 220 of the present invention can support near-real-time search and alert by the video analysis engine 214 .
- the unique track identifier 550 of a video surveillance system operating in accordance with the present invention is depicted in FIG.
- the distributed creation of unique track identifiers is a particular advantage of the present invention because it eliminates the need for a centralized solution, which can add a significant load to a server.
- the particular track identifier is just one possible embodiment within the scope of the present invention; for example, a separate entry can be added to indicate the particular video analytic applied to the video corresponding to the video track.
- the view identifier 552 can be used to indicate both the unique view being analyzed and the particular video analytic applied to the view.
- certain classes of data are “pushed” to the database server 222 of the electronic database system 220 of the video surveillance system 200 in real time, while other categories of data are saved locally.
- the locally-saved data is then pushed to the database server 222 during off-peak situations when the video surveillance system network is experiencing relatively light traffic.
- a local data storage structure 570 for storing data created by video analysis engines operating in accordance with the present invention is depicted FIG. 13 .
- data files are organized in a temporal manner with the unique view 572 identifier as the root directory for a camera view.
- the next three layers of the structure correspond to the year ( 574 , 576 , 578 ), month ( 580 , 582 , 584 ) and day ( 586 , 588 and 590 ) when a particular track was created.
- the directories can be further subdivided into those corresponding to hours and minutes.
- a particular day directory can contain videos 592 , background images 594 , keyframes 596 , foreground models 598 and XMLS subdirectories 600 .
- the XMLS subdirectory 600 is divided into three additional subdirectories: BGImages 602 , TrackData 604 and TrackSummary 606 .
- file limitations can be instituted which would cause new directories to be created when a file limitation has been exceeded.
- Particular advantages of the present invention associated with a temporal-based data structure are: scalable growth of track data over the long term; easy access to data segments based on temporal information; online and offline archival management based on temporal information; ease of temporal-information-based data package composition and decomposition; and ease of web-based access based on temporal information.
- Another feature of the present invention is that data associated with a particular view of a video surveillance camera for many consecutive time segments can be easily packaged and unpackaged. This can be accomplished by an XML document residing directly inside the ViewID directory structured as follows:
- the foregoing temporal-based database architecture concerned local data storage strategies in video surveillance systems operating in accordance with the present invention.
- the following description concerns details associated with the database architecture for performing near-real-time alerts and for archiving data generated by the video analysis engines 214 of a video surveillance system for later retrieval in response to queries.
- a feature of the present invention comprises methods and apparatus for use in establishing and managing a database architecture for use in a video surveillance system capable of performing, for example, near-real-time alerts based on video analysis activities.
- the control flow associated with a database schema management service is depicted in FIG. 14 .
- the database schema manager 620 receives a database name specification 622 ; the details associated with track data 624 ; and a mapping protocol 626 and creates 628 a database architecture in dependence on this information.
- FIG. 15 A method for creating and managing a database for use in a video surveillance system capable of performing near-real-time alerts based on the results of video analysis activities is depicted in FIG. 15 .
- a unique identifier is selected for the particular database being established; this is a particular advantage of the present invention because it alleviates the need for central database management in a distributed video surveillance system comprising many databases.
- the track data specification is retrieved.
- the track data specification identifies and specifies the categories, classes and file types of data that the database will be receiving from the video analysis engines.
- the categories of track data are mapped to database tables.
- next steps concern another particular advantage of the invention involving the automated forwarding of certain categories of data to video surveillance analysts.
- the information forwarding specification is retrieved.
- the information forwarding specification specifies what data will be forwarded, and when it will be forwarded.
- Applications are then programmed at step 656 to identify information satisfying the information forwarding criteria specified in the information forwarding specification, and at step 657 to forward the data.
- FIG. 16 Other operations available for establishing and managing an electronic database system for use in a distributed video surveillance system are depicted in FIG. 16 .
- the operations comprise: deleting the database 682 ; retrieving a pre-determined database schema 684 ; modifying a pre-existing database schema 686 ; migrating an existing database schema to a new database 688 ; and exporting a database 690 .
- FIG. 17 A schematic depiction of the table architecture 705 of a database server 222 made in accordance with the methods and apparatus of the present invention is depicted in FIG. 17 .
- the XML documents from which the track summary; track data and background information was culled are saved.
- tables 716 , 722 and 726 contain pointers to the original track summary 520 ; track data 530 ; and background image XML documents.
- the database server also contains tables 718 , 720 comprised of pointers to video segments and keyframes, respectively.
- the database server contains raw data 724 created by the video analysis engine 214 . Video surveillance analysts in certain situations may desire to perform searches of data created by the video analysis engines 214 .
- FIGS. 18-19 The ingestion of data created by the video analysis engine 214 into the database server of the electronic database system is depicted in FIGS. 18-19 .
- track summary 520 ; track data 530 ; and background image 540 XML documents are transmitted to the database server 220 where the data contained in the documents is remapped into the tables of the database server 222 according to schema mapping 740 .
- a particular advantage of the present invention is the ability to “push” data generated by the video analysis engines 214 to a video surveillance analyst in real- or near-real-time.
- certain categories of high-priority data are transmitted continuously to the database server of the electronic database system in near-real-time.
- the information received is mapped to the database tables by the schema mapping 740 established in accordance with the methods described with reference to FIGS. 14-16 .
- This data is then available for operation on by, for example, alarm applications depicted in FIG. 4 .
- vehicle velocity data can be transmitted in near-real time continuously to the database server 222 .
- the vehicle speed limit ala in application When a speed limit criterion is exceeded, the vehicle speed limit ala in application generates an alarm condition that is transmitted to the video surveillance analyst.
- the alarm can take many forms in accordance with the present invention; for example, an alarm associated with a keyframe containing the offending vehicle; an alarm associated with a text message identifying the offending vehicle; or an alarm associated with a looping video clip of the vehicle exceeding the speed limit.
- all non-crucial data created by the video analysis engine 214 for a given day can be saved to a jar file for local storage on a hard disk associated with the video analysis engine.
- the jar file is transmitted to the database server as depicted in FIG. 20 .
- the data comprising the jar file is then remapped by schema mapping 740 for storage in the database server 222 .
- the local archiving of data is controlled by a data archiving service manager operating either as part of the video analysis engine or independently.
- the data archiving service manager actively monitors a view result; archives data; and manage data ingestion into the local temporal-based database.
- the service relies on user-specified policy parameters to indicate how data should be archived and/or when an archival ingestion should happen. For example, a video surveillance analyst could specify that only data from today should be kept and the rest of the data should be archived, and at the end of today to send the electronic database system a view jar file of today to ingest.
- the following basic functions can be used to implement the service:
- a particular advantage of the present invention is the rich database of object and activity statistics and track data. This information can be accessed in near-real-time to assist in, for example, policing activities, or much later in time to assist in, for example, investigative activities.
- the event retrieval feature of the present invention automates the retrieval of event and object data in response to automatic triggers and surveillance analyst queries.
- the event retrieval service After receiving a query, the event retrieval service performs a search of the database to retrieve information stored in the database server responsive to the request.
- the data and control flow associated with this activity is illustrated in the block diagram depicted in FIG. 21 .
- the event retrieval manager receives a surveillance analyst queries that generally concern an object or event constraint 781 ; or a time span 782 ; or a region under surveillance 784 and formulates a database query 786 based on the original user query.
- the database server then returns the results at 787 and the track data is forwarded to the video surveillance analyst at step 784 .
- FIG. 22 depicts the steps of a method for performing event and object data retrieval in a video surveillance system operating in accordance with the present invention.
- the event retrieval manager 780 receives a query.
- the event retrieval manager parses the query, and formulates at step 824 a new query in a format expected by the database server 222 .
- step 826 based on the database query, responsive information is retrieved.
- FIG. 23 illustrates other operations that can be performed by the electronic database server. Those operations are retrieve track summary property list 840 ; retrieve track data property list 842 ; retrieve spatial property list; and retrieve temporal property list 846 .
- Various user queries can be formulated by a video surveillance analyst.
- the “DoSummaryQuery” is a request for track summary information, and p 1 , etc. are search parameters relating to, for example, object classes; object characteristics; or event class.
- the “DoDataQuery” is a request for track data information comprising, for example, object trajectory information.
- the “DoSDQuery” is a combination request seeking track summary and track data information.
- Various queries can be posed to the database of video track information. For example, after the event retrieval manager 780 has parsed the user query, and reformulated the query in a format the database server can understand, the event retrieval manager may issue one of the following queries. For example, the following query will retrieve tracks identifying all members of a particular class observed by the video surveillance system:
- the following query finds tracks containing a specific object (for example, a particular automobile) or a specific individual (for example, a particular person):
- the following query finds tracks associated with activity of a specified type (for example, tracks of all people entering a particular building):
- the following query can be very useful to detect loitering when duration is very long and detect noise (caused by, for example, sudden camera movements, swing trees, rain drops) when the duration is very short:
- FIG. 24 depicts the control and data flow associated with another feature of the present invention, the track information retrieval service.
- the track information manager 850 receives a request 852 for information associated with a particular track identifier and issues a database query 856 to retrieve information associated with that track identifier.
- the track information manager receives in return information 858 associated with the track identifier, which may consist of track summary properties; track data properties; and background properties.
- the information may further comprise pointers to keyframes, videos or statistics related to that track identifier.
- the information returned is then transmitted 854 to the video surveillance analyst.
- FIG. 25 depicts the steps of a method for performing track information retrieval in a distributed video surveillance system operating in accordance with the methods and apparatus of the present invention.
- the track information retrieval manager receives a track identifier 860 .
- the track information retrieval manager formulates a database query for retrieving information associated with the track identifier.
- the query is transmitted to the database server 222 .
- the track information retrieval manager receives track information responsive to the query, which is then forwarded to the process or person that requested the information at step 868 .
- FIG. 26 depicts additional operations that may be performed by the track information retrieval service, for example, retrieve track summary information 890 ; retrieve track summary file information 892 ; retrieve track data information 894 ; retrieve track data file information 896 ; retrieve keyframe information 898 ; and retrieve video pointer information 900 .
- FIG. 27 depicts the transformation service of the present invention.
- the database server and video server will be retrieve information responsive to the queries.
- the information will not be in a format suitable for presentation to the video surveillance analyst in its native format in which it is stored in the database server.
- the transformation manager 910 receives information in the format in which it is stored in the electronic database from the track information manager 920 and formats the information for presentation to the video surveillance analyst. This formatting typically comprises several diverse operations to accommodate the various formats stored in the database.
- FIG. 28 depicts the steps of a method associated with the transformation service of the present invention.
- the information is retrieved from the electronic database.
- the transformation application determines whether a custom presentation template has been specified for the retrieved information. If so, the custom template is applied at step 944 . If not, a standard template is applied at step 948 . In either case, the information is then transmitted to the video surveillance analyst at steps 946 or 950 .
- FIG. 29 depicts particular transformation options available in accordance with the present invention.
- the desired transformation operation 952 comprises formatting the data into an XML document. This may occur in situations where the information will be used by another video surveillance system component expecting data formatted according to a particular XML standard. In other embodiments, the information will be displayed to a video surveillance analyst. In such situations, the transformation service will perform an operation 954 to retrieve an HTML template for application to the data prior to display. In particular situations, this may comprise retrieving particular table 956 or report formats 958 . In other situations, this may comprise retrieving the HTML format for keyframe 962 or video presentations 960 . Finally in situations where video is to be played back, a video player will be launched 964 .
- FIG. 30 is a block diagram depicting the control flow associated with the instant alert management service of the present invention.
- An alert is implemented in the following manner.
- a video surveillance analyst programs the alert manager through alert constraints or criteria 1012 and desired action and name of action at 1014 .
- the instant alert manager 1010 interacts with the database server 222 .
- the instant alert manager creates a trigger in the database server at 1018 , and once the trigger is established, receives alert notifications from the database at 1020 .
- the instant alert manager 1010 is also programmed to take system actions 1022 and possibly return track data 1024 and track summary properties 1026 when issuing alert actions at 1016 .
- FIG. 31 is a block diagram depicting the track data flow associated with the instant alert manager service in a video surveillance system operating in accordance with the present invention.
- the instant alert manager 1010 acts upon pre-programmed alerts stored in instant alert library 1030 .
- the pre-programmed alerts comprise alert constraints or criteria 1032 and actions 1034 to be taken upon occurrence of the alert criteria.
- Activities shown flowing into the instant alert manager comprise activity reporting by the database server 222 , and may further comprise pre-programmed triggers. Trigger events need not be specifically programmed in the database server; rather in some embodiments the instant alert manager 1010 may monitor a data stream received from the database server 222 .
- the alert event and track data associated with the alert event are stores in an alert activity archive 1036 .
- the alert event and track data associated with the alert event can be retrieved by the event retrieval manager 780 .
- the instant alert service manager receives information that has been forwarded by the video analysis engine 214 in real- or near-real time. Then at step 1042 , the instant alert service manager retrieves the instant alert alarm criterion which is typically expressed in terms of some data condition, for example, “vehicle speed is greater than x mph”, and compares the data received in real time. If there is a match, an alarm is issued at step 1044 . In an alarm condition, the information that is typically issued in association with the alarm is collected at step 1046 and forwarded to the process or video surveillance analyst to be alerted at step 1048 .
- the instant alert service manager receives information that has been forwarded by the video analysis engine 214 in real- or near-real time. Then at step 1042 , the instant alert service manager retrieves the instant alert alarm criterion which is typically expressed in terms of some data condition, for example, “vehicle speed is greater than x mph”, and compares the data received in real time. If there is a match, an alarm is issued at
- Additional operations that may be performed by an instant alert manager operating in accordance with the present invention are depicted in FIG. 33 and comprise retrieving track summary information to be issued with the alert 1050 ; retrieving track data information to be issued with the alert 1052 ; retrieving spatial property information to be issued with the alert 1054 ; retrieving temporal property information to be issued with the alert 1058 ; retrieving keyframes to be issued with the alert 1058 ; retrieving video to be issued with the alert 1060 ; and deleting the alert 1052 .
- FIGS. 34 and 35 depict definitions of alerts formulated by a video surveillance technician and implemented in the video surveillance system by the instant alert manager 1010 .
- the “create alert” instruction it is comprised of an alert name; an alert condition; and an action to be taken upon occurrence of the alert condition.
- the respective names for the alerts are “speedingDetectorM” and “speedingDetectorW”.
- the respective alert conditions are the same—“velStatMax>40”—meaning that a vehicle has exceeded 40 mile per hour.
- the alert action in 1070 is to e-mail an alert, and in 1072 to update a browser page.
- the respective names for the alerts are “loiteringDetectorM” and “loiteringDetectorW”.
- the respective loitering conditions are the same—“end_time ⁇ start_time>60”—meaning that a person has been loitering in a field of view of a video surveillance camera for longer than sixty minutes.
- the alert action in 1080 is to e-mail an alert, and in 1082 to update a browser page.
- FIG. 36 depicts web-based alerts issued by an instant alert manager operating in accordance with the present invention.
- the alerts were accumulated in a browser page 1090 .
- the individual alerts were added as they occurred to the browser page 1090 .
- the individual alerts comprise a title and keyframe. For example there are four speeding alerts recorded in the browser page 1094 ; 1098 ; 1104 and 1106 ; three large vehicle alerts 1096 ; 1100 and 1102 ; and one loitering alert 1092 .
- alerts are issued accompanied by keyframes.
- alert information may comprise a text message indicating time of occurrence; type of alert; and track identifier.
- a video segment may be replayed.
- all data associated with the alert can be instantly retrieved. For instance, in the embodiment depicted in FIG. 36 , when a video surveillance analyst places a cursor over the keyframe, additional information may be displayed in a bubble message.
- a drop-down menu may appear with various options; for example, to recover track summary or track data information; to retrieve velocity information derived from video analysis; to retrieve other categories of temporal or spatial statistics describing the appearance and activities of an object or person; or to retrieve a video segment capturing the event which caused the alert.
- FIGS. 37-48 depict the activity summary service of the present invention. Another advantage of the present invention is the ability to provide a wide range of summary information summarizing activities under video surveillance and video analysis.
- FIG. 37 is a block diagram depicting the control and data flow associated with the summary information manager 1140 .
- the summary manager 1140 receives queries for summary information subject to various conditions, for example object or activity constraints 1142 ; time span 1144 ; resolution 1146 ; and region 1148 .
- Resolution 1146 refers to time increments and Region refers to a specific portion of a field of view under surveillance.
- the summary queries are generally formulated by video surveillance analysts interested in statistical information that may be derived from track and other categories of information generated by video analysis engines 214 .
- the activity summary manager 1140 relays the constraints 1150 to the event retrieval manager 780 ; this begins the information collection process by retrieving track identifiers satisfying the constraints. The track identifiers are returned at 1152 .
- the activity summary manager can retrieve track information associated with the track identifiers if this additional information is required to create the statistics sought by the video surveillance analyst. After the information necessary to generate the statistics has been retrieved, the activity summary manager generates the statistics and issues a summary at 1154 .
- the activity summary manager interacts with the information transformation manager 910 to format the activity summary information in a manner specified by the video surveillance analyst.
- a method for performing activity summarization is depicted in FIG. 38 and comprises at least the following steps.
- the activity summarization service receives a time period selection from a process or video surveillance analyst desiring activity summary data corresponding to that time period.
- the activity summarization service retrieves the activity data associated with the selected time period from the database server at step 1182 .
- the activity summarization service then generates the desired statistics at step 1184 , and formats the statistics into the desired output format at step 1186 .
- Additional operations that may be performed by the activity summary manager are depicted in FIG. 39 and correspond to retrieving a track summary property list 1210 ; retrieving a track data property list 1212 ; retrieving a spatial property list 1214 ; retrieving a temporal property list 1216 ; retrieving a temporal summary 1218 subject to start time, end time, resolution and region constraints; retrieving a spatial summary 1220 subject to start time, end time, region and resolution constraints; and retrieving a property summary 1222 subject to property, start time, end time and region constraints.
- FIG. 40 depicts browser pages 1232 , 1234 returned in response to a query 1230 for a summary of human activity over the course of day divided by hour.
- the browser page 1232 on the left is a chart depicting the number of persons passing through a field of view of a video surveillance camera on a given day divided by hour.
- the browser page 1234 on the right side is a collection of keyframes that correspond to one of the hour segments from the chart on the left.
- a track is created whenever a specified surveillance criterion is satisfied.
- a track is created whenever a person enters the field of view, and a keyframe of the event is generated as well.
- the track contains track summary information, and the track summary information comprises at least a pointer to the keyframe.
- a video analysis criterion had been specified by a video surveillance analyst to monitor the appearance of persons over the course of a day for a particular camera view.
- a track was created, as well as a keyframe.
- a video surveillance analyst poses a query 1230 as depicted in FIG. 40 , and the summary service acts on the track information generated through application of the specified criterion.
- the keyframes memorializing human activity can be recovered by selecting one of the bars of the bar chart. For example, if the video surveillance analyst selected the first bar, three keyframes would be depicted in the browser on the right. In 1234 , a different time period has been selected.
- FIG. 41 depicts a summary query 1240 and information returned by the video surveillance system in response to the summary query.
- the video surveillance analyst requested information summarizing car activity for a given day on an hour-by-hour basis.
- the browser page 1244 on the right simply shows keyframes of the last nineteen automobiles to be detected by the video surveillance system.
- the browser page 1242 on the left displays a chart containing the desired statistics.
- FIG. 42 depicts two summary queries 1250 , 1260 that request summary information on a day-by-day basis.
- the statistical information returned is depicted in the browser pages 1252 and 1262 .
- FIG. 43 depicts a table 970 summarizing arrival and departure statistics returned in response to a query.
- the information was first retrieved from the database server, and then the transformation service applied the pre-selected table format to the data.
- FIGS. 44-46 Another category of information that may be generated by the activity summary service is illustrated by FIGS. 44-46 .
- FIGS. 44-46 show background images 1280 , 1290 , and 1330 with object tracks indicating movement superimposed. This feature of the present invention is particularly useful for capturing traffic flow patterns and depends on the ability of a video analysis engine operating in accordance with the present invention to recover object trajectory information from a video stream.
- FIG. 47 depicts an HTML document 1310 returned in response to a query for people entering and leaving the building under surveillance.
- HTML document keyframes corresponding to particular video segments are inserted into the document. If the video surveillance analyst desired to watch a video segment corresponding to a keyframe, the video surveillance analyst could select a particular keyframe displayed. The video surveillance system would then retrieve the video segment for playback.
- FIG. 48 depicts a similar HTML document 1320 returned in response to a query for mid-sized cars passing through the field of view observed by the video surveillance camera.
- each keyframe comprising an HTML document an object could be outlined with a color-coded boundary box and contains a color-coded trajectory overlying the original keyframe image.
- the keyframe is the representative frame of the track video clip best depicting an object viewed by a surveillance camera.
- the additional keyframes can be obtained by directing active surveillance cameras to capture the close-up appearance of a tracked object. By clicking a keyframe, an associated video track clip is played back immediately.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Library & Information Science (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computational Linguistics (AREA)
- Closed-Circuit Television Systems (AREA)
Abstract
Systems and methods are disclosed for analyzing, managing and cataloguing video tracks generated by a video surveillance system. Systems and methods operating in accordance with the present invention receive video tracks from a video surveillance system; analyze them for particular categories of, for example, objects or events, assign unique identifying tags to the video tracks; and save them in highly-cross-referenced databases so that video tracks containing, for example, particular objects or events can be easily searched for and recalled. In addition, the present invention provides browsing facilities for viewing events that have just elapsed, events that are occurring in near real-time.
Description
-
CROSS REFERENCE TO RELATED APPLICATIONS
-
This application hereby claims priority from provisional U.S. patent application Ser. Nos. xxx and yyy filed on Oct. 12, 2004 and assigned to the same assignee as the present application. These provisional applications are hereby incorporated by reference in their entirety as if fully restated herein. This application is also related to U.S. patent application Ser. No. vvv filed on even date and assigned to the same assignee as the present application. U.S. patent application Ser. No. vvv is also incorporated by reference in its entirety as if fully restated herein.
TECHNICAL FIELD
-
The present invention generally concerns video surveillance systems and more particularly concerns methods and systems for analyzing, managing and cataloguing of video tracks. The present invention also more particularly concerns systems and methods for organizing video tracks and associated event and statistical information in well-indexed databases whereby video tracks containing certain types of, for example, objects; events; and activities, are easily managed and accessed through web services and browsers.
BACKGROUND
-
In the present environment of heightened concern over criminal and terrorist activities, those skilled in the art have recognized the need for a more widespread application of video surveillance systems. This recognition has resulted in the desire for a more flexible suite of physical video surveillance assets (for example, highly programmable video cameras having the innate ability to handle a multitude of surveillance situations) and firmware for integrating and managing a distributed system of video surveillance assets. U.S. patent application Ser. No. yyy, assigned to the same assignee as the present invention, discloses such methods and systems and is hereby incorporated by reference in its entirety as if fully restated herein.
-
The desire for flexibility also extends to the establishment and management of video surveillance activities. Current systems for establishing and managing video surveillance activities lack flexibility and therefore are difficult to modify to respond to evolving threat environments. In particular, current systems for establishing and managing video surveillance activities were often constructed to respond to a relatively limited number of threats and to perform little or no video analysis. In order to take advantage of the flexibility and programmability of video surveillance cameras and video analysis engines, new video surveillance system architectures are required.
-
Those skilled in the art desire video surveillance system architectures that adopt a modular approach to video analysis applications and operations. For example, those skilled in the art desire a software and middleware framework that accommodates the rapid addition of video analysis applications to a suite of pre-existing video analysis applications. Such a framework would make it far easier to tailor video surveillance system assets to evolving threat environments.
-
In addition, state-of-the-art video surveillance systems, particularly those with large numbers of a video surveillance cameras and video analysis engines, create a great deal of video and data. Obviously, it would be prohibitively expensive and impractical to employ a human user to monitor the output each video camera of such a highly integrated and distributed video surveillance system. In addition, it would thwart one of the objectives of such systems, that is, to give a cadre of surveillance system analysts a holistic view of a particular surveillance environment through highly distributed video analysis operations. Instead, the surveillance system analysts would develop only a tunnel vision view of a particular surveillance environment.
-
Another problem results simply from the number of video cameras that comprise such a distributed and large video surveillance system. Since it would be prohibitively expensive to employ security personnel to monitor each video camera, there must be a system for recording, analyzing and cataloguing the output of the video surveillance system, and for alerting surveillance system analysts in dependence on evolving threats revealed by analysis. Otherwise, the data created by such a system would simply go to waste as so much un-reviewed data.
-
An additional problem arises from the fact that security system analysts in next generation video surveillance systems will not be continuously monitoring a relatively limited number of video feeds giving them the ability to develop a contextual understanding of fields of view as events evolve in the field of view. Rather, due to the large number of available views, it is simply impossible to monitor all of them. As a result, some way must be devised to analyze, manage, catalogue and present video tracks so that the situational awareness of surveillance system analysts is substantially improved over current video surveillance systems.
-
Thus, those skilled in the art desire a modular, extensible and distributed video surveillance system architecture that easily accommodates the addition and management of video analysis applications. Those skilled in the art also desire a system for analyzing, managing, cataloguing and presenting video tracks. In particular, those skilled in the art desire a system for assigning meaningful track identifying tags to simplify and ease the cataloguing of such tracks. In addition, those skilled in the art desire a suite of video analysis applications capable of generating a wide degree of track information from key frames suitable for improving the situational awareness of surveillance analysts, to statistical information summarizing monitored object activities and events.
SUMMARY OF THE PREFERRED EMBODIMENTS
-
The foregoing and other problems are overcome, and other advantages are realized, in accordance with the presently preferred embodiments of these teachings. The present invention comprises methods and apparatus for performing and managing video surveillance activities.
-
In particular, a first alternate embodiment of the present invention comprises a video surveillance system comprising: at least one video surveillance system module comprising: a video surveillance camera, wherein the video surveillance camera is remotely programmable; a video encoder to encode video gathered by the video surveillance camera; and a video analysis engine coupled to the video surveillance camera to analyze video gathered by the video surveillance camera and to create data derived from the video; at least one video surveillance system network coupled to the at least one video surveillance system module; at least one database and application system coupled to the at least one video surveillance system network to store video gathered by the video surveillance camera and data created by the video analysis engine, and to receive and process queries for the video and data; a video surveillance system control interface coupled to the at least one video surveillance system network to control the operations of the video surveillance system; and a video surveillance application and control interface coupled to the at least one video surveillance system network for controlling video surveillance activities, the video surveillance application and control interface comprising: a program for controlling video surveillance activities; a display for displaying video gathered by the video surveillance camera and data created by the video analysis engine; and at least one input device for creating queries of the database and application system.
-
A second alternate embodiment of the present invention comprises: a signal-bearing medium tangibly embodying a program of machine-readable instructions executable by a digital processing apparatus of a computer system to perform operations comprising management of video surveillance activities using an electronic video surveillance system, the operations comprising: applying a video analysis application tool to video gathered by a video surveillance camera in real time, wherein the video analysis application tool comprises at least one user-specified video analysis criterion, and where application of the video analysis application tool to the video generates a data stream related to the at least one user-specified video analysis criterion; transmitting the data stream to an electronic database and application system; storing information comprising the data stream in the electronic database and application system; selecting at least a portion of the information comprising the data stream for forwarding to a video surveillance analyst performing video surveillance activities, wherein the selection is made based on at least one user-specified information forwarding criterion; and forwarding the selected information to the video surveillance analyst for display on an interface device of the electronic video surveillance system.
-
Thus it is seen that embodiments of the present invention overcome the limitations of the prior art. In particular, the methods and apparatus of the present invention provide a highly flexible middleware architecture for adding new video analysis applications in order to accommodate the availability of new technology or to respond to new threat environments.
-
In addition, methods and apparatus of the present invention achieve continuous situational awareness through around-the-clock automated object and event detection; object tracking and object classification. In prior video surveillance systems, the systems often depended on the constant vigilance of a video surveillance analyst, a situation fraught with danger due to the inherent limitations of human beings, for example the inability to remain at a high degree of attentiveness. In contrast to the prior art, in the present invention computer-controlled automated systems perform most of the object and event detection, object tracking and object classification, meaning that the lack of attentiveness of a video surveillance analyst will be much less of an issue.
-
Continuing, the methods and apparatus of the present invention heighten the situational awareness of video surveillance analysts by performing video surveillance in a more structured manner. In the prior art, video surveillance often consisted of setting up a bank of video cameras and associated video displays and monitoring them. Such a low-level approach did not take into consideration likely threats; instead the video monitors are simply observed for “anything which might happen”. In contrast, the methods and apparatus of the present invention institute highly regimented video surveillance procedures and activities which take into consideration likely threats and desired responses. For example, video analytics are selected based on the threat environment. In addition, alarm events are pre-instituted, meaning that the potential negative effect of human judgment is reduced.
-
Further, the methods and apparatus of the present invention apply video analytics continuously, and sometimes supply the results in real time, thereby providing the video surveillance analyst with a wealth of focused information that simply was not available in prior-art systems. The systematic way in which video analytics is performed and the resulting data saved means that information is available both for real-time alarm situations and near- and far-term investigative situations. This is particularly important as video surveillance analysts simply cannot observe patterns which only become visible over many months or years.
-
In addition, the continuous recording and cataloguing of objects and events means that a video surveillance analyst always has available just-occurred-event information. Often the most important information for deciding how to respond to a threat situation is event information that is only minutes or tens of minutes old. In prior art situations where the information simply was not recorded or, if recorded, was not immediately or easily accessible (for example, video tape) the ability to make decisions based on an almost immediate review of just-occurred-event information was either very difficult or impossible. In the present invention, just-occurred-event information is saved continuously and is immediately available in response to simple queries to assist a decision maker in deciding how to respond to a threat situation.
-
In summary, the foregoing alternate embodiments of the present invention are exemplary and non-limiting. For example, one of ordinary skill in the art will understand that one or more aspects or steps from one alternate embodiment can be combined with one or more aspects or steps from another alternate embodiment to create a new embodiment within the scope of the present invention.
BRIEF DESCRIPTION OF THE DRAWINGS
-
The foregoing and other aspects of these teachings are made more evident in the following Detailed Description of the Preferred Embodiments, when read in conjunction with the attached Drawing Figures, wherein:
- FIG. 1A
depicts a block diagram of a front end of an electronic video surveillance camera system organized in relation to video surveillance cameras comprising the video surveillance camera system and made in accordance with the present invention;
- FIG. 1B
depicts a block diagram of a front end of a video surveillance camera system organized in relation to programmed views of video surveillance cameras comprising the video surveillance camera system and made in accordance with the present invention;
- FIG. 2
depicts the data flow architecture of a distributed video surveillance camera system made in accordance with the present invention;
- FIG. 3
depicts the control flow architecture of a distributed video surveillance camera system made in accordance with the present invention;
- FIG. 4
depicts the data flow architecture of an alternate distributed video surveillance camera system made in accordance with the present invention;
- FIG. 5
depicts an alternate video surveillance system architecture made in accordance with the present invention;
- FIG. 6
depicts the steps of a method for automating video analysis operations and data collection and distribution activities of a distributed video surveillance camera system;
- FIGS. 7A and 7B
depict video analysis criteria and information forwarding criteria available for automating video analysis and data collection and distribution activities of a distributed video surveillance camera system;
- FIG. 8
depicts a schematic representation of a track data model in accordance with the present invention;
- FIG. 9
depicts a track summary document in accordance with one embodiment of the present invention;
- FIG. 10
depicts a track data document in accordance with one embodiment of the present invention;
- FIG. 11
depicts a background image document in accordance with one embodiment of the present invention;
- FIG. 12
depicts a track identification format in accordance with the present invention;
- FIG. 13
depicts a temporal-based hierarchical data structure for organizing video track data in accordance with the present invention;
- FIG. 14
depicts a conceptual block diagram illustrating the control flow in establishing the schematic architecture of an electronic database system for use in a distributed video surveillance camera system made in accordance with the present invention;
- FIG. 15
depicts the steps of a method in establishing the architecture of an electronic database system for use in a distributed video surveillance camera system operating in accordance with the present invention;
- FIG. 16
depicts additional operations available for managing the schematic architecture of an electronic database system for use in a distributed video surveillance camera system operating in accordance with the present invention;
- FIG. 17
depicts a schematic representation of the data organization of an electronic database system operating in accordance with the present invention;
- FIG. 18
depicts in schematic representation of data flow from a video analysis engine to an electronic database system in an electronic video surveillance system operating in accordance with one embodiment of the present invention;
- FIG. 19
depicts the flow of data from a video analysis engine to an electronic database system that in certain embodiments may occur in real- or near-real-time in a video surveillance camera system operating in accordance with the present invention;
- FIG. 20
depicts the flow of data from a video analysis engine to an electronic database system that in certain embodiments may occur in background mode or off-peak hours in a video surveillance camera system operating in accordance with the present invention;
- FIG. 21
depicts a conceptual block diagram of the control flow in an event retrieval module incorporated in a distributed video surveillance camera system operating in accordance with the present invention;
- FIG. 22
depicts steps of a method for retrieving event information from an electronic database system of a video surveillance camera system operating in accordance with the present invention;
- FIG. 23
depicts additional operations available for managing the retrieval of event information from an electronic database system of a video surveillance camera system operating in accordance with the present invention;
- FIG. 24
depicts a conceptual block diagram of the control and data flow in a track information retrieval service incorporated in a distributed video surveillance camera system operating in accordance with the present invention;
- FIG. 25
depicts the steps of a method for managing track information retrieval in a distributed video surveillance camera system operating in accordance with the present invention;
- FIG. 26
depicts additional operations available for managing track information retrieval in a distributed video surveillance camera system operating in accordance with the present invention;
- FIG. 27
depicts the control and data flow of a data transformation service incorporated in a distributed video surveillance camera system operating in accordance with the present invention;
- FIG. 28
depicts the steps of a method for transforming the format of data retrieved from the electronic database system in a distributed video surveillance camera system operating in accordance with the present invention;
- FIG. 29
depicts additional operations available in the transformation service of the present invention;
- FIG. 30
is a block diagram depicting the control and data flow of an instant alert management service incorporated in a video surveillance system operating in accordance with the present invention;
- FIG. 31
is a block diagram depicting the track data flow in an instant alert management service operating in accordance with the present invention;
- FIG. 32
depicts the steps of a method performing the instant alert management service of the present invention;
- FIG. 33
depicts additional operations available in the instant alert management service of the present invention;
- FIG. 34
depicts the establishment of a speeding alert definition for use by the instant alert management service of the present invention;
- FIG. 35
depicts the establishment of a loitering alert definition for use by the instant alert management service of the present invention;
- FIG. 36
depicts a browser page comprising an instant alert summary generated by the instant alert management service of the present invention;
- FIG. 37
is a block diagram depicting the control and data flow of a summary management service operating in accordance with the present invention;
- FIG. 38
depicts the steps of a method performing the summary management service of the present invention;
- FIG. 39
depicts additional operations available in the summary management service of the present invention;
- FIG. 40
depicts a chart and browser page of keyframes generated by the summary management service of the present invention summarizing human activity;
- FIG. 41
depicts a chart and browser page of keyframes generated by the summary management service of the present invention summarizing hourly automobile activity;
- FIG. 42
depicts charts generated by the summary management service of the present invention summarizing daily object activity;
- FIG. 43
depicts a chart generated by the summary management service of the present invention summarizing arrival and departure times;
- FIG. 44-46
are background images depicting object movement tracks generated by the summary management service of the present invention;
- FIG. 47
is a browser page comprised of keyframes associated with human activity generated by the summary management service of the present invention; and
- FIG. 48
is a browser page comprised of keyframes associated with automobile activity generated by the summary management service of the present invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
-
Particular embodiments and features of the present invention are illustrated in the schematic block diagrams depicted in
FIGS. 1-5. The block diagram in
FIG. 1Adepicts a portion of a video surveillance system
front end100 comprised of a root
video surveillance cluster102, which is in turn comprised of individual video
surveillance camera clusters112, 114. Each
cluster112, 114 is in turn comprised of at least one video surveillance camera; for example,
cluster1 comprises four
video surveillance cameras122, 124, 126, 128, while
cluster114 comprises one
video surveillance camera130.
-
Individual cameras comprising the videos surveillance camera clusters are preferably highly programmable with programmable pan, tilt, zoom and digital zoom features. The individual cameras also have view-programming capability, wherein a view comprised of a specific set of operations or movements selected from at least one of the pan, tilt, zoom or digital zoom options are programmed into a sequence of operations or movements. In certain embodiments, each camera may have multiple programmed views as illustrated by
camera124 in
FIG. 1A, which has
views142, 144, and 146,
camera128 which has
views148 and 150; and
camera130 which has
views152 and 154.
-
In addition to view programming, the video corresponding to a
particular view146 gathered by a
video surveillance camera124 is analyzed by a
video analysis engine214 according to one or more user-specified criteria. The
video analysis214 engine creates
track data162, 164 by analyzing the video gathered by the video surveillance cameras, for example,
video surveillance camera124. As is shown in
FIG. 1A, the video analysis occurs on a
specific view146 of
video surveillance camera124. The track data comprises at least a pointer to video satisfying the user-specified criteria. For example, if the criteria correspond to the event of an automobile entering the view of the camera, the video analysis engine will create track data comprising at least a pointer to each segment of video recording the occurrence of an automobile entering the field of view.
-
In contrast to the embodiment depicted in
FIG. 1A, the embodiment depicted in
FIG. 1Bis organized in clusters of programmed camera views. As in the case of the embodiment depicted in
FIG. 1A, the video surveillance camera
front end170 comprises a
root cluster171 further comprised of
clusters172 and 173. Each of the
clusters172, 173 is comprised of programmed camera views selected from different cameras. For example,
cluster172 is comprised of a view 1 (174) selected from
camera1; view 3 (175) selected from a
camera2; and a view 2 (176) selected from a
camera5; and
cluster173 is comprised of a view 4 (177) from a
camera7 and a view 3 (178) from a
camera9. In addition, video analysis operations can be applied to a cluster of views, or to individual views comprising the cluster. Further, data and alarm conditions derived from video analysis applied to a cluster of views can be reported from a cluster-centric perspective.
-
This feature illustrates a particular advantage of the present invention in that video surveillance and analysis activities can be easily and continually customized to a particular threat environment. A video surveillance analyst is not confronted with a situation as in the case of the prior art where video surveillance options were predetermined at the establishment of the video surveillance system; rather, when practicing the methods and apparatus of the present invention the video surveillance analyst can continually adjust the system to evolving threat situations.
- FIG. 2
depicts data management details and
FIG. 3depicts control management details of a
video surveillance system200 made in accordance with the present invention. In particular,
video surveillance system200 comprises, in part, multiple video
surveillance system modules210 comprising
video cameras212;
video analysis engines214;
video encoders216; and
camera controllers218. At least one video surveillance system module comprises a video surveillance camera cluster in the embodiment depicted in
FIGS. 2 and 3. The video
surveillance system modules210 are a particular advantage of the present invention because they provide a high degree of programmability, modularity, flexibility and extensibility. Each
video surveillance camera212 in a video surveillance system module is remotely programmable through
camera controller218. In addition, the
video analysis engine214 is remotely programmable, permitting a user to specify video analysis criteria remotely. In the data flow depicted in
FIG. 2, data derived from the analysis of video gathered by the
video surveillance cameras212 is transmitted by the video analysis engine to the
electronic database system220 over a video
surveillance system network215. Although the video surveillance system module is depicted as incorporating both a video analysis engine and video encoder, in other embodiments video analysis engines and video encoders may be distributed in other ways; for example, a video encoder may perform encoding operations for multiple cameras.
-
In other embodiments the video surveillance system module further comprises a keyframe generation system to create a keyframe of a specific field of view viewed by the video surveillance camera. Keyframes are generated whenever a user-specified video analysis criterion is satisfied, and generally correspond to a still image of a field of view best capturing an object or event being observed. In further embodiments, a keyframe may be a headshot of a person entering a facility. The headshot may be created by momentarily zooming the video surveillance camera on the person's face.
-
A
unitary system network215 is depicted in
FIGS. 2 and 3, but in various alternate embodiments of the present invention the video surveillance system network can be segmented into various component networks operating in accordance with various protocols. For example, portions of the video surveillance system network can be hard-wired while other portions can be wireless. In addition, different wired and wireless network protocols can be used at various levels of the system.
-
The
electronic database system220 comprises a
database server222 for receiving and cataloguing data generated by the
video analysis engines214; a video manager for receiving video encoded by
video encoders216; and an
application server226 for managing video surveillance applications. In the particular embodiment depicted in
FIGS. 2 and 3both implementation and management of the
video surveillance system200, and video surveillance activities are managed through the
electronic database system220 of the
video surveillance system200. In alternate embodiments, these activities can be managed through separate instrumentalities.
-
In addition, the database system is depicted in
FIGS. 2 and 3as integrating both a
database server222 and
video manager224. In alternate embodiments, these can be separated; and in still other embodiments there may be
multiple database servers222 and
video managers224 in a large-scale distributed video surveillance system operating in accordance with the present invention. In fact, the use of globally-unique identifiers for identifying video surveillance camera clusters; video surveillance cameras; video surveillance camera views and track data greatly reduces database management overhead since there is no need to centrally manage identification activities, and makes it possible to implement a distributed video surveillance system.
-
In the embodiment depicted in
FIGS. 2 and 3two
applications228 and 230 interact with the
video surveillance system200 through the
electronic database system220. In particular,
application228 is a control application for implementing and managing the video surveillance system, while
application230 is a surveillance application for controlling surveillance activities accomplished using the
video surveillance system200. For example, in various embodiments, the
surveillance application230 would automatically disseminate video and data gathered by the
video surveillance system200 to video surveillance analysts performing surveillance activities. In other embodiments, the surveillance application would receive queries of the
electronic database system200 formulated by the video surveillance analysts; locate the data and video satisfying the queries, and display the data and video in
browser format240 as shown in
FIGS. 2 and 3. In further embodiments, the
surveillance application230 would generate alarm conditions for alerting the video surveillance analysts performing video surveillance activities. These various options can also be simultaneously available in a single embodiment.
-
Particular advantages of the methods and apparatus of the present invention are apparent in
FIGS. 2 and 3. For example, video analysis can be performed continuously by the
video analysis engines214, thereby creating a rich database of object and event data that can be accessed either in near-real-time in aid of, for example, policing activities, or much later in aid of investigative activities. In addition, it is apparent the simplicity with which a video surveillance system having the architecture depicted in
FIGS. 2 and 3can be extended through the addition of modular video
surveillance system modules210. This is further eased by the assigning of unique identifiers to video surveillance camera clusters; video surveillance cameras; and video surveillance camera views. The assignment of unique identifiers aids not only in the cataloguing and archiving of data and video in the
electronic database system220, but also in the development of situational awareness by video surveillance analysts whose awareness is increased through the association of video surveillance cameras and views with unique identifiers.
-
In addition, the unique identifiers assigned to the clusters, cameras and views can also function as web address in a web-based system, greatly simplifying interactions with the electronic video surveillance system. For example, if a video surveillance analyst desired to see a live feed associated with a particular camera view, the analyst need merely enter the address of the camera view using the unique identifier to form a portion of the address to access the camera view.
-
An alternate video surveillance system operating in accordance with the methods and apparatus of the present invention is depicted in
FIG. 4. In
FIG. 4, the
video analysis engine214 is programmed with an
object detection application250 to perform object detection; an
object tracking application252 to perform object tracking; an
object classification application254 to perform object classification; and an
activity analysis application256 to perform activity analysis. In the embodiment depicted in
FIG. 4, the analysis activities are applied to video gathered from a single
video surveillance camera212, illustrating that multiple video analysis activities can be simultaneously applied to the video.
-
The
video encoder216 comprises multiple video encoding applications, providing a wide range of options for video surveillance analysts using the video surveillance system. In particular,
video encoder216 comprises an
MPEG4 encoder260; a
Windows Media encoder262; and a
Verint Encoder264.
-
The
database server222 depicted in
FIG. 4is programmed to perform various alerts in response to data received from the
video analysis engine214. For example, the programming of the database server comprises a speeding
alert application270 to issue speeding alerts; a
vehicle alert application272 to issue vehicle alerts; a
loitering application274 to issue loitering alerts; and another
application276 to issue other alerts. In other embodiments (for example, those depicted in
FIGS. 2-3) alarms can also be implemented locally to be performed by at least one of the
video analysis engines214.
-
Additional details of a
video surveillance system200 made in accordance with the present invention are depicted in the conceptual block diagram depicted in
FIG. 5. The
video surveillance system200 includes at least one
data processor301 coupled to at least one
system bus302 through which the
data processor301 may address a
memory sub-system303, also referred to herein simply as the
memory303. The
memory303 may include RAM, ROM and fixed and removable disks and/or tape. The
memory303 is assumed to store at least one program comprising instructions for causing the
data processor301 to execute methods in accordance with the teachings of the invention. The
data processor301 is also coupled through the
bus302 to a user interface, preferably a graphical user interface (“GUI”) 305 that includes a
user input device306, such as one or more of a keyboard, a mouse, a trackball, a voice recognition interface, as well as a
user display device307, such as a high resolution graphical CRT display terminal, an LCD display terminal, or any suitable display device. With these input/output devices, a user can perform the steps of the methods of the present invention where user-specified values are required.
-
The
data processor301 may also be coupled through the
bus302 to a
network interface308 that provides bidirectional access to the video
surveillance system network215, such as an intranet and/or the internet. In various embodiments of the present invention, both system implementation and management, as well as surveillance activities can be performed through the
display307 and
input devices306.
-
In general, these teachings may be implemented using at least one software program running on a personal computer, a server, a microcomputer, a mainframe computer, a portable computer, an embedded computer, or by any suitable type of
programmable data processor301 operating in combination with video surveillance
system management application228 and
video surveillance application230 residing on the
electronic database system200. A video surveillance system technician can also manage the video
surveillance system modules210 comprising at least one video
surveillance camera cluster211 through the
display305 and
input device306.
-
One of the advantages of the present invention is the ease in which a distributed video surveillance system can be quickly and easily constructed. This is the result of the modular concept embodied in the system, reflected in
FIGS. 3-5, wherein the firmware and software for establishing and managing the video surveillance system are designed to operate with video surveillance cameras having a high degree of programmability and video analysis engines providing a wide range of video analytics.
-
The foregoing description generally concerned the architecture of a video surveillance system operating in accordance with the methods and apparatus of the present invention. The following description will concern establishment of video surveillance activities—particularly the establishment of automatic video analysis operations—and the handling of data created during video surveillance operations.
-
For example,
FIG. 6depicts a block diagram illustrating the steps of a method for establishing video analysis operations and data distribution in a distributed video surveillance system operating in accordance with the present invention. In particular, the
first step410 comprises selecting a view of a video surveillance camera. Next, at
step420 the video surveillance analyst selects a video analysis criterion to be applied to video gathered by the view of the video surveillance camera. Then, at
step430, the video surveillance analyst selects an information forwarding criterion, related to the video analysis criterion selected at
step420, which governs when and what categories of data generated by the
video analysis engine214 and, in some instances the
database server220, will be forwarded to the video surveillance analyst. Then, the
video analysis engine214 performs video analysis on video gathered by the video surveillance camera at
step440, and transmits a data stream created by the video analysis operations to the
database server222 of the
electronic database system220 for storage at
step450. Next, at
step460, the
database server220 applies the information forwarding criterion selected by the video surveillance analyst to the data stream received from the
video analysis engine214, and selects information meeting the information forwarding criterion established by the video surveillance analyst. Information meeting the criterion is then forwarded to the video surveillance analyst at
step470.
- FIGS. 7A and 7B
depict various video analysis criteria and information forwarding criteria available for selection by video surveillance analysts performing video surveillance activities aided by a video surveillance system operating in accordance with the present invention. In particular, available video analysis criteria listed in
FIG. 7Aare
object detection481; object tracking 482; object
classification483;
activity detection484;
activity monitoring485; and
activity analysis486. Available information forwarding criteria are depicted in
FIG. 7Band comprise a speeding
alert491; an
accident alert492; a
vehicle appearance alert493; a loitering
alert494; and a
view monitor495 for forwarding general information derived from monitoring a particular view of a video surveillance camera. Many other video analysis criteria and information forwarding criteria can be selected for use in a video surveillance system operating in accordance with the present invention.
FIGS. 7A and 7Bdepict the modularity of the present invention at the video analysis and alert levels. In particular, a video surveillance analyst can easily reprogram a video surveillance system to accommodate new threat environments by adding new video analysis applications and alert criteria.
-
The immediately preceding description concerned the selection of video analysis activities and general information distribution details of a system operating in accordance with the present invention. The following description will concern details associated with data identification and storage in a distributed video surveillance system operating in accordance with the present invention.
-
In general, in video surveillance systems operating in accordance with the present invention, video gathered by a particular programmed view of a video surveillance camera is continuously analyzed by a
video analysis engine214 of the video surveillance system to generate tracks, and the data derived from the analysis is then transmitted to the
electronic database system220 for storage in the
database server222. Each track is comprised of track data and captures the activity of an object inside the field of view of the video surveillance camera. The track data comprises at least a pointer to a video segment containing the object, and can further comprise in various embodiments location data, trajectory data, motion data, visual features (for example, color texture and shape) and object class or identity. In addition to pointers to video segments, the track data can also comprise digital video clips, keyframes, foreground models and background models.
FIG. 8depicts one possible data model for an electronic video surveillance system operating in accordance with the present invention. In the particular data model depicted in FIG. 8,
track data510 can be comprised of four parts:
video media512;
background information514;
object information516 and track
summary information518.
-
In one alternate embodiment of the present invention, primitive tracks output from a
video analysis engine214 are represented as XML documents. In this particular alternate embodiment there are three types of XML documents: track summary documents, track data documents and background image documents.
FIG. 9depicts an example of a
track summary document520;
FIG. 10depicts an example of a
track data document530; and
FIG. 11depicts an example of a
background image document540. The
track summary document520 depicted in
FIG. 9is structured as follows:
-
<?xml version=“1.0” encoding=“UTF-8”?> <Tracks> <TrackSummary> . . . </TrackSummary> * * * <TrackSummary> . . . </TrackSummary> </Tracks> This document type allows multiple occurrences of <TrackSummary> so that a
video analysis engine214 can send an optimal number of track summary documents to the
database server222 of the electronic database system in dependence on load conditions and the urgency of real time data delivery. Each
track summary document520 is associated with a particular video surveillance camera view that is identified through use of a globally-unique identifier.
-
Track data documents 530 each contain a root tag <TrackData> . . . </TrackData> and are structured as follows:
-
<?xml version=“1.0” encoding=“UTF-8”?> <TrackData> <TrackDataFragment> . . . </TrackDataFragment> * * * <TrackDataFragment> . . . </TrackDataFragment> </TrackData> As in the case of the track summary document, the track data document type allows multiple occurrences of <TrackDataFragment> so that the
video analysis engine214 can send the optimal amount of <TrackDataFragment> to the
database server222 of the
electronic database system220 in dependence on load conditions and the urgency of real time data delivery. Each <TrackDataFragment> contains all the meta-data of a primitive track.
-
The background
image XML document540 contains a root tag <BGImages> . . . </BGImages> and is structured as follows:
-
<?xml version=“1.0” encoding=“UTF-8”?> <BGImages> <BGImage> . . . </BGImage> * * * <BGImage> . . . </BGImage> </BGImages> As in the case of the track summary and track data documents, the background image document type allows multiple occurrences of <BGImage> so that the
video analysis engine214 can send the optimal amount of <BGImage> to the
database server222 of the
electronic database system220 in dependence on load conditions and the urgency of real time data delivery.
-
The
track summary520,
track data530 and
background image540 XML documents created during video analysis activities of a video analysis engine operating on video gathered by a particular view of a video surveillance camera each contain at least a sub-tag <ViewID> nnnnn </ViewID> to uniquely identify in a global manner the view under analysis by the
video analysis engine214.
-
<TrackSummary> and <TrackDataFragment> also contain at least a sub-tag <TrackID> nnnnn </TrackID> to identify the primitive track extracted by the
video analysis engine214. This identifier is globally unique across the
electronic database system220. In response to a query, the electronic database system can use the unique identifier to access all data associated with a primitive track. Combining the unique track identifier with the multiple occurrences XML document design, the
electronic database system220 of the present invention can support near-real-time search and alert by the
video analysis engine214. The
unique track identifier550 of a video surveillance system operating in accordance with the present invention is depicted in
FIG. 12, and is generated by combining the
unique view identifier552, the number of seconds of track creation time in relation to a
standard reference time553, and the
ticks554 of a video analysis engine track counter. A “tick” indicates multiple occurrences of objects, events or activities occurring nearly simultaneously. The tick serves to distinguish between multiple tracks created in such situations. There are
reserved bits551 for a distributed electronic database system. The distributed creation of unique track identifiers is a particular advantage of the present invention because it eliminates the need for a centralized solution, which can add a significant load to a server. The particular track identifier is just one possible embodiment within the scope of the present invention; for example, a separate entry can be added to indicate the particular video analytic applied to the video corresponding to the video track. Alternatively the
view identifier552 can be used to indicate both the unique view being analyzed and the particular video analytic applied to the view.
-
In various embodiments of the present invention, certain classes of data are “pushed” to the
database server222 of the
electronic database system220 of the
video surveillance system200 in real time, while other categories of data are saved locally. The locally-saved data is then pushed to the
database server222 during off-peak situations when the video surveillance system network is experiencing relatively light traffic. A local
data storage structure570 for storing data created by video analysis engines operating in accordance with the present invention is depicted
FIG. 13. In response to the open-ended data growth associated with continuous video analysis activities in some embodiments operating in accordance with the present invention, data files are organized in a temporal manner with the
unique view572 identifier as the root directory for a camera view. The next three layers of the structure correspond to the year (574, 576, 578), month (580, 582, 584) and day (586, 588 and 590) when a particular track was created. In alternate embodiments, the directories can be further subdivided into those corresponding to hours and minutes.
-
In one possible embodiment where the smallest temporal directory corresponds to a particular day, all track data generated during that day for a particular view of a video surveillance camera would be stored in the same directory, for example, /ViewID/Yxxxx/Mxx/Dxx. In various embodiments of the present invention, a particular day directory can contain
videos592,
background images594,
keyframes596,
foreground models598 and
XMLS subdirectories600. Furthermore, the
XMLS subdirectory600 is divided into three additional subdirectories:
BGImages602,
TrackData604 and
TrackSummary606. To prevent access inefficiency for large directories, file limitations can be instituted which would cause new directories to be created when a file limitation has been exceeded.
-
Particular advantages of the present invention associated with a temporal-based data structure are: scalable growth of track data over the long term; easy access to data segments based on temporal information; online and offline archival management based on temporal information; ease of temporal-information-based data package composition and decomposition; and ease of web-based access based on temporal information.
-
Another feature of the present invention is that data associated with a particular view of a video surveillance camera for many consecutive time segments can be easily packaged and unpackaged. This can be accomplished by an XML document residing directly inside the ViewID directory structured as follows:
-
<?xml version=“1.0” encoding=“UTF-8”?> <ViewManifest> <ViewStruct> <ViewID> vvvvvvvv </ViewID> <TemporalUnit>Day</TemporalUnit> <StartTime>yyyy-mm-dd</StartTime> <EndTime>yyyy-mm-dd</EndTime> <MaxDirSize>100</MaxDirSize> <ContainVideos>Yes</ContainVideos> <ContainBGImages>Yes</ContainBGImages> <ContainKeyframes>Yes</ContainKeyframes> <ContainFGModels>Yes</ContainFGModels> <ContainXMLS>Yes</ContainXMLS> <ViewStruct> * * * </ViewStruct> </ViewManifest> This format allows multiple occurrences of <ViewStruct> . . . </ViewStruct>. Each <ViewStruct> provides the track data structure for a continuous time span. The manifest file allows packaging and un-packaging of data files for many consecutive time spans.
-
The foregoing temporal-based database architecture concerned local data storage strategies in video surveillance systems operating in accordance with the present invention. The following description concerns details associated with the database architecture for performing near-real-time alerts and for archiving data generated by the
video analysis engines214 of a video surveillance system for later retrieval in response to queries.
-
A feature of the present invention comprises methods and apparatus for use in establishing and managing a database architecture for use in a video surveillance system capable of performing, for example, near-real-time alerts based on video analysis activities. The control flow associated with a database schema management service is depicted in
FIG. 14. The
database schema manager620 receives a
database name specification622; the details associated with
track data624; and a
mapping protocol626 and creates 628 a database architecture in dependence on this information.
-
A method for creating and managing a database for use in a video surveillance system capable of performing near-real-time alerts based on the results of video analysis activities is depicted in
FIG. 15. In step 652 a unique identifier is selected for the particular database being established; this is a particular advantage of the present invention because it alleviates the need for central database management in a distributed video surveillance system comprising many databases. Then, at
step653 the track data specification is retrieved. The track data specification identifies and specifies the categories, classes and file types of data that the database will be receiving from the video analysis engines. Next, at
step654 the categories of track data are mapped to database tables.
-
The next steps concern another particular advantage of the invention involving the automated forwarding of certain categories of data to video surveillance analysts. At
step655, the information forwarding specification is retrieved. The information forwarding specification specifies what data will be forwarded, and when it will be forwarded. Applications are then programmed at
step656 to identify information satisfying the information forwarding criteria specified in the information forwarding specification, and at
step657 to forward the data.
-
Other operations available for establishing and managing an electronic database system for use in a distributed video surveillance system are depicted in
FIG. 16. The operations comprise: deleting the
database682; retrieving a
pre-determined database schema684; modifying a
pre-existing database schema686; migrating an existing database schema to a
new database688; and exporting a
database690.
-
A schematic depiction of the
table architecture705 of a
database server222 made in accordance with the methods and apparatus of the present invention is depicted in
FIG. 17. As is apparent from the schematic representation, there are tables corresponding to track
summary information710; track
data information712 and
background image information714. In certain embodiments, the XML documents from which the track summary; track data and background information was culled are saved. In these embodiments, tables 716, 722 and 726 contain pointers to the
original track summary520;
track data530; and background image XML documents. In addition, the database server also contains tables 718, 720 comprised of pointers to video segments and keyframes, respectively. Further, the database server contains
raw data724 created by the
video analysis engine214. Video surveillance analysts in certain situations may desire to perform searches of data created by the
video analysis engines214.
-
The ingestion of data created by the
video analysis engine214 into the database server of the electronic database system is depicted in
FIGS. 18-19. As is
apparent track summary520;
track data530; and
background image540 XML documents are transmitted to the
database server220 where the data contained in the documents is remapped into the tables of the
database server222 according to
schema mapping740.
-
A particular advantage of the present invention is the ability to “push” data generated by the
video analysis engines214 to a video surveillance analyst in real- or near-real-time. In
FIG. 19certain categories of high-priority data are transmitted continuously to the database server of the electronic database system in near-real-time. The information received is mapped to the database tables by the
schema mapping740 established in accordance with the methods described with reference to
FIGS. 14-16. This data is then available for operation on by, for example, alarm applications depicted in
FIG. 4. For example, in one particular embodiment, vehicle velocity data can be transmitted in near-real time continuously to the
database server222. When a speed limit criterion is exceeded, the vehicle speed limit ala in application generates an alarm condition that is transmitted to the video surveillance analyst. The alarm can take many forms in accordance with the present invention; for example, an alarm associated with a keyframe containing the offending vehicle; an alarm associated with a text message identifying the offending vehicle; or an alarm associated with a looping video clip of the vehicle exceeding the speed limit.
-
Not all data created by the video analysis engines of a video surveillance system needs to be transmitted to the
database server222 of the
electronic database system220 in real time. In such situations data can be stored locally in accordance with the methods described with respect to
FIG. 13. For example, all non-crucial data created by the
video analysis engine214 for a given day can be saved to a jar file for local storage on a hard disk associated with the video analysis engine. At a pre-selected off-peak time of the day, the jar file is transmitted to the database server as depicted in
FIG. 20. The data comprising the jar file is then remapped by
schema mapping740 for storage in the
database server222.
-
The local archiving of data is controlled by a data archiving service manager operating either as part of the video analysis engine or independently. The data archiving service manager actively monitors a view result; archives data; and manage data ingestion into the local temporal-based database. The service relies on user-specified policy parameters to indicate how data should be archived and/or when an archival ingestion should happen. For example, a video surveillance analyst could specify that only data from today should be kept and the rest of the data should be archived, and at the end of today to send the electronic database system a view jar file of today to ingest. The following basic functions can be used to implement the service:
-
- Jar-File JarView(Root-View-Path)
- Jar-File JarTemporalView(Root-View-Path, BGIMGES|FGIMAGES|VIDEOS|KEYFRAMES|XMLS, Temporal-Path-1, Temporal-Path-2
- Jar-File ListTemporalView(Jar-File, BGIMGES|FGIMAGES|VIDEOS|KEYFRAMES|XMLS, Temporal-Path-1, Temporal-Path-2)
- Jar-File ExtractTemporalView(Jar-File, BGIMGES|FGIMAGES|VIDEOS|KEYFRAMES|XMLS, Temporal-Path-1, Temporal-Path-2)
- Jar-File MergeTemporalView(Jar-File-1, Jar-File-2)
- UnjarTemporalView(Jar-File, BGIMGES|FGIMAGES|VIDEOS|KEYFRAMES|XMLS, Temporal-Path-1, Temporal-Path-2
-
The immediately preceding description concerned data archiving in accordance with the methods and apparatus of the present invention. The next feature to be described concerns event retrieval. A particular advantage of the present invention is the rich database of object and activity statistics and track data. This information can be accessed in near-real-time to assist in, for example, policing activities, or much later in time to assist in, for example, investigative activities. The event retrieval feature of the present invention automates the retrieval of event and object data in response to automatic triggers and surveillance analyst queries.
-
After receiving a query, the event retrieval service performs a search of the database to retrieve information stored in the database server responsive to the request. The data and control flow associated with this activity is illustrated in the block diagram depicted in
FIG. 21. In the data and control flow diagram, the event retrieval manager receives a surveillance analyst queries that generally concern an object or
event constraint781; or a
time span782; or a region under
surveillance784 and formulates a
database query786 based on the original user query. The database server then returns the results at 787 and the track data is forwarded to the video surveillance analyst at
step784.
- FIG. 22
depicts the steps of a method for performing event and object data retrieval in a video surveillance system operating in accordance with the present invention. At
step820, the
event retrieval manager780 receives a query. Then at
step822, the event retrieval manager parses the query, and formulates at step 824 a new query in a format expected by the
database server222. Next, at
step826, based on the database query, responsive information is retrieved.
- FIG. 23
illustrates other operations that can be performed by the electronic database server. Those operations are retrieve track
summary property list840; retrieve track
data property list842; retrieve spatial property list; and retrieve
temporal property list846.
-
Various user queries can be formulated by a video surveillance analyst. For example, the following are three examples of property-based queries that may be created by video surveillance analysts and received by the event retrieval manager at step 820:
-
- DoSummaryQuery(“PCS(p1, p2, p3, . . . , pm)”)
- DoDataQuery(“PCD(p1, p2, p3, . . . , pn)”)
- DoSDQuery(“PCS(p1, p2, p3, . . . pm)”, (“PCD(p1, p2, p3, . . . , pm)”)
-
The “DoSummaryQuery” is a request for track summary information, and p1, etc. are search parameters relating to, for example, object classes; object characteristics; or event class. The “DoDataQuery” is a request for track data information comprising, for example, object trajectory information. The “DoSDQuery” is a combination request seeking track summary and track data information.
-
The following are examples of temporal-based queries:
-
- DoTSummaryQuery(“PCS(p1, p2, p3, . . . , pm)”, “TC(start_time, end_time)”)
- DoTDataQuery(“PCD(p1, p2, p3, . . . , pn)”, “TC(start_time, end_time)”)
- DoTSDQuery(“PCS(p1, p2, p3, . . . , pm)”, “PCD(p1, p2, p3, . . . , pn)”, “TC(start_time, end_time)”)
These requests are formulated similar to the property-based requests except they are further delimited by a time criterion “TC(start_time, end_time)”.
-
The following are example of spatial-based queries:
-
- DoSSummaryQuery(“PCS(p1, p2, p3, . . . , pm)”, “SC(l1, l2, l3, . . . , ll)”)
- DoSDataQuery(“PCD(p1, p2, p3, . . . , pn)”, “SC(l1, l2, l3, . . . , ll)”)
- DoSSDQuery(“PCS(p1, p2, p3, . . . , pm)”, (“PCD(p1, p2, p3, . . . , pn)”, “SC(l1, l2, l3, . . . , ll”)
Spatial-based queries concern specific portions of a field of view of a video surveillance camera and seek information concerning objects entering the specified portion of a field of view, or events or activities occurring in the specified field of view. In embodiments with this capability the video analysis engine 214 creates data concerning where in a field of view an object entered or an event occurred.
-
The following is an example of a spatial-temporal based query:
-
- DoSTSummaryQuery(“PCS(p1, p2, p3, . . . , pm)”, “SC(l1, l2, . . . , ll) “TC(start_time, end_time)”
This request is a combination temporal-spatial request seeking information concerning activities or events occurring in specified portion of a field of view over a specified time period.
- DoSTSummaryQuery(“PCS(p1, p2, p3, . . . , pm)”, “SC(l1, l2, . . . , ll) “TC(start_time, end_time)”
-
Various queries can be posed to the database of video track information. For example, after the
event retrieval manager780 has parsed the user query, and reformulated the query in a format the database server can understand, the event retrieval manager may issue one of the following queries. For example, the following query will retrieve tracks identifying all members of a particular class observed by the video surveillance system:
-
Select Track ID From MILS_Track_Summary_Tab Where IdentityClassID = -
The following query finds tracks containing a specific object (for example, a particular automobile) or a specific individual (for example, a particular person):
-
Select TrackID From MILS_TRACK_SUMMARY_TAB Where Identity = -
The following query finds tracks associated with activity of a specified type (for example, tracks of all people entering a particular building):
-
Select TrackID From MILS_TRACK_SUMMARY_TAB Where ActivityClassID = -
The following query finds tracks containing automobiles and other vehicles traveling faster than a specified speed:
-
Select TrackID From MILS_TRACK_SUMMARY_TAB Where AreaStatMax > -
The following query finds tracks containing automobiles and other vehicles with a minimum speed slower than a given value:
-
Select TrackID From MILS_TRACK_SUMMARY_TAB Where VelStatMin < -
The following query finds tracks containing similar properties:
-
Select TrackID, SFunction(?. FVBlob) From MILS_TRACK_SUMMARY_TAB Order By SFunction(?, FVBlob) ASC -
The following query can be very useful to detect loitering when duration is very long and detect noise (caused by, for example, sudden camera movements, swing trees, rain drops) when the duration is very short:
-
Select TrackID From MILS_TRACK_SUMMARY_TAB Where EndTimestamp-StartTimeStamo < -
This query finds all tracks with centroid passing through a bounding box:
-
Select TrackID From MILS_TRACK_SUMMARY_TAB Where CentroidX > and CentroidX < and CentroidY > and CentroidY < When the track identifiers responsive to any of these example queries are returned, they are forwarded to whatever process or person requested them.
- FIG. 24
depicts the control and data flow associated with another feature of the present invention, the track information retrieval service. As is apparent from the control and data flow, the
track information manager850 receives a
request852 for information associated with a particular track identifier and issues a
database query856 to retrieve information associated with that track identifier. The track information manager receives in
return information858 associated with the track identifier, which may consist of track summary properties; track data properties; and background properties. In addition, the information may further comprise pointers to keyframes, videos or statistics related to that track identifier. The information returned is then transmitted 854 to the video surveillance analyst.
- FIG. 25
depicts the steps of a method for performing track information retrieval in a distributed video surveillance system operating in accordance with the methods and apparatus of the present invention. At
step860, the track information retrieval manager receives a
track identifier860. Then at
step862, the track information retrieval manager formulates a database query for retrieving information associated with the track identifier. Next, at
step864 the query is transmitted to the
database server222. Then, at
step866 the track information retrieval manager receives track information responsive to the query, which is then forwarded to the process or person that requested the information at
step868.
- FIG. 26
depicts additional operations that may be performed by the track information retrieval service, for example, retrieve
track summary information890; retrieve track
summary file information892; retrieve
track data information894; retrieve track data file
information896; retrieve
keyframe information898; and retrieve
video pointer information900.
- FIG. 27
depicts the transformation service of the present invention. In response to queries from video surveillance analysts, the database server and video server will be retrieve information responsive to the queries. Typically the information will not be in a format suitable for presentation to the video surveillance analyst in its native format in which it is stored in the database server. Accordingly, the
transformation manager910 receives information in the format in which it is stored in the electronic database from the
track information manager920 and formats the information for presentation to the video surveillance analyst. This formatting typically comprises several diverse operations to accommodate the various formats stored in the database.
- FIG. 28
depicts the steps of a method associated with the transformation service of the present invention. In the
first step940, the information is retrieved from the electronic database. Then, at
step942 the transformation application determines whether a custom presentation template has been specified for the retrieved information. If so, the custom template is applied at
step944. If not, a standard template is applied at
step948. In either case, the information is then transmitted to the video surveillance analyst at
steps946 or 950.
- FIG. 29
depicts particular transformation options available in accordance with the present invention. For example, in certain embodiments the desired
transformation operation952 comprises formatting the data into an XML document. This may occur in situations where the information will be used by another video surveillance system component expecting data formatted according to a particular XML standard. In other embodiments, the information will be displayed to a video surveillance analyst. In such situations, the transformation service will perform an
operation954 to retrieve an HTML template for application to the data prior to display. In particular situations, this may comprise retrieving particular table 956 or report formats 958. In other situations, this may comprise retrieving the HTML format for
keyframe962 or
video presentations960. Finally in situations where video is to be played back, a video player will be launched 964.
-
Another feature of the present invention is the instant alert management service. As expressed throughout this specification, a particular advantage of a distributed electronic video surveillance system operating in accordance with the present invention is the ability of the system to perform real-time or nearly-real-time alerts based on video analysis performed by the electronic video surveillance system.
FIG. 30is a block diagram depicting the control flow associated with the instant alert management service of the present invention. An alert is implemented in the following manner. A video surveillance analyst programs the alert manager through alert constraints or
criteria1012 and desired action and name of action at 1014. As indicated in the control flow diagram, the
instant alert manager1010 interacts with the
database server222. The instant alert manager creates a trigger in the database server at 1018, and once the trigger is established, receives alert notifications from the database at 1020. The
instant alert manager1010 is also programmed to take
system actions1022 and possibly return
track data1024 and
track summary properties1026 when issuing alert actions at 1016.
- FIG. 31
is a block diagram depicting the track data flow associated with the instant alert manager service in a video surveillance system operating in accordance with the present invention. The
instant alert manager1010 acts upon pre-programmed alerts stored in
instant alert library1030. The pre-programmed alerts comprise alert constraints or
criteria1032 and
actions1034 to be taken upon occurrence of the alert criteria. Activities shown flowing into the instant alert manager comprise activity reporting by the
database server222, and may further comprise pre-programmed triggers. Trigger events need not be specifically programmed in the database server; rather in some embodiments the
instant alert manager1010 may monitor a data stream received from the
database server222. As alerts are issued, the alert event and track data associated with the alert event are stores in an
alert activity archive1036. The alert event and track data associated with the alert event can be retrieved by the
event retrieval manager780.
-
A method in accordance with the instant alert service is depicted in
FIG. 32. At
step1040, the instant alert service manager receives information that has been forwarded by the
video analysis engine214 in real- or near-real time. Then at
step1042, the instant alert service manager retrieves the instant alert alarm criterion which is typically expressed in terms of some data condition, for example, “vehicle speed is greater than x mph”, and compares the data received in real time. If there is a match, an alarm is issued at
step1044. In an alarm condition, the information that is typically issued in association with the alarm is collected at
step1046 and forwarded to the process or video surveillance analyst to be alerted at
step1048.
-
Additional operations that may be performed by an instant alert manager operating in accordance with the present invention are depicted in
FIG. 33and comprise retrieving track summary information to be issued with the alert 1050; retrieving track data information to be issued with the alert 1052; retrieving spatial property information to be issued with the alert 1054; retrieving temporal property information to be issued with the alert 1058; retrieving keyframes to be issued with the alert 1058; retrieving video to be issued with the alert 1060; and deleting the
alert1052.
- FIGS. 34 and 35
depict definitions of alerts formulated by a video surveillance technician and implemented in the video surveillance system by the
instant alert manager1010. As is apparent from consideration of the format of the “create alert” instruction, it is comprised of an alert name; an alert condition; and an action to be taken upon occurrence of the alert condition.
-
In the speeding
alert instructions1070, 1072 depicted in
FIG. 34the respective names for the alerts are “speedingDetectorM” and “speedingDetectorW”. The respective alert conditions are the same—“velStatMax>40”—meaning that a vehicle has exceeded 40 mile per hour. The alert action in 1070 is to e-mail an alert, and in 1072 to update a browser page.
-
In the loitering
alert instructions1080, 1082 depicted in
FIG. 35the respective names for the alerts are “loiteringDetectorM” and “loiteringDetectorW”. The respective loitering conditions are the same—“end_time−start_time>60”—meaning that a person has been loitering in a field of view of a video surveillance camera for longer than sixty minutes. The alert action in 1080 is to e-mail an alert, and in 1082 to update a browser page.
- FIG. 36
depicts web-based alerts issued by an instant alert manager operating in accordance with the present invention. As is apparent, the alerts were accumulated in a
browser page1090. The individual alerts were added as they occurred to the
browser page1090. The individual alerts comprise a title and keyframe. For example there are four speeding alerts recorded in the
browser page1094; 1098; 1104 and 1106; three
large vehicle alerts1096; 1100 and 1102; and one
loitering alert1092.
-
In the embodiment depicted in
FIG. 36alerts are issued accompanied by keyframes. In alternate embodiments, alert information may comprise a text message indicating time of occurrence; type of alert; and track identifier. In other embodiments, a video segment may be replayed. In any case, all data associated with the alert can be instantly retrieved. For instance, in the embodiment depicted in
FIG. 36, when a video surveillance analyst places a cursor over the keyframe, additional information may be displayed in a bubble message. In other embodiments, a drop-down menu may appear with various options; for example, to recover track summary or track data information; to retrieve velocity information derived from video analysis; to retrieve other categories of temporal or spatial statistics describing the appearance and activities of an object or person; or to retrieve a video segment capturing the event which caused the alert.
- FIGS. 37-48
depict the activity summary service of the present invention. Another advantage of the present invention is the ability to provide a wide range of summary information summarizing activities under video surveillance and video analysis.
FIG. 37is a block diagram depicting the control and data flow associated with the
summary information manager1140. As is apparent from
FIG. 37, the
summary manager1140 receives queries for summary information subject to various conditions, for example object or
activity constraints1142;
time span1144;
resolution1146; and
region1148.
Resolution1146 refers to time increments and Region refers to a specific portion of a field of view under surveillance. The summary queries are generally formulated by video surveillance analysts interested in statistical information that may be derived from track and other categories of information generated by
video analysis engines214.
-
Based on queries received from video surveillance analysts, the
activity summary manager1140 relays the
constraints1150 to the
event retrieval manager780; this begins the information collection process by retrieving track identifiers satisfying the constraints. The track identifiers are returned at 1152. In addition, in other embodiments not depicted, the activity summary manager can retrieve track information associated with the track identifiers if this additional information is required to create the statistics sought by the video surveillance analyst. After the information necessary to generate the statistics has been retrieved, the activity summary manager generates the statistics and issues a summary at 1154.
-
In further embodiments, the activity summary manager interacts with the
information transformation manager910 to format the activity summary information in a manner specified by the video surveillance analyst.
-
A method for performing activity summarization is depicted in
FIG. 38and comprises at least the following steps. At
step1180, the activity summarization service receives a time period selection from a process or video surveillance analyst desiring activity summary data corresponding to that time period. The activity summarization service then retrieves the activity data associated with the selected time period from the database server at
step1182. The activity summarization service then generates the desired statistics at
step1184, and formats the statistics into the desired output format at
step1186.
-
Additional operations that may be performed by the activity summary manager are depicted in
FIG. 39and correspond to retrieving a track
summary property list1210; retrieving a track
data property list1212; retrieving a
spatial property list1214; retrieving a
temporal property list1216; retrieving a
temporal summary1218 subject to start time, end time, resolution and region constraints; retrieving a
spatial summary1220 subject to start time, end time, region and resolution constraints; and retrieving a
property summary1222 subject to property, start time, end time and region constraints.
- FIG. 40
depicts
browser pages1232, 1234 returned in response to a
query1230 for a summary of human activity over the course of day divided by hour. The
browser page1232 on the left is a chart depicting the number of persons passing through a field of view of a video surveillance camera on a given day divided by hour. The
browser page1234 on the right side is a collection of keyframes that correspond to one of the hour segments from the chart on the left. In the present embodiment, a track is created whenever a specified surveillance criterion is satisfied. In the particular embodiment depicted in
FIG. 40, a track is created whenever a person enters the field of view, and a keyframe of the event is generated as well. The track contains track summary information, and the track summary information comprises at least a pointer to the keyframe.
-
Accordingly, in the video surveillance system of
FIG. 40, a video analysis criterion had been specified by a video surveillance analyst to monitor the appearance of persons over the course of a day for a particular camera view. As a person entered the field of view, a track was created, as well as a keyframe. In order to recover the information depicted in
FIG. 40, a video surveillance analyst poses a
query1230 as depicted in
FIG. 40, and the summary service acts on the track information generated through application of the specified criterion. The keyframes memorializing human activity can be recovered by selecting one of the bars of the bar chart. For example, if the video surveillance analyst selected the first bar, three keyframes would be depicted in the browser on the right. In 1234, a different time period has been selected.
- FIG. 41
depicts a
summary query1240 and information returned by the video surveillance system in response to the summary query. In this example, the video surveillance analyst requested information summarizing car activity for a given day on an hour-by-hour basis. In contrast to
FIG. 40, where all keyframes were recalled and displayed memorializing the event of an appearance of a human in the field of view of the video surveillance camera for a given hour of a given day, the
browser page1244 on the right simply shows keyframes of the last nineteen automobiles to be detected by the video surveillance system. The
browser page1242 on the left displays a chart containing the desired statistics.
- FIG. 42
depicts two
summary queries1250, 1260 that request summary information on a day-by-day basis. The statistical information returned is depicted in the
browser pages1252 and 1262.
- FIG. 43
depicts a table 970 summarizing arrival and departure statistics returned in response to a query. The information was first retrieved from the database server, and then the transformation service applied the pre-selected table format to the data.
-
Another category of information that may be generated by the activity summary service is illustrated by
FIGS. 44-46.
FIGS. 44-46 show background images1280, 1290, and 1330 with object tracks indicating movement superimposed. This feature of the present invention is particularly useful for capturing traffic flow patterns and depends on the ability of a video analysis engine operating in accordance with the present invention to recover object trajectory information from a video stream.
- FIG. 47
depicts an
HTML document1310 returned in response to a query for people entering and leaving the building under surveillance. In the HTML document, keyframes corresponding to particular video segments are inserted into the document. If the video surveillance analyst desired to watch a video segment corresponding to a keyframe, the video surveillance analyst could select a particular keyframe displayed. The video surveillance system would then retrieve the video segment for playback.
FIG. 48depicts a
similar HTML document1320 returned in response to a query for mid-sized cars passing through the field of view observed by the video surveillance camera.
-
In other embodiments alternate and more sophisticated presentation strategies can be employed. For example, in each keyframe comprising an HTML document an object could be outlined with a color-coded boundary box and contains a color-coded trajectory overlying the original keyframe image. The keyframe is the representative frame of the track video clip best depicting an object viewed by a surveillance camera. Optionally, there can be more than one keyframe for each track or event. The additional keyframes can be obtained by directing active surveillance cameras to capture the close-up appearance of a tracked object. By clicking a keyframe, an associated video track clip is played back immediately.
-
Thus it is seen that the foregoing description has provided by way of exemplary and non-limiting examples a full and informative description of the best method and apparatus presently contemplated by the inventors for implementing a distributed video surveillance system. One skilled in the art will appreciate that the various embodiments described herein can be practiced individually; in combination with one or more other embodiments described herein; or in combination with interactive tabular query interfaces differing from those described herein. Further, one skilled in the art will appreciate that the present invention can be practiced by other than the described embodiments; that these described embodiments are presented for the purposes of illustration and not of limitation; and that the present invention is therefore limited only by the claims which follow.
Claims (21)
129. A video surveillance system comprising:
at least one video surveillance system module comprising:
a video surveillance camera, wherein the video surveillance camera is remotely programmable;
a video encoder to encode video gathered by the video surveillance camera; and
a video analysis engine coupled to the video surveillance camera to analyze video gathered by the video surveillance camera and to create data derived from the video, where at least a portion of the created data is generated in dependence on at least one user-specified video analysis criterion, where the created data comprises at least one pointer to a portion of the video gathered by the video surveillance system satisfying the specified video analysis criterion;
at least one video surveillance system network coupled to the at least one video surveillance system module;
at least one database and application system coupled to the at least one video surveillance system network to store video gathered by the video surveillance camera and data created by the video analysis engine, and to receive and process queries for the video and data;
a video surveillance system control interface coupled to the at least one video surveillance system network to control the operations of the video surveillance system; and
a video surveillance application and control interface coupled to the at least one video surveillance system network for controlling video surveillance activities, the video surveillance application and control interface comprising:
a program for controlling video surveillance activities;
a display for displaying video gathered by the video surveillance camera and data created by the video analysis engine; and
at least one input device for creating queries of the database and application system.
130. The video surveillance system of
claim 129, wherein:
the video analysis engine further comprises:
a track identification system for assigning a unique identifier to the at least one track.
131. The video surveillance system of
claim 130, wherein the track further comprises track summary information.
132. The video surveillance system of
claim 131, where the track identification system applies the unique identifier assigned to the at least one track to the track summary information associated with the at least one track.
133. The video surveillance system of
claim 131, wherein the track summary information further comprises at least one of timestamp information memorializing a creation time of the track, object class information specifying an object class, object area information, object velocity information, activity class information, activity class information corresponding to loitering activity, a pointer to a video segment and a pointer to a keyframe.
134. The video surveillance system of
claim 130, wherein the at least one database and application system further comprises:
a database server to save the at least one track using the unique identifier.
135. The video surveillance system of
claim 134, where the video surveillance application and control interface further comprises: a track query formulation application to formulate queries to locate tracks saved in the database server satisfying a user-specified criterion, where the database and application system further comprises: an event retrieval system to accept user-specified track location queries and to locate tracks satisfying the criterion contained in the user-specified queries; and a track identifier forwarding system to forward track identifiers satisfying the user-specified query to the user requesting the track identifiers.
136. The video surveillance system of
claim 135, where the video surveillance application and control interface further comprises: a track information request application to request track information associated with forwarded track identifiers, where the at least one database and application system further comprises: a track information retrieval application to retrieve track information associated with track identifiers; and a transformation application for transforming the track information into a desired output format.
137. The video surveillance system of
claim 130, where the video surveillance system further comprises:
a local storage database associated with the at least one video surveillance system module and coupled to the at least one video surveillance system network for storing data created by the video analysis engine.
138. The video surveillance system of
claim 130, wherein the video surveillance system module further comprises:
a keyframe generation system to create a keyframe of a specific field of view viewed by the video surveillance camera satisfying the specified video analysis criterion.
139. The video surveillance system of
claim 138, where the track further comprises a pointer to the keyframe, wherein the keyframe generation system performs the following operation: generating a keyframe satisfying the at least one specified video analysis criterion, wherein the video analysis engine performs the following operation: creating tracks in dependence on the at least one specified video analysis criterion, wherein the at least one database and application system performs the following operation: saving the keyframe and track created when the specified criterion is satisfied.
140. The video surveillance system of
claim 139, wherein the video surveillance application and control interface performs the following operations: receiving a query for tracks corresponding to the specified criterion; and transmitting the query for the tracks to the at least one database and application system, wherein the at least one database and application system performs the following operations: receiving the query for tracks corresponding to the specified criterion; and locating tracks corresponding to the specified criterion, wherein the at least one database and application system performs the following operations: locating the keyframe corresponding to the specified criterion by using the tracks; and transmitting the keyframe satisfying the specified criterion to the video surveillance application and control interface.
141. The video surveillance system of
claim 130, wherein the video analysis criterion corresponds to at least one of an object appearing in the video gathered by the video surveillance camera, an event appearing in video gathered by the video surveillance camera and an activity appearing in video gathered by the video surveillance camera.
142. The video surveillance system of
claim 129, where the at least one database and application system further comprises:
an alert system, wherein the alert system further comprises:
an alert criterion specification system to specify an alert based on data created by the video analysis engine; and
an alert action specification system to specify alert actions to be taken when the alert criterion is satisfied.
143. The video surveillance system of
claim 142, where the alert criterion concerns at least one of vehicle speed and loitering activity.
144. The video surveillance system of
claim 142, where the alert action comprises at least one of an e-mail issued to a video surveillance analyst and a browser page comprising alert information, where the alert information comprises at least one of a keyframe, track identifiers and a video segment capturing activity that satisfied the alert criterion.
145. The video surveillance system of
claim 129, where:
the video surveillance camera has a multiple programmable view feature, the video surveillance camera further comprising:
a view programming system for programming views of the video surveillance camera; and
the video surveillance system control interface further comprises:
a video surveillance camera view programming system to program views of the video surveillance camera by interacting with the view programming system of the video surveillance camera.
146. The video surveillance system of
claim 145, where the video analysis engine further comprises:
a video surveillance camera view selection system to select a particular view of the video surveillance camera having the multiple programmable view feature and to analyze video associated with the view; and
a track identification system for assigning a unique identifier to the track created by the video analysis engine, where the unique identifier references at least the selected view of the video surveillance camera.
147. The video surveillance system of
claim 129, wherein the at least one database and application system performs the following operation:
saving the at least one track generated when the at least one user-specified video analysis criterion is satisfied.
148. The video surveillance system of
claim 147, wherein the video surveillance system application and control interface performs the following operations: receiving a query for tracks corresponding to the at least one user-specified video analysis criterion, and transmitting the query for the tracks to the at least one database and application system; wherein the at least one database and application system performs the following operations: receiving the query for the tracks corresponding to the specified criterion, and locating tracks corresponding to the at least one user-specified video analysis criterion.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/076,656 US20110211070A1 (en) | 2004-10-12 | 2011-03-31 | Video Analysis, Archiving and Alerting Methods and Appartus for a Distributed, Modular and Extensible Video Surveillance System |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US61834204P | 2004-10-12 | 2004-10-12 | |
US61809504P | 2004-10-12 | 2004-10-12 | |
US10/967,048 US7746378B2 (en) | 2004-10-12 | 2004-10-14 | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
US12/462,926 US20090322881A1 (en) | 2004-10-12 | 2009-08-11 | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
US13/076,656 US20110211070A1 (en) | 2004-10-12 | 2011-03-31 | Video Analysis, Archiving and Alerting Methods and Appartus for a Distributed, Modular and Extensible Video Surveillance System |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/462,926 Continuation US20090322881A1 (en) | 2004-10-12 | 2009-08-11 | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
Publications (1)
Publication Number | Publication Date |
---|---|
US20110211070A1 true US20110211070A1 (en) | 2011-09-01 |
Family
ID=36145291
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/967,048 Expired - Fee Related US7746378B2 (en) | 2004-10-12 | 2004-10-14 | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
US12/462,926 Abandoned US20090322881A1 (en) | 2004-10-12 | 2009-08-11 | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
US13/076,656 Abandoned US20110211070A1 (en) | 2004-10-12 | 2011-03-31 | Video Analysis, Archiving and Alerting Methods and Appartus for a Distributed, Modular and Extensible Video Surveillance System |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/967,048 Expired - Fee Related US7746378B2 (en) | 2004-10-12 | 2004-10-14 | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
US12/462,926 Abandoned US20090322881A1 (en) | 2004-10-12 | 2009-08-11 | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
Country Status (1)
Country | Link |
---|---|
US (3) | US7746378B2 (en) |
Cited By (27)
* Cited by examiner, † Cited by third partyPublication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080291274A1 (en) * | 2006-09-08 | 2008-11-27 | Marcel Merkel | Method for Operating at Least One Camera |
US20080303902A1 (en) * | 2007-06-09 | 2008-12-11 | Sensomatic Electronics Corporation | System and method for integrating video analytics and data analytics/mining |
US20110044498A1 (en) * | 2009-08-18 | 2011-02-24 | Wesley Kenneth Cobb | Visualizing and updating learned trajectories in video surveillance systems |
US20120198089A1 (en) * | 2011-01-31 | 2012-08-02 | Cbs Interactive, Inc. | System and method for custom segmentation for streaming video |
US8510644B2 (en) * | 2011-10-20 | 2013-08-13 | Google Inc. | Optimization of web page content including video |
US20130215267A1 (en) * | 2008-07-16 | 2013-08-22 | Verint Systems Inc. | System and Method for Capturing, Storing, Analyzing and Displaying Data Related to the Movements of Objects |
WO2013138700A1 (en) * | 2012-03-15 | 2013-09-19 | Behavioral Recognition Systems, Inc. | Alert volume normalization in a video surveillance system |
US20140214885A1 (en) * | 2013-01-31 | 2014-07-31 | Electronics And Telecommunications Research Institute | Apparatus and method for generating evidence video |
WO2014172624A2 (en) * | 2013-04-19 | 2014-10-23 | James Carey | Video identification and analytical recognition system |
US20150116498A1 (en) * | 2012-07-13 | 2015-04-30 | Abb Research Ltd | Presenting process data of a process control object on a mobile terminal |
US20150130934A1 (en) * | 2013-11-14 | 2015-05-14 | Click-It, Inc. | Entertainment device safety system and related methods of use |
WO2015099675A1 (en) * | 2013-12-23 | 2015-07-02 | Pelco, Inc. | Smart view selection in a cloud video service |
CN105045820A (en) * | 2015-06-25 | 2015-11-11 | 浙江立元通信技术股份有限公司 | Method for processing video image information of mass data and database system |
US20150341599A1 (en) * | 2013-03-15 | 2015-11-26 | James Carey | Video identification and analytical recognition system |
US9786113B2 (en) | 2013-03-15 | 2017-10-10 | James Carey | Investigation generation in an observation and surveillance system |
WO2018106437A3 (en) * | 2016-12-06 | 2018-07-26 | Vivint, Inc. | Street watch |
US20190172293A1 (en) * | 2013-03-15 | 2019-06-06 | James Carey | Investigation generation in an observation and surveillance system |
CN110191278A (en) * | 2018-11-30 | 2019-08-30 | 浙江宇视科技有限公司 | Image processing method and device |
US10522013B2 (en) | 2016-05-20 | 2019-12-31 | Vivint, Inc. | Street watch |
US10853882B1 (en) * | 2016-02-26 | 2020-12-01 | State Farm Mutual Automobile Insurance Company | Method and system for analyzing liability after a vehicle crash using video taken from the scene of the crash |
US10861305B2 (en) | 2016-05-20 | 2020-12-08 | Vivint, Inc. | Drone enabled street watch |
US10972704B2 (en) | 2013-03-15 | 2021-04-06 | James Carey | Video identification and analytical recognition system |
CN112737859A (en) * | 2021-01-04 | 2021-04-30 | 中车青岛四方车辆研究所有限公司 | Vehicle-mounted flow auditing and alarming linkage system and flow abnormity judging method |
US11113937B2 (en) | 2016-03-01 | 2021-09-07 | James Carey | Theft prediction and tracking system |
US11417202B2 (en) | 2016-03-01 | 2022-08-16 | James Carey | Theft prediction and tracking system |
US11710392B2 (en) | 2020-09-11 | 2023-07-25 | IDEMIA National Security Solutions LLC | Targeted video surveillance processing |
US11743431B2 (en) | 2013-03-15 | 2023-08-29 | James Carey | Video identification and analytical recognition system |
Families Citing this family (116)
* Cited by examiner, † Cited by third partyPublication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10271017B2 (en) * | 2012-09-13 | 2019-04-23 | General Electric Company | System and method for generating an activity summary of a person |
US7746378B2 (en) * | 2004-10-12 | 2010-06-29 | International Business Machines Corporation | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
US20080195637A1 (en) * | 2004-12-17 | 2008-08-14 | Takehiko Anegawa | Photographed Information Management System and Method Thereof, and Photographed Information Search System |
US20060184553A1 (en) * | 2005-02-15 | 2006-08-17 | Matsushita Electric Industrial Co., Ltd. | Distributed MPEG-7 based surveillance servers for digital surveillance applications |
WO2006103744A1 (en) * | 2005-03-29 | 2006-10-05 | Fujitsu Limited | Video managing system |
US7680773B1 (en) * | 2005-03-31 | 2010-03-16 | Google Inc. | System for automatically managing duplicate documents when crawling dynamic documents |
US7843491B2 (en) * | 2005-04-05 | 2010-11-30 | 3Vr Security, Inc. | Monitoring and presenting video surveillance data |
US8089563B2 (en) * | 2005-06-17 | 2012-01-03 | Fuji Xerox Co., Ltd. | Method and system for analyzing fixed-camera video via the selection, visualization, and interaction with storyboard keyframes |
WO2007031697A1 (en) * | 2005-09-16 | 2007-03-22 | Trevor Burke Technology Limited | Method and apparatus for classifying video data |
US7599957B2 (en) * | 2006-02-15 | 2009-10-06 | Panasonic Corporation | System and method for high performance template driven metadata schema mapping and data storage for surveillance and sensor devices |
US20080122932A1 (en) * | 2006-11-28 | 2008-05-29 | George Aaron Kibbie | Remote video monitoring systems utilizing outbound limited communication protocols |
US20100194868A1 (en) * | 2006-12-15 | 2010-08-05 | Daniel Peled | System, apparatus and method for flexible modular programming for video processors |
CN100585656C (en) * | 2007-03-14 | 2010-01-27 | 中国科学院自动化研究所 | A rule-based all-weather intelligent video analysis and monitoring method |
US8965042B2 (en) * | 2007-03-20 | 2015-02-24 | International Business Machines Corporation | System and method for the measurement of retail display effectiveness |
EP2151128A4 (en) | 2007-04-25 | 2011-11-16 | Miovision Technologies Inc | Method and system for analyzing multimedia content |
US20090031381A1 (en) * | 2007-07-24 | 2009-01-29 | Honeywell International, Inc. | Proxy video server for video surveillance |
US8355961B1 (en) | 2007-08-03 | 2013-01-15 | Sprint Communications Company L.P. | Distribution center head-up display |
US8558893B1 (en) * | 2007-08-03 | 2013-10-15 | Sprint Communications Company L.P. | Head-up security display |
US8451333B2 (en) * | 2007-08-06 | 2013-05-28 | Frostbyte Video, Inc. | Video capture system and method |
US8711224B2 (en) * | 2007-08-06 | 2014-04-29 | Frostbyte Video, Inc. | Image capture system and method |
US8264422B1 (en) | 2007-11-08 | 2012-09-11 | Sprint Communications Company L.P. | Safe head-up display of information |
KR100885734B1 (en) * | 2007-12-17 | 2009-02-26 | 한국전자통신연구원 | Image Information Processing System and Method |
JP5213105B2 (en) * | 2008-01-17 | 2013-06-19 | 株式会社日立製作所 | Video network system and video data management method |
US8427552B2 (en) * | 2008-03-03 | 2013-04-23 | Videoiq, Inc. | Extending the operational lifetime of a hard-disk drive used in video data storage applications |
US9325951B2 (en) | 2008-03-03 | 2016-04-26 | Avigilon Patent Holding 2 Corporation | Content-aware computer networking devices with video analytics for reducing video storage and video communication bandwidth requirements of a video surveillance network camera system |
US9420233B2 (en) * | 2008-03-26 | 2016-08-16 | Pelco, Inc. | Method and apparatus for dynamically controlling a video surveillance system |
JP4585580B2 (en) * | 2008-04-24 | 2010-11-24 | 東芝テック株式会社 | Human flow tracking system |
US8797404B2 (en) * | 2008-07-14 | 2014-08-05 | Honeywell International Inc. | Managing memory in a surveillance system |
EP2304955B1 (en) * | 2008-07-26 | 2017-03-29 | Thomson Licensing | A real-time transport protocol (rtp) packetization method for fast channel change applications using scalable video coding (svc) |
JP2010041535A (en) * | 2008-08-07 | 2010-02-18 | Hitachi Ltd | Data distribution apparatus |
US8502869B1 (en) | 2008-09-03 | 2013-08-06 | Target Brands Inc. | End cap analytic monitoring method and apparatus |
EP2164003A1 (en) * | 2008-09-12 | 2010-03-17 | March Networks Corporation | Distributed video surveillance system |
US8255820B2 (en) * | 2009-06-09 | 2012-08-28 | Skiff, Llc | Electronic paper display device event tracking |
IL201129A (en) * | 2009-09-23 | 2014-02-27 | Verint Systems Ltd | System and method for automatic camera hand off using location measurements |
US8908033B2 (en) * | 2009-09-29 | 2014-12-09 | Avaya Inc. | Utilizing presence information for the purpose of enhanced surveillance |
US9111287B2 (en) * | 2009-09-30 | 2015-08-18 | Microsoft Technology Licensing, Llc | Video content-aware advertisement placement |
US9848114B2 (en) * | 2009-12-07 | 2017-12-19 | Cobra Electronics Corporation | Vehicle camera system |
US9152707B2 (en) * | 2010-01-04 | 2015-10-06 | Martin Libich | System and method for creating and providing media objects in a navigable environment |
WO2011102416A1 (en) * | 2010-02-19 | 2011-08-25 | 株式会社 東芝 | Moving object tracking system and moving object tracking method |
US8817094B1 (en) | 2010-02-25 | 2014-08-26 | Target Brands, Inc. | Video storage optimization |
US8438175B2 (en) * | 2010-03-17 | 2013-05-07 | Lighthaus Logic Inc. | Systems, methods and articles for video analysis reporting |
CN101848377B (en) * | 2010-05-26 | 2012-06-06 | 苏州安杰瑞电子科技发展有限公司 | Device and method for intelligent linkage of multi-video recording device based on cloud computing and mass video searching |
US11082665B2 (en) | 2010-11-05 | 2021-08-03 | Razberi Secure Technologies, Llc | System and method for a security system |
US10477158B2 (en) | 2010-11-05 | 2019-11-12 | Razberi Technologies, Inc. | System and method for a security system |
US10157526B2 (en) | 2010-11-05 | 2018-12-18 | Razberi Technologies, Inc. | System and method for a security system |
US9860490B2 (en) | 2010-11-05 | 2018-01-02 | Tom Galvin | Network video recorder system |
US8922658B2 (en) | 2010-11-05 | 2014-12-30 | Tom Galvin | Network video recorder system |
US9251188B2 (en) * | 2010-11-05 | 2016-02-02 | Nec Corporation | Information processing device |
AU2010257454B2 (en) * | 2010-12-24 | 2014-03-06 | Canon Kabushiki Kaisha | Summary view of video objects sharing common attributes |
JP2012155655A (en) * | 2011-01-28 | 2012-08-16 | Sony Corp | Information processing device, notification method, and program |
US10803724B2 (en) * | 2011-04-19 | 2020-10-13 | Innovation By Imagination LLC | System, device, and method of detecting dangerous situations |
US8970349B2 (en) * | 2011-06-13 | 2015-03-03 | Tyco Integrated Security, LLC | System to provide a security technology and management portal |
EP2727328B8 (en) | 2011-06-28 | 2020-02-26 | NCR Corporation | Image processing to prevent access to private information |
JP2014529930A (en) * | 2011-08-05 | 2014-11-13 | フォックス スポーツ プロダクションズ,インコーポレイティッド | Selective capture and display of a portion of a native image |
US11039109B2 (en) | 2011-08-05 | 2021-06-15 | Fox Sports Productions, Llc | System and method for adjusting an image for a vehicle mounted camera |
JP6190369B2 (en) * | 2011-09-22 | 2017-08-30 | フィリップス ライティング ホールディング ビー ヴィ | Imaging services using outdoor lighting networks |
EP2772049A4 (en) * | 2011-10-24 | 2015-06-17 | Intel Corp | Multiple stream processing for video analytics and encoding |
WO2013074060A1 (en) * | 2011-11-14 | 2013-05-23 | Intel Corporation | Video analytic encoding |
US9569878B2 (en) | 2011-12-07 | 2017-02-14 | Intel Corporation | Sample culling based on linearized 5D edge equations |
CN102572215A (en) * | 2011-12-14 | 2012-07-11 | 深圳市贝尔信智能系统有限公司 | City-class visual video analysis method and server |
US8712126B2 (en) * | 2012-03-12 | 2014-04-29 | Xerox Corporation | Web-based system and method for video analysis |
RU2484529C1 (en) * | 2012-03-21 | 2013-06-10 | Общество с ограниченной ответственностью "Синезис" | Method of ranking video data |
US8825368B2 (en) * | 2012-05-21 | 2014-09-02 | International Business Machines Corporation | Physical object search |
US8856082B2 (en) | 2012-05-23 | 2014-10-07 | International Business Machines Corporation | Policy based population of genealogical archive data |
US20180288461A1 (en) * | 2012-06-21 | 2018-10-04 | Google Inc. | Web Analytics for Video Level Events |
US9213781B1 (en) | 2012-09-19 | 2015-12-15 | Placemeter LLC | System and method for processing image data |
CN103412859B (en) * | 2012-10-11 | 2016-09-14 | 华迪计算机集团有限公司 | Massive video method for quickly retrieving based on media asset management system and device |
US8934754B2 (en) | 2012-11-13 | 2015-01-13 | International Business Machines Corporation | Providing emergency access to surveillance video |
US9681103B2 (en) | 2012-11-13 | 2017-06-13 | International Business Machines Corporation | Distributed control of a heterogeneous video surveillance network |
US9041812B2 (en) | 2012-11-13 | 2015-05-26 | International Business Machines Corporation | Automated authorization to access surveillance video based on pre-specified events |
US9684719B2 (en) * | 2012-12-10 | 2017-06-20 | Verint Systems Ltd. | Object search by description |
US9230599B2 (en) | 2013-01-23 | 2016-01-05 | Fleye, Inc. | Storage and editing of video and sensor data from athletic performances of multiple individuals in a venue |
JP6171374B2 (en) * | 2013-02-06 | 2017-08-02 | ソニー株式会社 | Information processing apparatus, information processing method, program, and information processing system |
US20140280557A1 (en) * | 2013-03-15 | 2014-09-18 | SaferAging, Inc. | High reliability alert delivery using web-based interfaces |
US9392099B2 (en) * | 2013-06-24 | 2016-07-12 | Cellco Partnership | Mobile surveillance system |
WO2015009463A2 (en) * | 2013-07-03 | 2015-01-22 | Lindblom Timothy P | Multiple retail device universal data gateway |
US9201581B2 (en) * | 2013-07-31 | 2015-12-01 | International Business Machines Corporation | Visual rules for decision management |
US20150055832A1 (en) * | 2013-08-25 | 2015-02-26 | Nikolay Vadimovich PTITSYN | Method for video data ranking |
CN103475882B (en) * | 2013-09-13 | 2017-02-15 | 北京大学 | Surveillance video encoding and recognizing method and surveillance video encoding and recognizing system |
US20160110972A1 (en) * | 2014-04-10 | 2016-04-21 | Smartvue Corporation | Systems and methods for automated cloud-based analytics for surveillance systems |
WO2015184440A2 (en) | 2014-05-30 | 2015-12-03 | Placemeter Inc. | System and method for activity monitoring using video data |
US9807337B2 (en) | 2014-09-10 | 2017-10-31 | Fleye, Inc. | Storage and editing of video of activities using sensor and tag data of participants and spectators |
US9396397B2 (en) * | 2014-10-21 | 2016-07-19 | Motorola Solutions, Inc. | Methods and systems for object-recognition and link integration in a composite video stream |
US10491796B2 (en) | 2014-11-18 | 2019-11-26 | The Invention Science Fund Ii, Llc | Devices, methods and systems for visual imaging arrays |
US11159854B2 (en) | 2014-12-13 | 2021-10-26 | Fox Sports Productions, Llc | Systems and methods for tracking and tagging objects within a broadcast |
US11758238B2 (en) | 2014-12-13 | 2023-09-12 | Fox Sports Productions, Llc | Systems and methods for displaying wind characteristics and effects within a broadcast |
US9965684B2 (en) | 2014-12-18 | 2018-05-08 | Sensormatic Electronics, LLC | Method and system for queue length analysis |
CN105898207B (en) * | 2015-01-26 | 2019-05-10 | 杭州海康威视数字技术股份有限公司 | Intelligent processing method and system for video data |
KR20170115082A (en) * | 2015-03-04 | 2017-10-16 | 가부시키가이샤 히타치 시스테무즈 | A system for checking the situation by camera image data, a method for checking the situation by the control device and the camera image data |
US11334751B2 (en) | 2015-04-21 | 2022-05-17 | Placemeter Inc. | Systems and methods for processing video data for activity monitoring |
US10043078B2 (en) | 2015-04-21 | 2018-08-07 | Placemeter LLC | Virtual turnstile system and method |
CN105208326A (en) * | 2015-08-28 | 2015-12-30 | 深圳市融创天下科技有限公司 | Video-cloud-based early warning method and system of public security threat of urban region |
US11100335B2 (en) * | 2016-03-23 | 2021-08-24 | Placemeter, Inc. | Method for queue time estimation |
KR102586962B1 (en) * | 2016-04-07 | 2023-10-10 | 한화비전 주식회사 | Surveillance system and controlling method thereof |
CN105959621A (en) * | 2016-04-28 | 2016-09-21 | 泰华智慧产业集团股份有限公司 | Quasi-real time control distribution system and method based on multi-source video structural data |
US10089844B2 (en) * | 2016-06-09 | 2018-10-02 | Avigilon Corporation | Camera with wireless power transmitter |
US9916493B2 (en) | 2016-08-03 | 2018-03-13 | At&T Intellectual Property I, L.P. | Method and system for aggregating video content |
KR102249498B1 (en) * | 2016-08-17 | 2021-05-11 | 한화테크윈 주식회사 | The Apparatus And System For Searching |
WO2018117753A1 (en) * | 2016-12-23 | 2018-06-28 | Samsung Electronics Co., Ltd. | Electronic device and method of controlling the same |
US10997421B2 (en) * | 2017-03-30 | 2021-05-04 | Hrl Laboratories, Llc | Neuromorphic system for real-time visual activity recognition |
US10373458B2 (en) * | 2017-04-20 | 2019-08-06 | Deep Sentinel Corp. | Automatic threat detection based on video frame delta information in compressed video streams |
US10186124B1 (en) * | 2017-10-26 | 2019-01-22 | Scott Charles Mullins | Behavioral intrusion detection system |
US10867398B2 (en) * | 2017-11-21 | 2020-12-15 | Reliance Core Consulting LLC | Methods, systems, apparatuses and devices for facilitating motion analysis in an environment |
US10819684B2 (en) | 2017-11-24 | 2020-10-27 | International Business Machines Corporation | Cognitive blockchain for internet of things |
US10686611B2 (en) | 2017-11-24 | 2020-06-16 | International Business Machines Corporation | Data anonymizing blockchain system |
US20200036945A1 (en) * | 2018-07-24 | 2020-01-30 | Comcast Cable Communications, Llc | Neighborhood Proximity Media Capture For Life Safety Alarm Events |
CN109685459A (en) * | 2018-12-17 | 2019-04-26 | 台州三石量子科技有限公司 | Case Computer Aided Analysis System |
MY194748A (en) * | 2018-12-21 | 2022-12-15 | Mimos Berhad | A system and method for video surveillance and monitoring |
CN109919009A (en) * | 2019-01-24 | 2019-06-21 | 北京明略软件系统有限公司 | The monitoring method of target object, apparatus and system |
CA3136497A1 (en) * | 2019-04-10 | 2020-10-15 | Scott Charles Mullins | Monitoring systems |
US11165954B1 (en) * | 2019-06-19 | 2021-11-02 | Objectvideo Labs, Llc | Scene-aware custom tuned video surveillance detection system |
US11587243B2 (en) * | 2019-10-25 | 2023-02-21 | 7-Eleven, Inc. | System and method for position tracking using edge computing |
WO2021100574A1 (en) * | 2019-11-18 | 2021-05-27 | パナソニックi-PROセンシングソリューションズ株式会社 | Investigation assistance system and investigation assistance method |
EP4068791A4 (en) * | 2019-11-26 | 2023-11-01 | Hanwha Vision Co., Ltd. | EVENT-ORIENTED MULTI-CHANNEL IMAGE SECURITY DEVICE AND METHOD THEREOF AND NETWORK SURVEILLANCE CAMERA SYSTEM THEREFOR |
US11328565B2 (en) * | 2019-11-26 | 2022-05-10 | Ncr Corporation | Asset tracking and notification processing |
US11475669B2 (en) | 2020-07-30 | 2022-10-18 | Ncr Corporation | Image/video analysis with activity signatures |
Citations (7)
* Cited by examiner, † Cited by third partyPublication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6271752B1 (en) * | 1998-10-02 | 2001-08-07 | Lucent Technologies, Inc. | Intelligent multi-access system |
US20020056082A1 (en) * | 1999-11-17 | 2002-05-09 | Hull Jonathan J. | Techniques for receiving information during multimedia presentations and communicating the information |
US20040090462A1 (en) * | 1997-12-22 | 2004-05-13 | Ricoh Company, Ltd. | Multimedia visualization and integration environment |
US20050132414A1 (en) * | 2003-12-02 | 2005-06-16 | Connexed, Inc. | Networked video surveillance system |
US6988245B2 (en) * | 2002-06-18 | 2006-01-17 | Koninklijke Philips Electronics N.V. | System and method for providing videomarks for a video program |
US20060072013A1 (en) * | 2004-09-23 | 2006-04-06 | Martin Renkis | Wireless video surveillance system and method with two-way locking of input capture devices |
US7746378B2 (en) * | 2004-10-12 | 2010-06-29 | International Business Machines Corporation | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
Family Cites Families (1)
* Cited by examiner, † Cited by third partyPublication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6525281B2 (en) * | 2001-07-11 | 2003-02-25 | Reitter & Schefenacker Gmbh & Co. Kg | Emergency call button for vehicles, in particular, motor vehicles |
-
2004
- 2004-10-14 US US10/967,048 patent/US7746378B2/en not_active Expired - Fee Related
-
2009
- 2009-08-11 US US12/462,926 patent/US20090322881A1/en not_active Abandoned
-
2011
- 2011-03-31 US US13/076,656 patent/US20110211070A1/en not_active Abandoned
Patent Citations (7)
* Cited by examiner, † Cited by third partyPublication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040090462A1 (en) * | 1997-12-22 | 2004-05-13 | Ricoh Company, Ltd. | Multimedia visualization and integration environment |
US6271752B1 (en) * | 1998-10-02 | 2001-08-07 | Lucent Technologies, Inc. | Intelligent multi-access system |
US20020056082A1 (en) * | 1999-11-17 | 2002-05-09 | Hull Jonathan J. | Techniques for receiving information during multimedia presentations and communicating the information |
US6988245B2 (en) * | 2002-06-18 | 2006-01-17 | Koninklijke Philips Electronics N.V. | System and method for providing videomarks for a video program |
US20050132414A1 (en) * | 2003-12-02 | 2005-06-16 | Connexed, Inc. | Networked video surveillance system |
US20060072013A1 (en) * | 2004-09-23 | 2006-04-06 | Martin Renkis | Wireless video surveillance system and method with two-way locking of input capture devices |
US7746378B2 (en) * | 2004-10-12 | 2010-06-29 | International Business Machines Corporation | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
Cited By (60)
* Cited by examiner, † Cited by third partyPublication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080291274A1 (en) * | 2006-09-08 | 2008-11-27 | Marcel Merkel | Method for Operating at Least One Camera |
US20080303902A1 (en) * | 2007-06-09 | 2008-12-11 | Sensomatic Electronics Corporation | System and method for integrating video analytics and data analytics/mining |
US8872922B2 (en) * | 2008-07-16 | 2014-10-28 | Verint Americas Inc. | System and method for capturing, storing, analyzing and displaying data related to the movements of objects |
US20130215267A1 (en) * | 2008-07-16 | 2013-08-22 | Verint Systems Inc. | System and Method for Capturing, Storing, Analyzing and Displaying Data Related to the Movements of Objects |
US8878937B2 (en) * | 2008-07-16 | 2014-11-04 | Verint Americas Inc. | System and method for capturing, storing, analyzing and displaying data related to the movements of objects |
US20110044498A1 (en) * | 2009-08-18 | 2011-02-24 | Wesley Kenneth Cobb | Visualizing and updating learned trajectories in video surveillance systems |
US8280153B2 (en) * | 2009-08-18 | 2012-10-02 | Behavioral Recognition Systems | Visualizing and updating learned trajectories in video surveillance systems |
US20120198089A1 (en) * | 2011-01-31 | 2012-08-02 | Cbs Interactive, Inc. | System and method for custom segmentation for streaming video |
US8583818B2 (en) * | 2011-01-31 | 2013-11-12 | Cbs Interactive Inc. | System and method for custom segmentation for streaming video |
US8510644B2 (en) * | 2011-10-20 | 2013-08-13 | Google Inc. | Optimization of web page content including video |
US11727689B2 (en) | 2012-03-15 | 2023-08-15 | Intellective Ai, Inc. | Alert directives and focused alert directives in a behavioral recognition system |
CN104254873A (en) * | 2012-03-15 | 2014-12-31 | 行为识别系统公司 | Alert volume normalization in a video surveillance system |
WO2013138700A1 (en) * | 2012-03-15 | 2013-09-19 | Behavioral Recognition Systems, Inc. | Alert volume normalization in a video surveillance system |
US11217088B2 (en) | 2012-03-15 | 2022-01-04 | Intellective Ai, Inc. | Alert volume normalization in a video surveillance system |
US12094212B2 (en) | 2012-03-15 | 2024-09-17 | Intellective Ai, Inc. | Alert directives and focused alert directives in a behavioral recognition system |
US9349275B2 (en) | 2012-03-15 | 2016-05-24 | Behavorial Recognition Systems, Inc. | Alert volume normalization in a video surveillance system |
US20150116498A1 (en) * | 2012-07-13 | 2015-04-30 | Abb Research Ltd | Presenting process data of a process control object on a mobile terminal |
US9208226B2 (en) * | 2013-01-31 | 2015-12-08 | Electronics And Telecommunications Research Institute | Apparatus and method for generating evidence video |
US20140214885A1 (en) * | 2013-01-31 | 2014-07-31 | Electronics And Telecommunications Research Institute | Apparatus and method for generating evidence video |
US9762865B2 (en) * | 2013-03-15 | 2017-09-12 | James Carey | Video identification and analytical recognition system |
US11546557B2 (en) | 2013-03-15 | 2023-01-03 | James Carey | Video identification and analytical recognition system |
US11756367B2 (en) | 2013-03-15 | 2023-09-12 | James Carey | Investigation generation in an observation and surveillance system |
US11743431B2 (en) | 2013-03-15 | 2023-08-29 | James Carey | Video identification and analytical recognition system |
US11869325B2 (en) | 2013-03-15 | 2024-01-09 | James Carey | Video identification and analytical recognition system |
US20200242876A1 (en) * | 2013-03-15 | 2020-07-30 | James Carey | Investigation generation in an observation and surveillance system |
US9786113B2 (en) | 2013-03-15 | 2017-10-10 | James Carey | Investigation generation in an observation and surveillance system |
US20150341599A1 (en) * | 2013-03-15 | 2015-11-26 | James Carey | Video identification and analytical recognition system |
US11881090B2 (en) * | 2013-03-15 | 2024-01-23 | James Carey | Investigation generation in an observation and surveillance system |
US11039108B2 (en) * | 2013-03-15 | 2021-06-15 | James Carey | Video identification and analytical recognition system |
US20190172293A1 (en) * | 2013-03-15 | 2019-06-06 | James Carey | Investigation generation in an observation and surveillance system |
US10347070B2 (en) | 2013-03-15 | 2019-07-09 | James Carey | Investigation generation in an observation and surveillance system |
US10972704B2 (en) | 2013-03-15 | 2021-04-06 | James Carey | Video identification and analytical recognition system |
US10432897B2 (en) * | 2013-03-15 | 2019-10-01 | James Carey | Video identification and analytical recognition system |
US10846971B2 (en) | 2013-03-15 | 2020-11-24 | James Carey | Investigation generation in an observation and surveillance system |
US10657755B2 (en) * | 2013-03-15 | 2020-05-19 | James Carey | Investigation generation in an observation and surveillance system |
WO2014172624A3 (en) * | 2013-04-19 | 2015-01-29 | James Carey | Video identification and analytical recognition system |
US11100334B2 (en) | 2013-04-19 | 2021-08-24 | James Carey | Video identification and analytical recognition system |
WO2014172624A2 (en) * | 2013-04-19 | 2014-10-23 | James Carey | Video identification and analytical recognition system |
US11587326B2 (en) | 2013-04-19 | 2023-02-21 | James Carey | Video identification and analytical recognition system |
US10078787B2 (en) | 2013-04-19 | 2018-09-18 | James Carey | Crowd-based video identification and analytical recognition system |
US11488409B2 (en) | 2013-11-14 | 2022-11-01 | James Carey | Entertainment device safety system and related methods of use |
US9773163B2 (en) * | 2013-11-14 | 2017-09-26 | Click-It, Inc. | Entertainment device safety system and related methods of use |
US10303935B2 (en) * | 2013-11-14 | 2019-05-28 | James Carey | Entertainment device safety system and related methods of use |
US12033417B2 (en) | 2013-11-14 | 2024-07-09 | James Carey | Entertainment device safety system and related methods of use |
US10915742B2 (en) | 2013-11-14 | 2021-02-09 | James Carey | Entertainment device safety system and related methods of use |
US20150130934A1 (en) * | 2013-11-14 | 2015-05-14 | Click-It, Inc. | Entertainment device safety system and related methods of use |
WO2015099675A1 (en) * | 2013-12-23 | 2015-07-02 | Pelco, Inc. | Smart view selection in a cloud video service |
CN106031165A (en) * | 2013-12-23 | 2016-10-12 | 派尔高公司 | Smart view selection in a cloud video service |
CN105045820A (en) * | 2015-06-25 | 2015-11-11 | 浙江立元通信技术股份有限公司 | Method for processing video image information of mass data and database system |
US10853882B1 (en) * | 2016-02-26 | 2020-12-01 | State Farm Mutual Automobile Insurance Company | Method and system for analyzing liability after a vehicle crash using video taken from the scene of the crash |
US11113937B2 (en) | 2016-03-01 | 2021-09-07 | James Carey | Theft prediction and tracking system |
US11417202B2 (en) | 2016-03-01 | 2022-08-16 | James Carey | Theft prediction and tracking system |
US11710397B2 (en) | 2016-03-01 | 2023-07-25 | James Carey | Theft prediction and tracking system |
US11183037B2 (en) | 2016-05-20 | 2021-11-23 | Vivint, Inc. | Street watch |
US10861305B2 (en) | 2016-05-20 | 2020-12-08 | Vivint, Inc. | Drone enabled street watch |
US10522013B2 (en) | 2016-05-20 | 2019-12-31 | Vivint, Inc. | Street watch |
WO2018106437A3 (en) * | 2016-12-06 | 2018-07-26 | Vivint, Inc. | Street watch |
CN110191278A (en) * | 2018-11-30 | 2019-08-30 | 浙江宇视科技有限公司 | Image processing method and device |
US11710392B2 (en) | 2020-09-11 | 2023-07-25 | IDEMIA National Security Solutions LLC | Targeted video surveillance processing |
CN112737859A (en) * | 2021-01-04 | 2021-04-30 | 中车青岛四方车辆研究所有限公司 | Vehicle-mounted flow auditing and alarming linkage system and flow abnormity judging method |
Also Published As
Publication number | Publication date |
---|---|
US7746378B2 (en) | 2010-06-29 |
US20090322881A1 (en) | 2009-12-31 |
US20060078047A1 (en) | 2006-04-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7746378B2 (en) | 2010-06-29 | Video analysis, archiving and alerting methods and apparatus for a distributed, modular and extensible video surveillance system |
CN100452871C (en) | 2009-01-14 | Video analysis, archiving and alerting methods and apparatus for a video surveillance system |
EP2781084B1 (en) | 2021-02-24 | Digital video system with intelligent video selection timeline |
US20110010624A1 (en) | 2011-01-13 | Synchronizing audio-visual data with event data |
US9342594B2 (en) | 2016-05-17 | Indexing and searching according to attributes of a person |
EP2936388B1 (en) | 2018-08-08 | System and method for detection of high-interest events in video data |
EP1024666A2 (en) | 2000-08-02 | Image recording/reproducing apparatus in monitor system |
US20150264296A1 (en) | 2015-09-17 | System and method for selection and viewing of processed video |
US20060077254A1 (en) | 2006-04-13 | Apparatus and methods for establishing and managing a distributed, modular and extensible video surveillance system |
US20170034483A1 (en) | 2017-02-02 | Smart shift selection in a cloud video service |
US20160357762A1 (en) | 2016-12-08 | Smart View Selection In A Cloud Video Service |
CA2569671A1 (en) | 2006-07-06 | Method and system for wide area security monitoring, sensor management and situational awareness |
US20060184553A1 (en) | 2006-08-17 | Distributed MPEG-7 based surveillance servers for digital surveillance applications |
WO2011063527A1 (en) | 2011-06-03 | Enterprise management system and auditing method employed thereby |
US10719552B2 (en) | 2020-07-21 | Focalized summarizations of a video stream |
JP4678043B2 (en) | 2011-04-27 | Image storage device, monitoring system, storage medium |
US20070226616A1 (en) | 2007-09-27 | Method and System For Wide Area Security Monitoring, Sensor Management and Situational Awareness |
CN115966313A (en) | 2023-04-14 | Integrated management platform based on face recognition |
US20240119736A1 (en) | 2024-04-11 | System and Method to Facilitate Monitoring Remote Sites using Bandwidth Optimized Intelligent Video Streams with Enhanced Selectivity |
EP3065398B1 (en) | 2017-11-22 | Dynamic layouts |
CN116010652B (en) | 2023-06-09 | Unstructured video data processing method and system |
CN112437270B (en) | 2021-09-28 | Monitoring video playing method and device and readable storage medium |
Perrott et al. | 2002 | Real-time multimedia tagging and content-based retrieval for CCTV surveillance systems |
WO2011005619A1 (en) | 2011-01-13 | Synchronizing audio-visual data with event data |
CN111858779A (en) | 2020-10-30 | Data analysis method and device, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
2015-02-22 | STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |