A home health care monitoring method and system are disclosed. In one embodiment, a method includes: capturing a plurality of video sequences from a plurality of respective cameras disposed in different locations within a patient's home, including capturing two-dimensional image data and depth data for each video sequence; defining a plurality of events to monitor associated with the patient, the events including, during a predetermined time period, at least (1) the patient's body being in a particular position for at least a predetermined amount of time; performing depth-enhanced video content analysis on the plurality of video sequences to determine whether the event (1) has occurred. The video content analysis may include: for each of at least two cameras: automatically detecting a human object based on the two-dimensional image data; using the depth data to confirm whether the human object is in the particular position; and based on the confirmation, tracking an amount of time that the human object is in the particular position; and then determining that the event (1) has occurred based on the collective tracked amount of time.
|
1. A method of monitoring the health of a person in home, the method comprising:
capturing a plurality of video sequences from a plurality of respective cameras disposed in different locations within the patient's home, including capturing two-dimensional image data and depth data for each video sequence;
defining a plurality of events to monitor associated with the patient, the events including, during a predetermined time period, at least (1) the patient's body being in a particular position for at least a predetermined amount of time;
performing depth-enhanced video content analysis on the plurality of video sequences to determine whether the event (1) has occurred, wherein the video content analysis includes:
for each of at least two cameras:
automatically detecting a human object based on the two-dimensional image data;
using the depth data associated with the detected human object to confirm whether the human object is in the particular position, without using depth data associated with other parts of the two-dimensional image in the confirmation; and
based on the confirmation, tracking an amount of time that the human object is in the particular position; and
determining that the event (1) has occurred based on the collective tracked amount of time.
10. A home health monitoring system, comprising:
a plurality of cameras in different rooms in a home, the different cameras configured to capture two-dimensional image data and depth data from a plurality of respective video images;
a central computer system configured to perform video content analysis based on the two-dimensional image data and the depth data; and
one or more communication media configured to transmit data from the plurality of cameras to the central computer system,
wherein the video content analysis includes:
setting a plurality of events to monitor associated with an inhabitant in the home, the events including, during a predetermined time period, (1) the patient's body being in a particular position for at least a predetermined period of time;
performing video content analysis on the plurality of video sequences to determine whether the event (1) has occurred, wherein the video content analysis includes:
for each of at least two cameras:
automatically detecting a human object based on the two-dimensional image data;
using the depth data associated with the detected human object to confirm whether the human object is in the particular position, without using depth data associated with other parts of the two dimensional image in the confirmation;
based on the confirmation, tracking an amount of time that the human object is in the particular position; and
determining that the event (1) has occurred based on the tracked amount of time.
2. The method of
3. The method of
4. The method of
5. The method of
6. The method of
7. The method of
9. The method of
11. The home health monitoring system of
|
This application claims the benefit of priority to U.S. Provisional Patent Application No. 61/587,186, filed Jan. 17, 2012, the contents of which are incorporated herein by reference in their entirety.
1. Field
This disclosure relates to a system for performing home health care monitoring. More specifically, this disclosure relates to a system for performing home health care monitoring using depth-enhanced video content analysis.
2. Background
Many people, particularly the elderly, need to take certain medications and maintain a certain level of physical activity to stay healthy. In addition, as people get older, there tends to be a greater likelihood of accidents or sudden emergencies, such as falling down, stroke, heart attacks, emergencies due to a failure to take medicine, etc. In addition, many elderly people suffer from depression, dementia, or other conditions that alter their daily behavior. Preventing and treating these accidents, emergencies, and conditions is very important but can be very expensive. For example, people may move into assisted living quarters or nursing homes, or may hire hospice care or home help.
To assist elderly patients, various less expensive alternatives have been used. For example, using Lifeline® by Philips, elderly patients can simply push a button if an emergency situation occurs, which alerts emergency response personnel. Other systems may use video or other electrical equipment to monitor a patient's health. For example, U.S. Pat. No. 5,544,649 to David et al., published on Aug. 13, 1996, describes a patient health monitoring system that includes video cameras in a person's home that connect to a central station for health care workers to monitor. Additional medical condition sensing and monitoring equipment, such as blood pressure, pulse, and temperature monitoring devices, may also be used at the patient's home. However, such a system uses a fairly large number of health care workers per patient, and therefore can still be fairly expensive and requires constant attention of the workers.
In hospital environments, cameras that perform automatic detection of certain patient behaviors have been proposed. For example, U.S. Patent Application Publication No. 2012/0075464 to Derenne et al., published on Mar. 29, 2012, proposes the use of video cameras in the hospital environment to determine certain patient conditions, such as whether a patient is sleeping, exiting a bed, walking, or falling. Proposed cameras for such a system include RGB cameras with depth sensors that may be used to provide full-body 3D motion detection, among other things.
However, systems such as proposed by Derenne et al., while monitoring individual events, do not monitor collective household activity such as to be expected of an elderly patient living at home. In addition, systems that provide full 3D motion detection, for example by analyzing three-dimensional data for all parts of a video scene (e.g., all pixels of a series of video frames) can be computationally complex, and may require special software and processing capability beyond the scope of traditional two-dimensional monitoring schemes, which may further increase the expense of such monitoring.
An example of a two-dimensional video content analysis (VCA) system is described in U.S. Pat. No. 7,932,923, issued to Lipton et al. on Apr. 26, 2011 (the '923 patent), the contents of which are incorporated herein by reference in their entirety. Some existing systems use RGB (red green blue), CMYK (cyan magenta yellow key), YCbCr, or other sensors that sense images in a two-dimensional manner and perform analysis of those images to perform object and event detection. Other existing systems use depth sensors, to generate three-dimensional data or depth maps, which are then analyzed using different software in order to perform object and event detection. In some ways, the systems that use depth sensors are more accurate than the two-dimensional systems. For example, the depth sensor systems may obtain more accurate three-dimensional information, and may deal better with occlusions. However, depth data and images determined by depth sensor systems are generally lower in resolution than RGB data, and may therefore include fewer details than RGB images. In addition, depth sensors are a relatively new technology for video analysis, and are still prone to error in determining three-dimensional coordinates. Further, certain information resulting from depth sensors often remains incomplete, such as depth information for objects with specularities, or depth information for featureless surfaces extracted from stereo.
Certain systems may combine both depth and RGB data in order to perform analysis on complex three-dimensional scenes. For example, as described in U.S. Pat. No. 7,831,087, depth data and optional non-depth data are used to generate a plan-view image, which plan view image can then be analyzed by classifying objects in the plan view image. However, systems such as this, which perform complex analysis on depth data and optional additional data in order to perform object detection or event detection, still suffer from the problems above relating the drawbacks of depth sensor systems. For example, some of the depth data may be missing or may be inaccurate, resulting in an analysis of faulty data. In addition, performing analysis on three-dimensional data generally requires more complex algorithms and may require a complete re-design of hardware and/or software that performs the analysis, compared to more traditional two-dimensional image analysis systems.
The embodiments described here address some of these problems of existing systems, and provide a new and simplified way to use depth data to assist in image analysis and video content analysis throughout a patient home environment. As a result, better home health care monitoring can be achieved using automated systems that are more accurate and reliable than prior systems.
The disclosed embodiments include a home health care monitoring system. The system includes a plurality of sensors throughout a patient's home, and performs depth-enhanced video content analysis on a plurality of video sequences captured by the plurality of cameras. In certain embodiments, the system performs initial video content analysis steps, such as object detection, on two-dimensional image data captured by the cameras. The system then performs additional analysis on the detected objects using depth data captured by the cameras. For example, the object detection may detect a human in a video frame, and the depth data may then be used to determine and/or confirm whether the human is lying down, sitting, or standing. Based on this process, information such as how long a patient is sitting, standing, and lying down throughout the home, or how often a patient visits certain rooms can be determined, and such information can be used to assist in monitoring the patient's health.
Example embodiments will be more clearly understood from the following detailed description taken in conjunction with the accompanying drawings. The figures represent non-limiting example embodiments as described herein.
The present disclosure will be described more fully hereinafter with reference to the accompanying drawings, in which various embodiments are shown. The invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. In the drawings, like numbers refer to like elements throughout.
It will be understood that when an element is referred to as being “connected” or “coupled” to or “in communication with” another element, it can be directly connected or coupled to or in communication with the other element or intervening elements may be present. In contrast, when an element is referred to as being “directly connected” or “directly coupled” or “in direct communication with” another element, there are no intervening elements present. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items and may be abbreviated as “/”.
It will be understood that, although the terms first, second, etc. may be used herein to describe various elements, these elements should not be limited by these terms. Unless indicated otherwise, these terms are only used to distinguish one element from another. For example, a first signal could be termed a second signal, and, similarly, a second signal could be termed a first signal without departing from the teachings of the disclosure.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” or “includes” and/or “including” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and/or the present application, and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
This disclosure includes particular terminology and descriptions that relate to home health care, video surveillance, and video content analysis. The descriptions are intended to provide a framework for certain terms and concepts, and are not intended to limit the scope of this disclosure unless explicitly stated.
In one embodiment, cameras such as 112, 122, 132, and 142 may be positioned to capture video frames likely to relate to events of interest for home health care monitoring. For example, if the amount of time sitting, standing, and lying down is to be monitored, then cameras may be positioned to capture video of the bed 114 (e.g., camera 112), the kitchen chairs 138 (e.g., camera 132), and the couch 146 (e.g., camera 142). As another example, if a number of times entering a kitchen or bathroom is to be monitored, then kitchen camera 132 and bathroom camera 122 may be positioned to capture video of the entranceways to the kitchen and bathroom. As yet another embodiment, if the taking of oral medication is to be monitored, then both the bathroom camera 122 and the kitchen camera 132 may be positioned to capture video of the respective sinks 126, 135, or the counter spaces.
As described further below, a video surveillance system (e.g., including cameras 112, 122, 132, and 142) can use depth data to assist in monitoring home health care in a patient's home. For example, video sequences from multiple devices in different rooms can be collectively analyzed to detect certain behaviors and monitor overall behavioral patterns of a patient at home. These sequences can be analyzed by a central computing system in the home, and/or by a remote computing system (e.g., connected via a telecommunications medium and over a network such as the Internet). Alternatively or additionally, some or all of the depth enhanced video sequences can be analyzed by the cameras that capture them. Events can be monitored, and as a result, alerts can be configured to warn the patient, a monitoring service, or outside professionals that an event occurred. For example, alerts can be sent to a health care professional via e-mail, phone, or other audio-visual means, and the alerts can include a snapshot or video clip or live feed of video from a camera in the patient's home. In one embodiment, an alerted professional can request a live audio or video communication with the patient prior to communicating. To address privacy concerns, in one embodiment, a video of the patient is only sent if the patient gives permission, or if an emergency alert is issued.
In one embodiment, each of the video cameras/depth sensors 112, 122, 132, 142 may be networked or otherwise in communication (e.g., hard wired or wirelessly) with a server (not shown). Each video camera may include a processor to perform video content analysis of the corresponding video images taken. The content analysis may analyze the two dimensional video image data with the depth information provided by the depth sensor associated with the video camera, and may also analyze the two dimensional video image data alone. On camera processors of each video camera may perform such content analysis to generate video primitives, also referred to herein as metadata, and stream the video primitives/metadata to the server. The video primitives/metadata may represent detected objects, detected classification and/or characteristics of the detected objects and/or actions and/or events (e.g., of the detected objects) detected in the corresponding video. The video primitives, or metadata, may be associated with each frame of the video sequence. By way of example, see U.S. Pat. No. 7,868,912 issued to Venetianer et al. and U.S. Pat. No. 7,932,923 issued to Lipton et al., both of which are incorporated herein by reference in their entirety, for exemplary details of video primitive (or metadata) generation and downstream processing (which may be real time processing or later processing) to obtain information from the video, such as event detection, using the generated video primitives. Depth data associated with the video image data may be provided to the server as metadata along with other metadata. Alternatively and/or in addition, height data derived from the depth data (e.g., from on camera processing) may be provided to the server as metadata along with other metadata. The depth metadata and/or height metadata may be associated with detected objects and may include depth and/or height of multiple elements of the detected object. The depth and/or height data and other metadata obtained from on camera processing of the video image data of the corresponding video camera may be streamed to the server.
Alternatively, the video camera/depth sensors 112, 122, 132, 142 may provide recorded video and associated depth data to the server or another computer without processing. In this example, each camera may stream to a server or to another computer the video image data together with the depth data. The server or the other computer may then process the video image data and depth data provided by the video cameras/depth sensors 112, 122, 132, 142. Such processing may also generate metadata derived from the video image data and depth metadata and/or height metadata as described previously.
The metadata may be processed to classify objects, and to detect actions and events without reprocessing the original video image data. Upon detecting an action/event of interest, the original video image data may be accessed by a user to verify the action/event detection or to review for other purposes.
As described in greater detail below, depth data added to a two-dimensional analysis can be used to detect various types of events. Further events can be based on an individual episode, or a collection of episodes involving the patient throughout the house. As an example, using the two-dimensional data and depth data, a home health care monitoring system can first use conventional two-dimensional processing to determine, for example that the patient is in a particular room. Then, using depth data, the system can determine a height of the patient in order to determine whether the person is standing, sitting, or lying down. That information can be monitored over time, in order to develop a behavioral profile indicating how often the person is lying down, sitting, and/or standing throughout a day and throughout the house.
As another example, events such as a patient falling down, getting out of bed, or staying in bed too long can be more accurately determined using depth data in combination with two-dimensional image data, also referred to as depth-enhanced video data in a video monitoring system. For example, in a traditional two-dimensional monitoring system, a person can be detected using, for example, face recognition or body shape recognition. However, the system may still have trouble determining if the person is standing, sitting up, or lying down. Therefore, the various embodiments discussed herein can better determine whether a person is standing up or lying down, or whether a person has recently fallen down, by using the depth data in combination with two-dimensional image data.
In other embodiments, based on information collected throughout a day, for example, the monitoring system can determine how many times a patient visited different rooms or performed different activities. For example, a count of bathroom visits or kitchen visits can be used to determine if the patient is exhibiting normal healthy behavior (e.g., if the person is eating enough). A determination of how many times a patient took medicine can also be used to determine if the patient is following recommended healthy patterns. Also, to assist the patient and prevent the patient from falling, the monitoring system can be integrated with a lighting system, in order to turn lights automatically on or off depending on whether a person has woken up, fallen asleep, gotten out of bed, etc.
In addition, in situations where a patient is occluded by one or more objects (e.g., a bedpost, a kitchen table, a pet, etc.), by using the depth data, a more accurate monitoring of the patient can be performed.
In order to appreciate the advantages of using a depth sensing system for home health care, a detailed discussion of certain existing video content analysis systems, as well a combined two-dimensional and depth sensing system according to various embodiments, is discussed below. Then, additional applications of such technology for home health care systems are described in greater detail.
VCA systems may use cameras that are calibrated in order to detect and identify objects. For example, rather than simply detecting an object based on its relative dimensions, which can represent, for example, a shape of pill box or a shape of a human being, calibrated VCA systems are able to detect a shape of an object as well as its real-world size. As a result, the system can more accurately detect certain objects. For example, in a non-calibrated system, a VCA system for counting a number of people that appear in a frame of a video stream may count the shapes of both actual people, and of miniature dolls in the frame as people. Similarly, a non-calibrated system attempting to count a number of pill boxes may count both garbage cans and actual pill boxes as pill boxes. To avoid this sort of error, VCA systems can be calibrated to provide scale and determine the actual sizes (e.g., actual height and width dimensions) of objects, which improves analysis accuracy.
As one example,
However, even a calibrated camera system can have some difficulties detecting real-world objects. For example, to determine an actual height of an object, such as a person, a calibrated system may search for the top of the object (e.g., the person's head) and the bottom of the object (e.g., the person's feet). However, part of a person's body, including the feet may be occluded by one or more objects, such as, for example, another person, or a shopping cart. In this case, the system may not be able to detect certain information about the person, such as the person's height. Similarly, if a person is lying down and a calibrated camera system detects the person's feet and head, it may determine a height of the person, but not a height above the floor at which the person is positioned. Therefore, the calibrated camera system may not be able to detect that a person is lying down after performing object detection.
As another example, if a second person is standing behind a first person, even if the system detects the second person, for example, based on an algorithm that detects human heads or faces, the system may not necessarily know the height of the second person. The second person may be taller than the first person and standing very close to the first person, or the second person may be shorter than the first person, but standing further away from the second person. In either case, however, the camera only sees the first person and the second person's head just above the first person.
Another example where a calibrated system may erroneously detect people or other objects is when shadows or reflections are involved, or where people are occluded by immobile objects, such as furniture, for example. A calibrated camera system may see a shadow or reflection, and may determine, erroneously, that it is an actual person, or may see two parts of a person separated by an immobile pole or other obstruction, and may fail to recognize both parts of the person as a single person.
To remedy these problems, particularly in home health care monitoring where patients will often be sitting, lying down, obstructed by furniture, etc., in one embodiment, a depth sensor is used together with the calibration information to help determine the real world height or size of a person or object and therefore to help determine a position of the person or object. The depth sensor information can be used to supplement, or verify information collected or determined by the calibrated camera system, and/or to help determine a position of a detected person or object.
As opposed to inferring distance based on geometric equations, certain depth sensors determine the distance of objects from a sensor device by obtaining a direct measurement. For example, the measurement may be made using an infrared projector and a monochromatic CMOS sensor. An exemplary system for determining depth of objects in a three-dimensional space is described in U.S. Patent Application Publication No. 2010/0199228, to Latta et al., published on Aug. 5, 2010, the contents of which are incorporated herein by reference in their entirety. However, depth determination is not limited to the method disclosed in Latta et al., and depth can be determined based on a plurality of different sources, such as lidar, stereopsis, or structured light, for example.
In one embodiment, depth information can be used to supplement camera image information to determine the identity of certain objects. For example, in one embodiment, camera image information can be used to determine all potential human beings in a camera's field of view. For example, a calibrated camera system may be configured to detect objects that are not part of the background (e.g., moving objects) and that have a shape approximately the same shape as a human being. Depth sensor information can then be used to determine a real-world height or size of each object detected as a potential human being, and as a result, the number and location of actual human beings, as well as their positioning, can be more accurately determined, for example, based on the potential human being objects that are at certain heights or that occupy a certain threshold volume. In another embodiment, the depth sensor information can be used as a filter to count certain groups of people, for example, if only adults are desired to be counted.
Many methods have been proposed on using depth data to perform scene analysis. In U.S. Pat. No. 8,238,607 and U.S. Patent Application Publication No. 2012/0314905, for example, stereo videos are used to generate disparity map and depth map, and human detection and tracking are performed on the computed depth map. In U.S. Pat. No. 7,831,087, “Plan-View” images are generated from both depth data and non-depth data, and object detection is performed on the “Plan-view” images through “Plan-view” templates. In U.S. Pat. No. 8,320,621 and U.S. Patent Application Publication No. 2012/0197393, a new 3D imaging device RGBD sensor is introduced which can provide both RGB and depth components for each pixel on the image. Humans and human body parts are detected and tracked on the depth map. In U.S. Patent Application No. 2005/0201612, stereo images are used to produce a height map, the human objects are detected by detecting heads using connected component analysis on the height map. In U.S. Patent Application Publication No. 2012/0293635, the above RGBD sensor is used to detect the head pose, and the head position and orientation are estimated by tracking head feature points in 3D space.
Most of the prior art performs the object detection and tracking in the depth space or 3D space. This usually results in a lower resolution and lost details on the objects of interest. Further, the accuracy and quality of the depth data is usually not as good as those RGB image data, and methods of how to deal with the noise and incompleteness of the depth data in the scene analysis have not been well addressed. In addition, processing for object detection and tracking using 3D space data for a whole scene can be computationally complex or even prohibitive. In the present application, a way to use aligned depth data to assist in object detection/tracking under the existing non-depth sensor framework is proposed. The approach is based on the existing RGB image sensor based framework, and uses additional depth information to solve or alleviate certain existing problems. The object detection and tracking is still performed on the traditional non-depth 2D image space, and the depth data is used to provide physical location and size information on objects of interest to help the object detection, segmentation, classification and tracking processes. As a result, particularly when used in a home health care environment, a video content analysis system can perform more accurate and reliable event detection, and can monitor patients in a way not previously attainable.
In one embodiment, the RGB components 304 may be processed by existing video content analysis algorithms, such as like described in U.S. Pat. No. 7,825,954, to Zhang et al., published on Nov. 2, 2010, the contents of which are incorporated herein by reference in their entirety. As such, the system may analyze the 2D (e.g., RGB) components 304 to first perform motion and change detection (step 305) to separate foreground from background. For example, in one embodiment, pixels that are detected as moving are indicated to be foreground data (e.g., by being labeled with a logic value, such as “1”), and pixels detected as non-moving are indicated to be background data (e.g., by being labeled with a different logic value, such as “0”). The output of step 305 may include a foreground mask for each frame. Next, the foreground regions may be divided into separate blobs by blob extraction (step 306). During blob extraction, in one embodiment, the individual foreground pixels are grouped spatially. Foreground pixels that are touching or close to each other are assumed to correspond to the same object and are combined into a single blob. As a result, for each frame, one or more blobs may be detected. Each blob or a part of each blob may correspond to one or more targets at each timestamp (where, for example, a particular timestamp may be associated with a frame of the video). In target tracking step 307 targets may be detected based on the blobs extracted in step 306, and each target may be tracked, where each target may correspond to an object in the scene that is desired to be tracked. The depth component 303 is used here to provide a more accurate determination of which blobs correspond to targets, as opposed to, for example, which blobs correspond to objects that are not targets and do not need to be tracked. Additionally, the depth component 303 may be used to better distinguish different targets from each other, or to determine a position of a target.
For example, in one embodiment, a blob may be detected, and depth information may be used to determine its size or volume. Alternatively, or additionally, facial recognition may be applied to the blob. The blob may correspond to a person. Based on the size or volume determination, and/or the facial recognition, the video content analysis system may determine that the blob is a person, and may indicate the person to be a target. Subsequently (or at the same time), the video content analysis system may use the depth information to determine a height of the person or the height of a face of the person where facial recognition is used, and therefore may estimate a position of the person based on the height (e.g., sitting, lying down, standing). This may occur even if part or all of the person's body is occluded, for example, by covers on a bed, or by the back of a chair. The person may be tracked based on changes in position, such as sitting up after lying down, getting out of bed, etc.
Finally, after a target and optional additional information about the target is detected, event detection step 308 performs event detection based on user-defined rules and the targets detected and tracked in step 307. In the embodiments discussed herein, video content analysis can be performed in real-time, or may be performed on video sequences stored previously, for example, by a DVR, NVR or other recording equipment attached to a camera, or in a central computer system
In one embodiment, though not shown in
As a result of the above steps, the following method may be performed. First, a video sequence that includes a plurality of frames may be captured, for example, by an RGBD sensor, such as a camera having depth detection capabilities. Each frame may include a video image. For each frame, two-dimensional (2D) image data (e.g., RGB data) may be extracted, and also depth data may be extracted. The 2D image data and depth data may then be transmitted to and received by a video content analysis system (e.g., one or more processors executing one or more algorithms for analyzing video content). The 2D image data of the video sequence may then be processed to differentiate foreground data from background data and to detect one or more blobs comprised of the foreground data. The one or more blobs may correspond to one or more real-world objects, and correspond to one or more potential targets. For each detected blob, the depth data may be used to determine whether at least part of the blob corresponds to at least part of a target, or to determine whether to track at least a part of the blob as a target. For example, it may be determined that an entire first blob corresponds to a single real-world object (e.g., single person), and so that the first blob is determined to correspond to a first target. Alternatively, it may be determined that a first blob actually corresponds to two different real-world objects (e.g. two people), and so part of that first blob is determined to correspond to a first target, and another part of the first blob is determined to correspond to a second target. In a third case, a blob may be determined to correspond to only part of a real-world object, and so that blob and an additional blob may collectively be determined to correspond to a single target (e.g., a single person).
After it is determined that at least part of a blob corresponds to at least part of a target, the target is tracked and at least one event associated with the target is detected.
Stated in a different way, as discussed in the examples above, a video sequence may be received that includes a plurality of frames, each frame including a video image. For each frame, image data of the video image and also depth data associated with the video image may be received (e.g., it may be extracted from the video sequence and received by a video content analysis system). The image data may then be analyzed to detect one or more objects depicted in the video sequence (e.g., a blob may be extracted, and the system initially assumes that the blob corresponds to a real-world object in the video sequence, for example, by treating the blob as a potential target). Next, using the depth data along with the one or more detected objects, at least a first object of the one or more detected objects is classified as an object to be tracked. For example the first object may be classified as a person to be tracked, an adult to be tracked, a vehicle to be tracked, etc. The object to be tracked may be treated as a target. Next, tracking is performed on at least the first classified object. For example, based on depth data, a person maybe tracked and it may be determined if the person is lying down, standing up, or sitting. Finally, event detection analysis is performed on the first classified object. In certain embodiments, the video content analysis described above is automatically performed by a computer system, such as a video content analysis system.
In one embodiment, the depth data 303 is mainly used in step 307 to help the target detection and tracking processes. The inputs to step 307 may be foreground image blobs extracted from the video frames based on change and motion detection. Each image blob may include a group of connected foreground pixels representing all or part of a physical object, or multiple physical objects. A correct understanding on what each image blob represents is important for the overall system performance. The disclosed embodiments use the depth data to help make the correct decision in step 307 regarding which targets to track.
As shown in
In one embodiment, the physical properties associated with an image blob are estimated by mapping some samples of the image pixels in the blob onto a number of Z-planes in 3D space as illustrated in
One advantage of the disclosed embodiments is that not every pixel in the RGB image needs to be mapped onto the Z-planes. For example, in one embodiment, only the foreground pixels that represent the image blobs are to be projected onto the discrete Z-planes, and background pixels do not need to be projected onto Z-planes. In addition, because the number of Z-planes mapped is quantized, not every pixel associated with a blob needs to be projected onto a Z-plane. Further, as described further below, convex hulls may be used to represent the object regions on Z-planes. One convex hull may be approximated by a few pivot points, and not every pixel of a blob in a particular Z-plane needs to be sampled in order to form the convex hull. Thus pixel sampling may be performed for each frame and within each image blob to further reduce the computational complexity. In addition, this approach relies less on the accuracy and completeness of the depth data on every image pixel, and is thus more robust despite inaccuracies that may be associated with the depth information.
The samples on a Z-plane mapped from the corresponding pixels from the same image blob may form different spatial regions on the Z-plane because they may correspond to spatially separated objects.
A physical distance threshold may be used to determine whether a sample point outside the cluster should belong to the cluster. Thus the clustering process can be described as follows. Given a list of sample points on a Z-plane which are mapped from sample pixels from an image blob, select a first sample and consider it as the first sample cluster. Then iterate through all the remaining sample points. For a given sample point, compute its distance to all the existing blob clusters. If the distance to a cluster is less than a distance threshold predetermined as a parameter, update this cluster by including this sample into the cluster convex hull. If one sample belongs to multiple clusters, merge all the corresponding cluster convex hulls into a new cluster. If a sample does not belong to any existing clusters, create a new cluster using the current sample. The exemplary method is a one-pass clustering process, and the distance computation only involves a limited number of pivot points. As a result, the clustering process, and the resulting target detection and tracking is computationally efficient.
Since a typical RGBD camera is not able to see through an object, a self-occlusion issue often occurs in the 3D space representation of an object.
In certain embodiments, a camera position may be directly above part of an object, or almost directly above the object.
The physical volume measurement may be used, for example, to perform target filtering and target classification. For example, it can increase the confidence on detecting a human object. A human blob should have a physical volume close to an average physical human. The change of human postures will change the image appearance and height but typically will only have small impact on the human volume. Meanwhile, the human pose change can be detected by tracking the changes of physical height and the projected areas on different Z-planes. The physical height and volume measurements can also be used to distinguishing different types of people from others, such as children from adults.
The physical volume measure may also be used to filter out spurious foreground blobs caused by illumination factors, such as shadows and reflections. These types of non-legitimate blobs usually have little physical volume. The physical height and volume information can be used to detect other types of targets such as vehicles or shopping carts, for example. The physical sizes at different Z-planes are strong cues to detect objects with different physical size and shapes. Just using a height map without volume information may incorrectly detect certain blobs, such as a shadow on a wall, as a person.
Due to the camera viewing perspective, multiple targets not close to one another may be connected in an RGB image and appear as a single blob. In one or more embodiments, they can be separated in the RGB image by using the depth data.
In some other scenarios, a single foreground object may be occluded by a static background object, or part of the object looks so similar as the background that the system may miss-detect that part as foreground. For example, a person may be occluded by part of a chair or by a bed post. When these happen, the RGB-based system will likely break a single image object into multiple image blobs. This type of problem may also be solved by the depth analysis.
One example of a general application of the combined calibration and depth detection system is shown in
In situation 320, however, one person stands in a space, but the person's shadow also appears on the floor. Because the depth detection can be used to remove the effects of the shadow (e.g., as discussed above), the shadow can be omitted from the mapping of people in the space in the depth map. Similarly, in situation 330, one person is partially occluded by a shopping cart, which also has a round object that may be detected as a potential person's head. However, after a height threshold is applied, the round object is assumed to not be a person and is not tracked, and the person is determined to be an actual person and can be tracked. As a result, only one person is included in the mapping of people after the height threshold has been applied. In each of these examples (320 and 330) a vision only person counting system (without depth detection) may have counted two people, thus over-counting the number of people in two of the examples.
After objects are identified as targets, those targets may be tracked within a scene in a video. However, because of the height mapping, the tracking may be analyzed from a top-down, two-dimensional perspective, even though there is no camera capturing images from a top-down view looking directly down at the scene. In one embodiment, a standard Kalman filter can be used to track the location of each object.
Event detection can then be performed based on the detected objects and their tracking information. For example, a virtual tripwire, as described in U.S. Pat. No. 6,696,945, issued to Venetianer et al. on Feb. 24, 2004, the contents of which are incorporated herein by reference in their entirety, can be used to perform counting of one or more people moving in or out of a certain area. An example of a virtual tripwire is shown in
Another example of a home health care application of the above embodiments is to perform object tracking to determine when a patient falls down. For example, a captured image may have the shape and size of a person, but the depth information showing that the person's head is near to the ground (e.g., one foot off the ground), may indicate that a person has fallen down or is lying down. One or more alarms can then be triggered based on an event of a person falling being detected. In one embodiment, the person can be mapped into a two-dimensional overhead view as long and narrow, as shown in
A method of performing video content analysis (VCA) using the disclosed depth sensing VCA system is shown in
As shown in
Singular Events:
Collective Events:
Many other events can be set. Groups of events can be set, and rules can be set based on the events. Events can be set to include information from different rooms monitored by different cameras in a house. For example, an event of lying down for X hours can be set to correspond to all of the cameras in a house so that whether a patient lies down on a couch or bed, the hours are counted. As discussed below, many of the events are more accurately determined by using depth data along with two-dimensional image data to perform depth-enhanced video content analysis.
Events can be relative to observed or learned patterns. For example, a depth-enhanced video content analysis monitoring system may learn that a patient opens a box, moves his hand toward his mouth, and then drinks a glass of water every morning around 9 a.m. The system may associated these patterns with a “patient takes medications” event. Events can be associated with different people. For example, if an elderly married couple lives in a house, a video content analysis monitoring system can recognize the different people and monitor for certain events based on the person. They can be recognized based on e.g. facial recognition, gender recognition, voice recognition, height, hair, etc. Similarly, if guests are present, they may be differentiated from the monitored people.
Events such as lying down, standing up, sitting, and falling, for example, can be more readily determined using depth data in combination with two-dimensional data, as discussed previously. For example, to detect a person lying in a bed, two-dimensional data of video frame can be analyzed to determine first that a person is in the bed. This may occur based, for example, on shape recognition, face recognition, and/or volume calculations. However, it may not be clear whether the person is lying down or sitting up in the bed. For example, the person's legs may be under the covers, and so only an upper torso and head of the person is detected. The two-dimensional data analysis may not be able to determine whether the torso and head is vertical or horizontal. As such, depth data may be used to determine a height of the person's head, for example. The height of the person's head may then be compared to the height of the bed, to determine whether the person is sitting up or lying down. Another position that may be difficult to distinguish is sitting up versus a person having their knees up.
In step 1702, a plurality of video sequences are monitored from a plurality of video cameras. For example, each of cameras 112, 122, 132, and 142 in
In step 1703, video content analysis steps are carried out on the two-dimensional image data to detect objects in the video sequences. For example, using analysis techniques such as facial recognition and shape analysis, the objects can be identified as particular targets. For example, a person in the video sequences can be identified as a person, or as a particular patient.
In step 1704, depth data is used to confirm and/or enhance/supplement information about the detected targets. For example, in one embodiment, depth data may be used to determine a height of the target, which in turn can be used to determine a position of the target (e.g., sitting, lying down, standing up). Although step 1703 is depicted as occurring before step 1704, in one embodiment, steps 1703 and 1704 may occur simultaneously, such that two-dimensional analysis is performed in conjunction with depth data confirmation to perform depth-enhanced video content analysis.
In step 1705, based on the information obtained in step 1704, an event may be detected (e.g., patient sits up, patient falls, etc.). A singular event may be detected in a first instance. In addition, a plurality of singular events may be combined to detect a collective event (e.g., person lies down for X hours in a day; person leaves house for X hours during the day, etc.). Based on the events detected, certain rules can be set that trigger alerts or alarms. For example, if a person is detected as lying down for at least X hours in a day (e.g., “person lies down for X hours in a day” event is detected), an alarm may be triggered because this may indicate that the person is having a health-related problem. Different types of and severity levels of alarms can be set based on different types of events.
To implement the system and methods described herein, various computing and optical components may be used, such as one or more of the following: a general purpose computer; supercomputer; a mainframe; a super mini-computer; a mini-computer; a workstation; a micro-computer; a server; an interactive television; a hybrid combination of a computer and an interactive television; a smart phone; a tablet; and application-specific hardware to emulate a computer and/or software. These may include one or more processors, one of more field programmable gate arrays (FPGAs), computer memory, a computer-readable medium such as, for example, any storage device used for storing data accessible by a computer (e.g., a processor may perform various algorithms on data received from a camera device, and a computer memory can then store the information about the various pixels and can store results of blob detection, target detection, and event detection). Examples of a computer-readable medium include: a magnetic hard disk; a floppy disk; an optical disk, such as a CD-ROM and a DVD; a magnetic tape; a memory chip; a solid state storage device; and a carrier wave used to carry computer-readable electronic data, such as those used in transmitting and receiving e-mail or in accessing a network. A tangible computer-readable medium includes computer-readable media, such as listed above, that are physically tangible. In addition, software may be used in combination with the computing and optical components to implement the methods described herein. Software may include rules and/or algorithms to operate a computer, and may include, for example, code segments, instructions, computer programs, and programmed logic. The various computers, cameras, and other image equipment described herein can be connected over a network, which may involve permanent connections such as cables or temporary connections such as those made through telephone or other communication links, and also may include wireless communication links. Examples of a network include: an internet, such as the Internet; an intranet; a local area network (LAN); a wide area network (WAN); and a combination of networks, such as an internet and an intranet. The various hardware and software examples described above are also described in greater detail in the patent documents incorporated by reference herein.
In an exemplary embodiment, depth data and a human model are used to more reliably detect a human arm and its movement. For example, reliable detection of an arm can be used to better detect when a patient reaches for certain items, such as reaching for a medicine cabinet for certain medications. Detecting arm motion only based on monocular vision is error prone, due to shadows, arms or clothing blending into the background, etc. The proposed combined system with the addition of depth and 3D information may significantly improve performance for these applications.
In one embodiment, the depth sensing VCA system can be used to better count and differentiate objects in different groups. For example, the system can use height thresholds to differentiate between adults and children. This may be useful in a home environment, for example, if children visit their grandparents, or if elderly patients have pets, in order to track only the grandparents for health monitoring, and not the children or animals.
The foregoing is illustrative of example embodiments and is not to be construed as limiting thereof. Although a few example embodiments have been described, those skilled in the art will readily appreciate that many modifications are possible in the example embodiments without materially departing from the novel teachings and advantages of the present disclosure.
Venetianer, Peter L., Zhang, Zhong, Myers, Gary W.
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
5544649, | Mar 25 1992 | CARDIOMEDIX, INC | Ambulatory patient health monitoring techniques utilizing interactive visual communication |
5553609, | Feb 09 1995 | Indiana Research and Technology Corporation; Indiana University Research and Technology Corporation | Intelligent remote visual monitoring system for home health care service |
6792319, | Oct 19 2000 | HOME DIRECTOR, INC | Home automation system and method |
7516888, | Jun 21 2004 | NCR Voyix Corporation | Method and apparatus for auditing transaction activity in retail and other environments using visual recognition |
7801330, | Jun 24 2005 | MOTOROLA SOLUTIONS, INC | Target detection and tracking from video streams |
7825954, | May 31 2005 | MOTOROLA SOLUTIONS, INC | Multi-state target tracking |
7831087, | Oct 31 2003 | Hewlett-Packard Development Company, L.P. | Method for visual-based recognition of an object |
7868912, | Oct 24 2000 | MOTOROLA SOLUTIONS, INC | Video surveillance system employing video primitives |
7932923, | Sep 29 2009 | AVIGILON FORTRESS CORPORATION | Video surveillance system employing video primitives |
8233660, | Jan 16 2009 | HONDA RESEARCH INSTITUTE EUROPE GMBH | System and method for object motion detection based on multiple 3D warping and vehicle equipped with such system |
8238607, | Jul 03 2007 | ShopperTrak RCT LLC | System and method for detecting, tracking and counting human objects of interest |
8320621, | Dec 21 2009 | Microsoft Technology Licensing, LLC | Depth projector system with integrated VCSEL array |
20040153671, | |||
20050201612, | |||
20070070190, | |||
20070127774, | |||
20080021731, | |||
20090063307, | |||
20090281392, | |||
20100197393, | |||
20100199228, | |||
20110080336, | |||
20110143779, | |||
20110200229, | |||
20120020518, | |||
20120025989, | |||
20120026289, | |||
20120026308, | |||
20120075464, | |||
20120087572, | |||
20120140068, | |||
20120293635, | |||
20120314905, | |||
20130041290, | |||
20130073093, | |||
20130182114, | |||
20130182904, | |||
20130182905, | |||
20130184887, | |||
KR100234196, | |||
WO2012037157, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jan 17 2013 | AVIGILON FORTRESS CORPORATION | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Date | Maintenance Schedule |
Nov 10 2018 | 4 years fee payment window open |
May 10 2019 | 6 months grace period start (w surcharge) |
Nov 10 2019 | patent expiry (for year 4) |
Nov 10 2021 | 2 years to revive unintentionally abandoned end. (for year 4) |
Nov 10 2022 | 8 years fee payment window open |
May 10 2023 | 6 months grace period start (w surcharge) |
Nov 10 2023 | patent expiry (for year 8) |
Nov 10 2025 | 2 years to revive unintentionally abandoned end. (for year 8) |
Nov 10 2026 | 12 years fee payment window open |
May 10 2027 | 6 months grace period start (w surcharge) |
Nov 10 2027 | patent expiry (for year 12) |
Nov 10 2029 | 2 years to revive unintentionally abandoned end. (for year 12) |