THE proportion of the elderly population is rising rapidly
|
|
|
- Barnaby Armstrong
- 9 years ago
- Views:
Transcription
1 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS Fall Detection Based on Body Part Tracking Using a Depth Camera Zhen-Peng Bian, Student Member, IEEE, Junhui Hou, Student Member, IEEE, Lap-Pui Chau, Senior Member, IEEE, and Nadia Magnenat-Thalmann Abstract The elderly population is increasing rapidly all over the world. One major risk for elderly people is the fall accidents, especially for those living alone. In this paper, we propose a robust fall detection approach by analyzing the tracked key joints of the human body using a single depth camera. Compared to the rivals that rely on the RGB inputs, the proposed scheme is independent of illumination of the lights and can work even in a dark room. In our scheme, a pose-invariant Randomized Decision Tree (RDT) algorithm is proposed for the key joint extraction, which requires low computational cost during the training and test. Then, the Support Vector Machine (SVM) classifier is employed to determine whether a fall motion occurs, whose input is the 3D trajectory of the head joint. The experimental results demonstrate that the proposed fall detection method is more accurate and robust compared with the state-of-the-art methods. Index Terms Computer vision, 3D, monocular, video surveillance, fall detection, head tracking. I. INTRODUCTION THE proportion of the elderly population is rising rapidly in most countries. In, the elderly population (6+ years old) is 759 million ( percent of the total population) all over the world []. Many studies have indicated that falls in elderly people are one of the most dangerous situations at home []. Approximately 8-35% of elderly people fall one time or more per year [3]. When an elderly person is living alone and has a fall accident, he/she may be lying on the floor for a long time without any help. This scenario mostly will lead to a serious negative outcome. Therefore, a fall accident detection system, which can automatically detect the fall accident and call for help, is very important for elderly people, especially for those living alone. In [], [4], [5], the authors reviewed principles and methods used in existing fall detection approaches. Nowadays, fall detection approaches could be classified as two main categories: non-vision-based method and vision-based method. Most methods of fall detection employ inertial sensors, such as accelerometers, since they are low costs. However, the methods based on inertial sensors are intrusive. As the vision technologies developed fast during the past few years, the Z.-P. Bian, Junhui Hou and L.-P. Chau are with the School of Electrical and Electronics Engineering, Nanyang Technological University, Singapore ( [email protected], [email protected], [email protected]). N. Magnenat-Thalmann is with the Institute for Media Innovation, Nanyang Technological University, , Singapore ( [email protected]). vision-based methods, which are non-intrusive, have become a focal point in the research of the fall detection. They can capture the object s motion and analyse the object environment and their relationship, such as the human lying on the floor. The feature of the vision-based systems can be posture [6], [7], [8], [9], shape in-activity/change [], [], spatio-temporal [], 3D head position [], [3], and 3-D silhouette vertical distribution [4]. To improve the accuracy, some researchers combined non-vision-based method and vision-based method [5]. Most of the fall detection methods based on vision try to execute in real-time using standard computers and low cost cameras. The fall motion is very fast, taking few hundred milliseconds, and the image processing is high computational complexity. Thus, most vision-based existing methods cannot capture the specific motion during the fall phase. Recent researches on fall detection based on computer vision showed some practical frameworks. However, the robustness as well as accuracy of vision-based methods still leave a wide open room for further fall detection research and development. The depth camera, such as Kinect [6], was used for fall detection [7], [8], [9], []. Thanks to the infra-red LED, the depth camera is independent of illumination of lights and can work well in weak light condition even in a dark room. It can also work well when the light condition significantly changes such as switching on or off the lights. As we know, some falls are caused by the weak light condition. In the depth image, each pixel value represents the depth information instead of the traditional color or intensity information. The depth value is the distance between the object and the camera. The depth information can be used to calibrate each pixel to a real world 3D location point. Compared with the traditional intensity or color camera, the depth camera provides several useful advantages in the object recognition. Depth cameras are useful for removing ambiguity in size scale. The object size in the color or intensity image is changed according to the distance between the object and the camera. That introduces ambiguity in size scale since the distance is unknown. In color or intensity images, the shadow greatly reduces the quality of background subtraction. Depth cameras can resolve silhouette ambiguity of the human body. The depth cameras simplify the tasks of background subtraction and floor detection. That can improve the robustness of the object recognition and can offer some useful information about the relationship between the human and the environment, such as the human hitting the floor. Furthermore, the realistic depth images of human can be much easier to synthesize. Therefore, a large and
2 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS Depth camera Fall alert Training data Depth image Fall confirmation Fast training algorithm Pose correction Fig.. Flow chart of proposed fall detection. RDT Extract and track the human joints Fall motion analysis by classifying the head joint trajectory based on SVM reliable training dataset can be built up for markerless motion capture []. In this paper, a robust fall detection system based on human body part tracking using a depth camera is proposed. To capture the fall motion, an improved randomized decision tree (RDT) algorithm is employed to extract the 3D body joints. By tracking the 3D joint trajectory, a support vector machine (SVM) classifier for fall detection is proposed. The 3D joint trajectory pattern is the input of the SVM classifier. Fig. shows the flow chart of proposed fall detection algorithm. The structure of this paper is as follows. Section II introduces the joint extraction method. Section III introduces the fall detection methods based on SVM. Experimental results are presented in Section IV. Conclusions are presented in Section V. II. JOINT EXTRACTION To capture the human fall motion, a markerless joint extraction is employed. The joint extraction is based on the proposed RDT algorithm, which is trained by large depth images dataset. A. The feature for joint extraction In [], the recognized feature is based on the difference of intensities of two pixels taken in the neighbourhood of a key point. This feature was further developed in the depth image by Shotton et al. in []. They employed a simple depth comparison feature instead of the intensity comparison feature, resulting in a wonderful success. Only one or two comparison features are very weak for discriminating objects, such as discriminating body parts. However, a RDT based on these comparison features are sufficient to discriminate objects. It can handle the noise of the depth image and can even work using the D silhouette []. The formulation of comparison feature in [] can be described as: f((x, y ) ( x, y ), ( x, y )) = z((x, y ) + ( x, y ) z(x, y ) ) z((x, y ) + ( x, y ) z(x, y ) ) () where (x, y ) is the test pixel of the depth image, ( x, y) is the offset related to the test pixel (x, y ), ( x, y ) and ( x, y ) are two different offset values, z(x, y) is the depth f ³ t L f <t value of the pixel (x, y). /z(x, y) is used to normalize the offset value, so that the feature can resolve the ambiguity in depth variation. As shown in Equation, this feature is three dimension translation invariant. The depth information around the test pixel describes the geometric surface around this pixel. The geometric surface around the test pixel can be described well enough by the depth differences between the neighbour pixels and the test pixel. Thus, the following feature can be used to describe the same geometric surface as Equation to recognize the test pixel: ( x, y) f((x, y ) ( x, y)) = z(x, y ) z((x, y ) + z(x, y ) ) () Compared with Equation, this feature is with a higher computational efficiency: there are only one division, one addition and one subtraction; it only looks up two depth pixels. That can save time and leave more time for real-time fall detection. f ³ t Fig.. Randomized decision tree. L f <t Fig. shows a randomized decision tree consisting of split and leaf nodes. A randomized decision forest includes several randomized decision trees. In many applications, the task of multi-class classifiers can be implemented in a high efficiency and high speed by RDT [], []. In each tree, each split node n has a parameter p n = (( x n, y n ), τ n ). ( x n, y n ) is the offset value. τ n is a scalar threshold for comparing with the feature value of the test pixel. The evaluating function of comparison is E((x, y ); p n ) = B(f((x, y ) ( x n, y n )) τ n ) (3) where B( ) is a binary function. When the value of (f((x, y ) ( x n, y n )) τ n ) is greater than or equal to, there will be B(f((x, y ) ( x n, y n )) τ n ) = ; otherwise B(f((x, y ) ( x n, y n )) τ n ) =. When E((x, y ); p n ) is one, the test pixel is split to the left branch child of node n. When E((x, y ); p n ) is zero, the test pixel is split to the right branch child of node n. The operation is repeated, and stop when it meets the leaf node l. There are some classification information in the leaf node l, such as the probability of body parts. To classify a pixel (x, y ), one starts at the root and repeatedly evaluates Equation 3, branching left or right according to the value of E until reaching the leaf node. After classifying each pixel, the joint position can be predicted by body part classification []. In the body part classification method, the joint position is presented by the mean of the same class pixels. Since the head and hip, which
3 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS 3 are used in the proposed fall detection, are the most visible body parts, we can use body part classification to extract them. Fig. 3 shows a classification result of the human body. Each color in Fig. 3 stands for the respective part of the human body with the highest probability. x y Entropy x 5 y 5 Fig. 4. map. Entropy map and 3D view. Map of entropy. 3D view of the Fig. 3. The classification result of each depth pixel, each color stands for the respective part of the human body with the highest probability. Blue: head, red: hip, green: other part. B. Aggregating predictions The result of RDT algorithm is the classification of each pixel. Then the joint position is found by a mean shift method. The mean shift is based on a Gaussian kernel and a weighting factor. For the mean shift, the density estimator [] of each body part c is defined as J c (L) N i= w ic exp( L L i b c ) (4) where c is the body part label, L is the 3D location in the 3D real world, w ic is a pixel weighting, N is the number of total test pixels in the test image I, L i is the 3D location of the test pixel (x i, y i ). b c is a bandwidth for body part c. w ic includes two factors: () the probability P from the RDT algorithm; () the world surface area related to the depth value z. The formulation of w ic is w ic = P (c (x i, y i ), I) z(x i, y i ) (5) This mean shift method result is more accurate than the result of the global centre of the same body part since there are some outlying pixels as shown in Fig. 3. C. Training In order to obtain an optimised parameter p n of each split node n of RDT, the computation complexity was very high in [], [3]. Based on Equation we propose a fast training algorithm which can reduce the number of candidate offsets significantly in the training phase. The training pixels with labels for each synthesized depth image are randomly down sampled. The tree is trained using the smallest Shannon entropy to split each node. At each node, a weak learner parameter p(( x, y), τ) ( ( x, y) ( X, Y ) is the pixel offset, τ T is the threshold value in Equation 3.) induces a partition of input example set Q = {I, (X, Y )} into left and right subsets by Equation 3: Q L (( x, y), τ) = {E((x, y ); p(( x, y), τ)) = } (6) Q R (( x, y), τ) = {E((x, y ); p(( x, y), τ)) = } (7) For each offset ( x, y), compute the τ giving the smallest Shannon entropy: τ = arg mins(q(( x, y), τ)) (8) τ T S(Q(( x, y), τ)) = Q sub (( x, y), τ) H(Q sub (( x, y), τ)) Q sub L,R where H(Q) is the Shannon entropy (computed on the probability of body part labels) of set Q. S(Q) is the sum of Shannon entropy. Fig. 4 is a map in a node by drawing Shannon entropy (given by p(( x, y), τ )) on the corresponding location ( x, y). Fig. 4 is a mesh view of Fig. 4. From Fig. 4, it can be noted that the Shannon entropy surface is smooth. The smallest Shannon entropy in this surface can be efficiently searched out by some search algorithms. Thus, just a few offsets need to be trained by Equations 6, 7 for each node. Equations 6, 7 should be tested by the whole set Q = {I, (X, Y )} of input examples in a node, and this operation takes a long training time. Therefore, using Equation and a suitable search algorithm, it can dramatically save training cost. In contrast, the feature in Equation requires two offsets, which require randomly sampling candidate offset pairs among (M + ) 4 candidate offset pairs in [], [3], where M is the range of x, y, i.e., x, y [ M, M]. The parameter p with the smallest Shannon entropy in all candidate offsets and thresholds is p(( x, y ), τ )= arg min ( x, y) ( X, Y ),τ T (9) S(Q(( x, y), τ)) = arg min {min S(Q(( x, y), τ))} ( x, y) ( X, Y ) τ T () If the depth of the tree is not too large, the training algorithm recurs for the left and right child nodes with example subsets Q L (( x, y ), τ ) and Q R (( x, y ), τ )), respectively, according to p(( x, y ), τ ). Table I demonstrates the performances of the algorithms based on [] and ours. Based on Equation and a search
4 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS 4 Fig. 6. After rotation, the head and hip centre joints can be extracted correctly. Fig. 5. Joint extraction results. algorithm, it takes less than four hours to train one tree with twenty levels from ten thousand images on a standard PC using one core by Matlab. There are training example pixels in each image, 33 candidate thresholds per offset and 4 dynamic candidate offsets. It takes 3 hours with pairs of candidate offset based on Equation. images with ground truth of the head joint are used to test the performance of these two algorithms. The test is implemented on a standard PC using one core by C++. Based on Equation, the test time per frame is.8 ms. Based on Equation, the test time per frame is 5. ms, which is 79% more than.8 ms. The mean error is measured on the head joint. Their mean errors, i.e., 3. cm (Equation ) and 3. cm (Equation ), are very similar without substantial difference. joint extraction process in the proposed fall detection system. Thanks to the high output frame rate of the depth camera, it is fast enough to track the human motion during falls. The torso orientation can be defined by the straight line through the hip centre and the head. The person s torso orientation changes very little per frame even during the fall. Thus, the torso orientation can be corrected well enough by rotating the torso based on the pose in the previous frame. Therefore, the joint extraction is pose-invariant. When the person walks into the view field at the first time, the head and hip centre are extracted. This information is used to rotate the torso orientation for the next frame, and the rotation angle is updated frame by frame. As shown in Fig. 7, when the person is falling down (Fig. 7), the torso orientation can be always rotated to be vertical in the image (Fig. 7) based on the previous frame. TABLE I COMPARISON OF ALGORITHMS BASED ON [] AND OURS. Training time per tree (hour) Test time per frame (ms) Mean error (cm) [] Ours D. Pose correction In order to track the joint trajectory well, the frame rate of the camera output should be high and the joint extraction should be fast. The frame rate of Kinect is 3 frames per second and the joint extraction based on RDT takes a few milliseconds. Thus, the joint trajectories can be tracked well. Based on RDT algorithm, an open license software has been released, i.e., Kinect for windows SDK [6]. Fig. 5 shows that the human joint extraction from the SDK is correct when the person is standing even moving a chair at the same time. However, the SDK joint extraction cannot work well under some scenarios, such as the person lying on a sofa as shown in Fig. 5. The degraded accuracy problem also appears in the moment of falling down due to the human body orientation changes dramatically while falling. We propose to rotate the depth image in Fig. 5 by 9 degree clockwise, and the key joint positions can be extracted accurately as shown in Fig. 6. As inspired from this, the human torso is always rotated to be vertical in the image before the Fig. 7. A simulated fall sequence. Original poses. After correction by proposed method. The re-initialization is required when the torso orientation tracking is lost. As shown in Fig. 8, the input image after the subject segmentation was rotated by several angles, such as, and 4, to generate several images. For each image, the head and the hip centre are extracted to obtain the torso orientation, and then the extracted torso orientation is corrected to be vertical. The extraction and correction are repeated several times. After the extractions and corrections, three candidate final rotation angles are obtained. It needs to select the best one from the three candidates. The selection metric is defined by the density estimator (Equation 4 ) during
5 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS 5 the processing of joint extraction. For simplicity, the decision can be made by using the density estimator of the head portion. Repeat Extraction & Correction Distance /m Distance /m Fig. 8. An example of re-initialization. To speed up the joint extraction, the rotation of the depth image can be taken place by rotating the offset parameter ( x n, y n ) of each split node n of tree. This rotation of parameters can be done off-line after training. The rotation angles are fixed, such as,,..., 35. III. FALL DETECTION BASED ON SVM This section describes the fall detection based on SVM, which employs the head joint distance trajectory as input feature vector. A. Fall motion analysis based on SVM Inadvertently coming to rest on the ground, floor or other lower level, excluding intentional change in position to rest in furniture, wall or other objects. is the definition of fall by World Health Organizations [4]. After the joint extraction, coming to the ground can be described in a technical word, i.e., some key joints coming to the ground. This feature can be used for fall detection. Referring to the above feature, the fall is an activity related to the floor. The floor information should be measured. In [7], they assume that the floor occupies a sufficiently large part in the image. However, this assumption is not always true such as a small bedroom with a bed. In the proposed system, the floor plane is defined by choosing three points from the input depth image when the depth camera is set-up. The floor plane equation can be described as Ax w + By w + Cz w + D = () A + B + C = () where A, B, C and D are coefficients, x w, y w and z w are real world coordinate variables. The coefficients A, B, C and D will be determined after choosing three points. Fig. 9. Two patterns of head distance trajectory from a same video sequence. Fall pattern. Non-fall pattern. The joint distance trajectory is defined as the trajectory of the distance between the joint and the floor. This signal includes the fall motion information, such as the acceleration, the joint velocity, the distance between the human joint and the floor and other hidden information. Fig. 9 shows a fall pattern and a non-fall pattern of the head distance trajectory. The fall motion can be classified by the joint distance trajectory pattern. A d dimensional feature vector is formed by the distance trajectory in d consecutive frames. d should be large enough to cover all the phases of falling including rapid movement period during the fall, the period before the fall and the period after the fall. The fall detection can be seen as two classes classification problem. SVM is very suitable for two classes classification problem. It can automatically learn the maximum-margin hyperplane for classification. The feature vector as aforementioned described can be used as the input feature vector of SVM classifier. B. Training dataset Since the 3D head joint trajectory has been tracking, the head joint motion can be analysed by the physics mechanics principle. During the falling phase, the joint motion can be seen as a free fall body. The free fall body is described as a simple formula h(t) = h + a(t t ) (3) where h(t) is the height at the time t, h is the height at the beginning of fall, a is the acceleration, t is the current time and t is the beginning time. The free fall body can be used to simulate the joint fall motion to generate fall patterns by computer. Fig. shows a free fall body trajectory fitting the head distance trajectory of a falling person. It can be noted that the free fall body curve fits well. The difference can be considered as Gaussian white noises. In order to improve the robustness, Gaussian white noises are added into the free fall body curve, as shown in Fig.. Some non-fall patterns can also be simulated by computer, as shown in Fig.. Based on the free fall body simulation, a large fall and non-fall patterns dataset can be built up. C. Fall confirmation In order to confirm the fall detection, the recover motion analysis after fall motion is required.
6 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS 6 Distance /m.5 fall free fall body 4 6 Distance /m.5 free fall body with noises 4 6 Fig.. Free fall body trajectory fits the head distance trajectory. Without noise. Free fall body with noises. Distance /m.5 walking sitting down sitting 4 6 Fig.. Non-fall patterns simulated by computer. There are two recover metrics: the heights of hip and head are higher than a recover threshold value T recover for a certain time; the height of the head is higher than a high recover threshold value T recover for a certain time. If one of these two metrics is satisfied, it means that the person is recovered. In our experiment, we set T recover =.5m and T recover =.8m. After the fall motion, if the person cannot recover within a certain time, fall detection will be confirmed. Without a fall confirmation, the fall alert will not be triggered. This stage can avoid the false alert. (6) Specificity (Sp): the capacity to detect non-fall events Sp = T N T N + F P (7) Accuracy (Ac): the correct classification rate Ac = T P + T N T P + T N + F P + F N (8) Error rate (Er): the incorrect classification rate Er = F P + F N T P + T N + F P + F N (5) (6) (7) A high sensitivity means that most falls are detected. A high specificity means that most non-falls are detected as non-fall. A good method for fall detection should have a high sensitivity and a high specificity. Besides, the accuracy should be high and the error rate should be low. B. Dataset The non-fall and fall activities were simulated as shown in Table II. They are suggested by [], but with more detailed description. During an impactive fall, the elderly person cannot keep the transient pose kneeling or sitting on the floor. Therefore, the person will lie on the floor after fall as shown in the first row and fifth row of scenarios in Table II. In Table II, Positive and means fall and non-fall, respectively. In total, there are scenarios. 5% are positive and 5% are negative. Each scenario is simulated several times. Totally, there are 38 samples. There are four subjects, and their heights, ages and weights are: 59-8 cm, 4-3 years and kg. Three are male and one is female. The experiments are in a real bedroom, as shown in Fig.. The camera is mounted.3m height on the wall. IV. EXPERIMENTAL RESULTS To test the proposed method, some normal activities (like crouching down, standing up, sitting down, walking) and falls, which are simulated by the human, have been tested. A. Performance evaluation metric The following parameters suggested by [] are used to analyse the detection results of the proposed algorithm. () True positives (TP): the number of fall events detected correctly. () True negatives (TN): the number of non-fall events detected correctly. (3) False positives (FP): the number of non-fall events detected as fall events. (4) False negatives (FN): the number of fall events detected as non-fall events. (5) Sensitivity (Se): the capacity to detect fall events Se = T P T P + F N (4) Fig.. Room plan for the evaluation. The research was approved by the Institutional Review Board of Nanyang Technological University, Singapore. C. Performance validation To test the performance of SVM classifier method, a large training dataset, which includes about k non-fall and fall patterns of head distance trajectory, is generated by computer.
7 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS 7 TABLE II SCENARIOS FOR THE EVALUATION OF FALL DETECTION []. Category Description Outcome Backward fall On the hip, then lying Positive Ending lying Positive Ending in lateral position Positive With recovery Forward fall On the knees, then lying Positive With forward arm protection Positive Ending lying Positive With rotation, ending in the lateral right position Positive With rotation, ending in the lateral to the left position Positive With recovery Lateral fall to the right Ending lying Positive With recovery Lateral fall to the left Ending lying Positive With recovery Syncope Vertical slipping against a wall finishing in sitting position Neutral To sit down on a chair then to stand up To lie down on the bed then to rise up Walk a few meters To bend down, catch something on the floor, then to rise up To cough or sneeze TABLE III THE RESULTS OF FALL MOTION DETECTION BASED ON SVM. TP TN FP FN Se(%) Sp(%) Ac(%) Er(%) After training, the SVM classifier is used to detect falls in the dataset of human simulated scenarios in Table II. The experimental results of fall motion analysis (detection) of SVM classifier is shown in Table III. For the fall motion detection, there is only eight error results, which are FN errors. The head distance trajectory during fall of an FN sample is shown in Fig. 3. There is an air mattress on the floor to protect the subject, as shown in Fig. 4. In this fall event, when the subject falls backward, the air mattress bounces the body of the subject quickly that the head distance trajectory is indicated by a circle in Fig. 3. The SVM classifier cannot have a correct decision on this head distance trajectory. If there is no air mattress, the dramatic rebound would not happen. In the fall confirmation stage, there is one fall event sample lost tracking the subject. In this sample, when the subject is lying on the floor, the system cannot segment the subject and recognises that there is no subject in the view field. The fall motion in this sample has been detected correctly by the SVM classifier. However, the system fails to confirm this fall event in the fall confirmation stage. It is due to failing in subject segmentation operated by Kinect SDK. In the fall confirmation stage, all the confirmation results are correct, except this lost tracking sample. Without a fall confirmation, the fall alert will not be triggered. Though the fall confirmation stage misses the lost tracking sample, this stage can avoid the false alert effectively. Thus, the system misses nine fall alerts, and there is no false alert. A video demonstration of fall detection based on SVM is Distance / m Time / s Fig. 3. The head distance trajectory during the fall of an FN sample of fall motion detection by SVM classifier. The circle indicates the rebound behaviour. Fig. 4. There is an air mattress on the floor to protect the subject. available in the website. D. Comparison To further evaluate the proposed algorithm, we compared it with two state-of-the-art approaches based on depth camera.
8 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS 8 TABLE IV COMPARISON OF DIFFERENT APPROACHES FOR FALL DETECTION. TP TN FP FN Se(%) Sp(%) Ac(%) Er(%) [7] [8] Proposed The approach of [7] is based on human silhouette centre height relative to the floor. The human silhouette centre is obtained by the whole foreground (with morphological filtering). When the silhouette centre is lower than a threshold, a fall is detected. The experimental results are shown in row [7] of Table IV. This algorithm can detect the most of fall events, but with a lot of false positives (FP). It cannot distinguish the fall accident and the non-impact initiative activities well since it does not consider the motion together. When most part of the foreground object is near to the floor, including slowly lying down or sitting down on the floor or bad segmentation, it is detected as fall. The centre location is easily distorted by moving object and bad segmentation. To these events, the two key joints, head and hip, are still tracked well by the proposed method in our experiment. The proposed joint tracking method has a better robustness. The approach of [8] makes use of the orientation of the body, which is based on the joints extracted from Kinect SDK, and the height information of the spine. If the orientation of the body is parallel to the floor and the spine distance to the floor is smaller than a threshold, a fall is detected. The experimental results are shown in row [8] of Table IV. The main disadvantage of this approach is the unreliable joints extraction. When the subject falls and is lying on the floor, the joint extraction is inaccurate and the orientation of the body obtained from inaccurate joints provides false information. Figure 5 shows two examples of the angle between the orientation extracted from Kinect SDK and the floor when the subject walked and fell down ending lying. From Figure 5, it can be noted that the orientations extracted from Kinect SDK were wrong when the subject fell and was lying on the floor. Based on un-robust joint extraction and predefined empirical thresholds, this method s capacity of prediction is limited. Combined with the fall confirmation, the error rate of the proposed method is.4%. As shown in Table IV, the proposed approach outperforms the existing state-of-the-art approaches. Furthermore, compared with [8], which is required to extract at least six joints, only two joints are required in the proposed algorithm. Combined with the higher efficiency feature in Equation, the proposed algorithm has a lower computational complexity. V. CONCLUSION The proposed fall detection approach uses the infra-red based depth camera, so the approach can operate even in the dark condition. The depth camera can measure the human body motion and the relationship between the body and the environment. The floor plane can be extracted from the depth images. To capture the human motion, an enhanced RDT Angle / degree Angle / degree Fig. 5. Two examples of the angle between the orientation extracted from Kinect SDK and the floor when the subject walked and fell down ending lying. Two depth image samples. Angle waveforms corresponded to the two sequences of. algorithm, which reduces the computational complexity based on the one offset feature, is employed to extract the human joints. Existing fall detection method based on joint extraction cannot extract human joints correctly when the subject lies down. The proposed rotation of the person torso orientation increases the accuracy of the joint extraction for fall detection. After extracting the joints, an SVM classifier is proposed to detect the fall based on the joint trajectory. The proposed approach is based on a single depth camera. Because the proposed motion analysis is based on the head tracking and the depth camera can be mounted close to the ceiling, it can avoid most occlusion situations. In the case of occlusion problem, multiple depth cameras based on the proposed approach can solve the problem. However, the proposed approach cannot detect the fall ending lying on furniture, for example, a wooden sofa, since the distance between the body and the floor is too high. The proposed RDT training algorithm based on the one offset feature reduces the number of candidate offsets significantly from to 4. Therefore, the computational complexity of building RDT for fall detection can be reduced by 83 times under the same training condition. Based on depth image sequence, by extracting and tracking the joints of the human body as well as investigating the joints behaviour after fall, the proposed approach can detect and confirm the human fall accurately. Experimental results show that the accuracy of the proposed algorithm is improved by.8% compared with the most recent state-of-the-art fall detection algorithm. ACKNOWLEDGMENT The authors would like to acknowledge the Ph.D. grant from the Institute for Media Innovation, Nanyang Technological
9 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS 9 University, Singapore. REFERENCES [] United-Nations, World population prospects: The 8 revision, 8. [] N. Noury, A. Fleury, P. Rumeau, A. K. Bourke, G. O. Laighin, V. Rialle, and J. E. Lundy, Fall detection-principles and methods, 9th IEEE International Conferenceon Engineering in Medicineand Biology Society (EMBS), pp , 7. [3] W. H. O. (WHO), Good health adds life to years, Global brief for World Health Day. [4] X. Yu, Approaches and principles of fall detectionfor elderly and patient, th IEEE International Conferenceon e-health Networking,Applications and Services (HealthCom), pp. 4 47, 8. [5] M. Mubashir, L. Shao, and L. Seed, A survey on fall detection: Principles and approaches, Neurocomputing, pp. 44 5, 3. [6] N. Thome, S. Miguet, and S. Ambellouis, A real-time, multiview fall detection system: A LHMM-based approach, IEEE Trans. Circuits Syst. Video Technol.(TCSVT), vol. 8, no., pp. 5 53, 8. [7] M. Yu, A. Rhuma, S. Naqvi, L. Wang, and J. Chambers, A posture recognition-based fall detection system for monitoring an elderly person in a smart home environment, IEEE Transactions on Information Technology in Biomedicine, vol. 6, no. 6, pp ,. [8] D. Brulin, Y. Benezeth, and E. Courtial, Posture recognition based on fuzzy logic for home monitoring of the elderly, IEEE Transactions on Information Technology in Biomedicine, vol. 6, no. 5, pp ,. [9] M. Yu, Y. Yu, A. Rhuma, S. Naqvi, L. Wang, and J. Chambers, An online one class support vector machine based person-specific fall detection system for monitoring an elderly individual in a room environment, IEEE Journal of Biomedical and Health Informatics, 3. [] H. Foroughi, B. S. Aski, and H. Pourreza, Intelligent video surveillance for monitoring fall detection of elderly in home environments, in th IEEE International Conference on Computer and Information Technology (ICCIT), 8, pp [] C. Rougier, J. Meunier, A. St-Arnaud, and J. Rousseau, Robust video surveillance for fall detection based on human shape deformation, IEEE Trans. Circuits Syst. Video Technol.(TCSVT), vol., pp. 6 6,. [] C. Rougier and J. Meunie, Fall detection using 3D head trajectory extracted from a single camera video sequence, in First International Work-shop on Video Processing for Security(VP4S), 6. [3] C. Rougier, J. Meunier, A. St-Arnaud, and J. Rousseau, 3D head tracking for fall detection using a single calibrated camera, Image Vision Comput., vol. 3, no. 3, pp , Mar. 3. [4] E. Auvinet, F. Multon, A. Saint-Arnaud, J. Rousseau, and J. Meunier, Fall detection with multiple cameras: An occlusion-resistant method based on 3-D silhouette vertical distribution, IEEE Transactions on Information Technology in Biomedicine, vol. 5, no., pp. 9 3,. [5] C. Doukas and I. Maglogiannis, Emergency fall incidents detection in assisted living environments utilizing motion, sound, and visual perceptual components, IEEE Transactions on Information Technology in Biomedicine, vol. 5, no., pp ,. [6] Microsoft, [7] C. Rougier, E. Auvinet, J. Rousseau, M. Mignotte, and J. Meunier, Fall detection from depth map video sequences, in Proceedings of the 9th international conference on Toward useful services for elderly and people with disabilities: smart homes and health telematics. Berlin, Heidelberg: Springer-Verlag,, pp. 8. [8] R. Planinc and M. Kampel, Introducing the use of depth data for fall detection, Personal Ubiquitous Computing,. [9] Z. P. Bian, L. P. Chau, and N. Magnenat-Thalmann, A depth video approach for fall detection based on human joints height and falling velocity, in International Conference on Computer Animation and Social Agents, May. [] Z.-P. Bian, L.-P. Chau, and N. Magnenat-Thalmann, Fall detection based on skeleton extraction, in Proceedings of the th ACM SIG- GRAPH International Conference on Virtual-Reality Continuum and its Applications in Industry. New York, NY, USA: ACM,, pp [] J. Shotton, A. Fitzgibbon, M. Cook, T. Sharp, M. Finocchio, R. Moore, A. Kipman, and A. Blake, Real-time human pose recognition in parts from single depth images, in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), june, pp [] P. Lagger and P. Fua, Randomized trees for real-time keypoint recognition, in in Proc. CVPR, 5. [3] K. Buys, C. Cagniart, A. Baksheev, T. D. Laet, J. D. Schutter, and C. Pantofaru, An adaptable system for RGB-D based human body detection and pose estimation, accepted for publication in the Journal of Visual Communication and Image Representation. [4] W. H. O., WHO global report on falls prevention in older age, World Health Organization (WHO) Library Cataloguing-in-Publication Data, 7. processing. Zhen-Peng Bian received the B. Eng degree in Microelectronics from South China University of Technology, Guangzhou, China in 7. He is currently pursuing the Ph.D degree from the School of Electrical & Electronic Engineering, Nanyang Technological University, Singapore. His current research interests include fall detection, motion capture, human computer interaction and image processing. Junhui Hou received the B. Eng degree in Information Engineering (Talented Students Program) from South China University of Technology, Guangzhou, China and the M. Eng in Signal and Information Processing from Northwestern Polytechnical University, Xi an, China in 9 and, respectively. He is currently pursuing the Ph.D degree from the School of Electrical & Electronic Engineering, Nanyang Technological University, Singapore. His current research interests include video compression, image processing and computer graphics
10 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS Lap-Pui Chau received the B. Eng degree with first class honours in Electronic Engineering from Oxford Brookes University, England, and the Ph.D. degree in Electronic Engineering from Hong Kong Polytechnic University, Hong Kong, in 99 and 997, respectively. In June 996, he joined Tritech Microelectronics as a senior engineer. Since March 997, he joined Centre for Signal Processing, a national research centre in Nanyang Technological University as a research fellow, subsequently he joined School of Electrical & Electronic Engineering, Nanyang Technological University as an assistant professor and currently, he is an associate professor. His research interests include fast signal processing algorithms, scalable video and video transcoding, robust video transmission, image representation for 3D content delivery, and image based human skeleton extraction. He involved in organization committee of international conferences including the IEEE International Conference on Image Processing (ICIP, ICIP 4), and IEEE International Conference on Multimedia & Expo (ICME ). He is a Technical Program Co-Chairs for Visual Communications and Image Processing (VCIP 3) and International Symposium on Intelligent Signal Processing and Communications Systems (ISPACS ). He was the chair of Technical Committee on Circuits & Systems for Communications (TC-CASC) of IEEE Circuits and Systems Society from to, and the chairman of IEEE Singapore Circuits and Systems Chapter from 9 to. He served as an associate editor for IEEE Transactions on Multimedia, IEEE Signal Processing Letters, and is currently serving as an associate editor for IEEE Transactions on Circuits and Systems for Video Technology, IEEE Transactions on Broadcasting and IEEE Circuits and Systems Society Newsletter. Besides, he is IEEE Distinguished Lecturer for 9-3, and a steering committee member of IEEE Transactions for Mobile Computing from -3. Nadia Magnenat-Thalmann has pioneered various aspects of research of virtual humans over the last 3 years. She obtained several Bachelor s and Master s degrees in various disciplines (Psychology, Biology and Biochemistry) and a PhD in Quantum Physics from the University of Geneva in 977. From 977 to 989, she was a Professor at the University of Montreal in Canada. In 989, she moved to the University of Geneva where she founded the interdisciplinary research group MIRALab. She is Editor-in-Chief of The Visual Computer Journal published by Springer Verlag, and editors of several other journals. During her Career, she has received more than 3 Awards. Among the recent ones, two Doctor Honoris Causa (Leibniz University of Hanover in Germany and University of Ottawa in Canada), the Distinguished Career Award from the Eurographics in Norrkoping, Sweden, and a Career Achievement Award from the Canadian Human Computer Communications Society in Toronto. Very recently, she received the prestigious Humboldt Research Award in Germany. Besides directing her research group MIRALab in Switzerland, she is presently visiting Professor and Director of the Institute for Media Innovation (IMI) at Nanyang Technological University, Singapore. For more information, please visit CurriculumVitae.aspx
Vision based approach to human fall detection
Vision based approach to human fall detection Pooja Shukla, Arti Tiwari CSVTU University Chhattisgarh, [email protected] 9754102116 Abstract Day by the count of elderly people living alone at home
Fall Detection System based on Kinect Sensor using Novel Detection and Posture Recognition Algorithm
Fall Detection System based on Kinect Sensor using Novel Detection and Posture Recognition Algorithm Choon Kiat Lee 1, Vwen Yen Lee 2 1 Hwa Chong Institution, Singapore [email protected] 2 Institute
Journal of Industrial Engineering Research. Adaptive sequence of Key Pose Detection for Human Action Recognition
IWNEST PUBLISHER Journal of Industrial Engineering Research (ISSN: 2077-4559) Journal home page: http://www.iwnest.com/aace/ Adaptive sequence of Key Pose Detection for Human Action Recognition 1 T. Sindhu
The Visual Internet of Things System Based on Depth Camera
The Visual Internet of Things System Based on Depth Camera Xucong Zhang 1, Xiaoyun Wang and Yingmin Jia Abstract The Visual Internet of Things is an important part of information technology. It is proposed
Privacy Preserving Automatic Fall Detection for Elderly Using RGBD Cameras
Privacy Preserving Automatic Fall Detection for Elderly Using RGBD Cameras Chenyang Zhang 1, Yingli Tian 1, and Elizabeth Capezuti 2 1 Media Lab, The City University of New York (CUNY), City College New
Building an Advanced Invariant Real-Time Human Tracking System
UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian
Tracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey [email protected] b Department of Computer
How does the Kinect work? John MacCormick
How does the Kinect work? John MacCormick Xbox demo Laptop demo The Kinect uses structured light and machine learning Inferring body position is a two-stage process: first compute a depth map (using structured
Machine Learning for Medical Image Analysis. A. Criminisi & the InnerEye team @ MSRC
Machine Learning for Medical Image Analysis A. Criminisi & the InnerEye team @ MSRC Medical image analysis the goal Automatic, semantic analysis and quantification of what observed in medical scans Brain
Spatio-Temporally Coherent 3D Animation Reconstruction from Multi-view RGB-D Images using Landmark Sampling
, March 13-15, 2013, Hong Kong Spatio-Temporally Coherent 3D Animation Reconstruction from Multi-view RGB-D Images using Landmark Sampling Naveed Ahmed Abstract We present a system for spatio-temporally
Automated Monitoring System for Fall Detection in the Elderly
Automated Monitoring System for Fall Detection in the Elderly Shadi Khawandi University of Angers Angers, 49000, France [email protected] Bassam Daya Lebanese University Saida, 813, Lebanon Pierre
Speed Performance Improvement of Vehicle Blob Tracking System
Speed Performance Improvement of Vehicle Blob Tracking System Sung Chun Lee and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA [email protected], [email protected] Abstract. A speed
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering
Human Pose Estimation from RGB Input Using Synthetic Training Data
Human Pose Estimation from RGB Input Using Synthetic Training Data Oscar Danielsson and Omid Aghazadeh School of Computer Science and Communication KTH, Stockholm, Sweden {osda02, omida}@kth.se arxiv:1405.1213v2
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA N. Zarrinpanjeh a, F. Dadrassjavan b, H. Fattahi c * a Islamic Azad University of Qazvin - [email protected]
A Real-Time Fall Detection System in Elderly Care Using Mobile Robot and Kinect Sensor
International Journal of Materials, Mechanics and Manufacturing, Vol. 2, No. 2, May 201 A Real-Time Fall Detection System in Elderly Care Using Mobile Robot and Kinect Sensor Zaid A. Mundher and Jiaofei
A DECISION TREE BASED PEDOMETER AND ITS IMPLEMENTATION ON THE ANDROID PLATFORM
A DECISION TREE BASED PEDOMETER AND ITS IMPLEMENTATION ON THE ANDROID PLATFORM ABSTRACT Juanying Lin, Leanne Chan and Hong Yan Department of Electronic Engineering, City University of Hong Kong, Hong Kong,
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
DINAMIC AND STATIC CENTRE OF PRESSURE MEASUREMENT ON THE FORCEPLATE. F. R. Soha, I. A. Szabó, M. Budai. Abstract
ACTA PHYSICA DEBRECINA XLVI, 143 (2012) DINAMIC AND STATIC CENTRE OF PRESSURE MEASUREMENT ON THE FORCEPLATE F. R. Soha, I. A. Szabó, M. Budai University of Debrecen, Department of Solid State Physics Abstract
Template-based Eye and Mouth Detection for 3D Video Conferencing
Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
Automatic Fall Detector based on Sliding Window Principle
Automatic Fall Detector based on Sliding Window Principle J. Rodriguez 1, M. Mercuri 2, P. Karsmakers 3,4, P.J. Soh 2, P. Leroux 3,5, and D. Schreurs 2 1 UPC, Div. EETAC-TSC, Castelldefels, Spain 2 KU
Fall detection in the elderly by head tracking
Loughborough University Institutional Repository Fall detection in the elderly by head tracking This item was submitted to Loughborough University's Institutional Repository by the/an author. Citation:
Vision-Based Blind Spot Detection Using Optical Flow
Vision-Based Blind Spot Detection Using Optical Flow M.A. Sotelo 1, J. Barriga 1, D. Fernández 1, I. Parra 1, J.E. Naranjo 2, M. Marrón 1, S. Alvarez 1, and M. Gavilán 1 1 Department of Electronics, University
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Automatic Traffic Estimation Using Image Processing
Automatic Traffic Estimation Using Image Processing Pejman Niksaz Science &Research Branch, Azad University of Yazd, Iran [email protected] Abstract As we know the population of city and number of
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
Recognition of Day Night Activity Using Accelerometer Principals with Aurdino Development Board
Recognition of Day Night Activity Using Accelerometer Principals with Aurdino Development Board Swapna Shivaji Arote R.S. Bhosale S.E. Pawar Dept. of Information Technology, Dept. of Information Technology,
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Circle Object Recognition Based on Monocular Vision for Home Security Robot
Journal of Applied Science and Engineering, Vol. 16, No. 3, pp. 261 268 (2013) DOI: 10.6180/jase.2013.16.3.05 Circle Object Recognition Based on Monocular Vision for Home Security Robot Shih-An Li, Ching-Chang
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION Saurabh Asija 1, Rakesh Singh 2 1 Research Scholar (Computer Engineering Department), Punjabi University, Patiala. 2 Asst.
A Reliability Point and Kalman Filter-based Vehicle Tracking Technique
A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video
Intrusion Detection via Machine Learning for SCADA System Protection
Intrusion Detection via Machine Learning for SCADA System Protection S.L.P. Yasakethu Department of Computing, University of Surrey, Guildford, GU2 7XH, UK. [email protected] J. Jiang Department
International Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014
Efficient Attendance Management System Using Face Detection and Recognition Arun.A.V, Bhatath.S, Chethan.N, Manmohan.C.M, Hamsaveni M Department of Computer Science and Engineering, Vidya Vardhaka College
Neural Network based Vehicle Classification for Intelligent Traffic Control
Neural Network based Vehicle Classification for Intelligent Traffic Control Saeid Fazli 1, Shahram Mohammadi 2, Morteza Rahmani 3 1,2,3 Electrical Engineering Department, Zanjan University, Zanjan, IRAN
CS231M Project Report - Automated Real-Time Face Tracking and Blending
CS231M Project Report - Automated Real-Time Face Tracking and Blending Steven Lee, [email protected] June 6, 2015 1 Introduction Summary statement: The goal of this project is to create an Android
A Survey on Vision-based Fall Detection
A Survey on Vision-based Fall Detection Zhong Zhang, Christopher Conly, and Vassilis Athitsos Department of Computer Science and Engineering University of Texas at Arlington Arlington, Texas, USA [email protected],
HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT
International Journal of Scientific and Research Publications, Volume 2, Issue 4, April 2012 1 HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT Akhil Gupta, Akash Rathi, Dr. Y. Radhika
Virtual Fitting by Single-shot Body Shape Estimation
Virtual Fitting by Single-shot Body Shape Estimation Masahiro Sekine* 1 Kaoru Sugita 1 Frank Perbet 2 Björn Stenger 2 Masashi Nishiyama 1 1 Corporate Research & Development Center, Toshiba Corporation,
Vision-based Walking Parameter Estimation for Biped Locomotion Imitation
Vision-based Walking Parameter Estimation for Biped Locomotion Imitation Juan Pedro Bandera Rubio 1, Changjiu Zhou 2 and Francisco Sandoval Hernández 1 1 Dpto. Tecnología Electrónica, E.T.S.I. Telecomunicación
3D Scanner using Line Laser. 1. Introduction. 2. Theory
. Introduction 3D Scanner using Line Laser Di Lu Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute The goal of 3D reconstruction is to recover the 3D properties of a geometric
CS 534: Computer Vision 3D Model-based recognition
CS 534: Computer Vision 3D Model-based recognition Ahmed Elgammal Dept of Computer Science CS 534 3D Model-based Vision - 1 High Level Vision Object Recognition: What it means? Two main recognition tasks:!
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals Chun Lei He, Ping Zhang, Jianxiong Dong, Ching Y. Suen, Tien D. Bui Centre for Pattern Recognition and Machine Intelligence,
Limits and Possibilities of Markerless Human Motion Estimation
Limits and Possibilities of Markerless Human Motion Estimation Bodo Rosenhahn Universität Hannover Motion Capture Wikipedia (MoCap): Approaches for recording and analyzing Human motions Markerless Motion
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006
Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY V. Knyaz a, *, Yu. Visilter, S. Zheltov a State Research Institute for Aviation System (GosNIIAS), 7, Victorenko str., Moscow, Russia
Removing Moving Objects from Point Cloud Scenes
1 Removing Moving Objects from Point Cloud Scenes Krystof Litomisky [email protected] Abstract. Three-dimensional simultaneous localization and mapping is a topic of significant interest in the research
A Study on M2M-based AR Multiple Objects Loading Technology using PPHT
A Study on M2M-based AR Multiple Objects Loading Technology using PPHT Sungmo Jung, Seoksoo Kim * Department of Multimedia Hannam University 133, Ojeong-dong, Daedeok-gu, Daejeon-city Korea [email protected],
Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior
Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior Kenji Yamashiro, Daisuke Deguchi, Tomokazu Takahashi,2, Ichiro Ide, Hiroshi Murase, Kazunori Higuchi 3,
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS Norbert Buch 1, Mark Cracknell 2, James Orwell 1 and Sergio A. Velastin 1 1. Kingston University, Penrhyn Road, Kingston upon Thames, KT1 2EE,
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS B.K. Mohan and S. N. Ladha Centre for Studies in Resources Engineering IIT
Local features and matching. Image classification & object localization
Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to
Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object
Professor, D.Sc. (Tech.) Eugene Kovshov MSTU «STANKIN», Moscow, Russia
Professor, D.Sc. (Tech.) Eugene Kovshov MSTU «STANKIN», Moscow, Russia As of today, the issue of Big Data processing is still of high importance. Data flow is increasingly growing. Processing methods
How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm
IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor Dominik Wȩgrzyn and Luís A. Alexandre IT - Instituto de Telecomunicações Dept. of Computer Science, Univ. Beira Interior, 6200-001 Covilhã, Portugal
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
Efficient Background Subtraction and Shadow Removal Technique for Multiple Human object Tracking
ISSN: 2321-7782 (Online) Volume 1, Issue 7, December 2013 International Journal of Advance Research in Computer Science and Management Studies Research Paper Available online at: www.ijarcsms.com Efficient
Tracking Groups of Pedestrians in Video Sequences
Tracking Groups of Pedestrians in Video Sequences Jorge S. Marques Pedro M. Jorge Arnaldo J. Abrantes J. M. Lemos IST / ISR ISEL / IST ISEL INESC-ID / IST Lisbon, Portugal Lisbon, Portugal Lisbon, Portugal
Low-resolution Character Recognition by Video-based Super-resolution
2009 10th International Conference on Document Analysis and Recognition Low-resolution Character Recognition by Video-based Super-resolution Ataru Ohkura 1, Daisuke Deguchi 1, Tomokazu Takahashi 2, Ichiro
Real time vehicle detection and tracking on multiple lanes
Real time vehicle detection and tracking on multiple lanes Kristian Kovačić Edouard Ivanjko Hrvoje Gold Department of Intelligent Transportation Systems Faculty of Transport and Traffic Sciences University
VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS
VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS Aswin C Sankaranayanan, Qinfen Zheng, Rama Chellappa University of Maryland College Park, MD - 277 {aswch, qinfen, rama}@cfar.umd.edu Volkan Cevher, James
Simultaneous Gamma Correction and Registration in the Frequency Domain
Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong [email protected] William Bishop [email protected] Department of Electrical and Computer Engineering University
ECE 533 Project Report Ashish Dhawan Aditi R. Ganesan
Handwritten Signature Verification ECE 533 Project Report by Ashish Dhawan Aditi R. Ganesan Contents 1. Abstract 3. 2. Introduction 4. 3. Approach 6. 4. Pre-processing 8. 5. Feature Extraction 9. 6. Verification
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
Face Recognition in Low-resolution Images by Using Local Zernike Moments
Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August14-15, 014 Paper No. 15 Face Recognition in Low-resolution Images by Using Local Zernie
Video Surveillance System for Security Applications
Video Surveillance System for Security Applications Vidya A.S. Department of CSE National Institute of Technology Calicut, Kerala, India V. K. Govindan Department of CSE National Institute of Technology
Introduction. www.imagesystems.se
Product information Image Systems AB Main office: Ågatan 40, SE-582 22 Linköping Phone +46 13 200 100, fax +46 13 200 150 [email protected], Introduction Motion is the world leading software for advanced
Modelling 3D Avatar for Virtual Try on
Modelling 3D Avatar for Virtual Try on NADIA MAGNENAT THALMANN DIRECTOR MIRALAB UNIVERSITY OF GENEVA DIRECTOR INSTITUTE FOR MEDIA INNOVATION, NTU, SINGAPORE WWW.MIRALAB.CH/ Creating Digital Humans Vertex
Automatic Labeling of Lane Markings for Autonomous Vehicles
Automatic Labeling of Lane Markings for Autonomous Vehicles Jeffrey Kiske Stanford University 450 Serra Mall, Stanford, CA 94305 [email protected] 1. Introduction As autonomous vehicles become more popular,
ExmoR A Testing Tool for Control Algorithms on Mobile Robots
ExmoR A Testing Tool for Control Algorithms on Mobile Robots F. Lehmann, M. Ritzschke and B. Meffert Institute of Informatics, Humboldt University, Unter den Linden 6, 10099 Berlin, Germany E-mail: [email protected],
Simple and efficient online algorithms for real world applications
Simple and efficient online algorithms for real world applications Università degli Studi di Milano Milano, Italy Talk @ Centro de Visión por Computador Something about me PhD in Robotics at LIRA-Lab,
Tracking performance evaluation on PETS 2015 Challenge datasets
Tracking performance evaluation on PETS 2015 Challenge datasets Tahir Nawaz, Jonathan Boyle, Longzhen Li and James Ferryman Computational Vision Group, School of Systems Engineering University of Reading,
Bernice E. Rogowitz and Holly E. Rushmeier IBM TJ Watson Research Center, P.O. Box 704, Yorktown Heights, NY USA
Are Image Quality Metrics Adequate to Evaluate the Quality of Geometric Objects? Bernice E. Rogowitz and Holly E. Rushmeier IBM TJ Watson Research Center, P.O. Box 704, Yorktown Heights, NY USA ABSTRACT
Support Vector Machine-Based Human Behavior Classification in Crowd through Projection and Star Skeletonization
Journal of Computer Science 6 (9): 1008-1013, 2010 ISSN 1549-3636 2010 Science Publications Support Vector Machine-Based Human Behavior Classification in Crowd through Projection and Star Skeletonization
Blog Post Extraction Using Title Finding
Blog Post Extraction Using Title Finding Linhai Song 1, 2, Xueqi Cheng 1, Yan Guo 1, Bo Wu 1, 2, Yu Wang 1, 2 1 Institute of Computing Technology, Chinese Academy of Sciences, Beijing 2 Graduate School
Robotics. Lecture 3: Sensors. See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information.
Robotics Lecture 3: Sensors See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review: Locomotion Practical
Open Access A Facial Expression Recognition Algorithm Based on Local Binary Pattern and Empirical Mode Decomposition
Send Orders for Reprints to [email protected] The Open Electrical & Electronic Engineering Journal, 2014, 8, 599-604 599 Open Access A Facial Expression Recognition Algorithm Based on Local Binary
The Implementation of Face Security for Authentication Implemented on Mobile Phone
The Implementation of Face Security for Authentication Implemented on Mobile Phone Emir Kremić *, Abdulhamit Subaşi * * Faculty of Engineering and Information Technology, International Burch University,
An Active Head Tracking System for Distance Education and Videoconferencing Applications
An Active Head Tracking System for Distance Education and Videoconferencing Applications Sami Huttunen and Janne Heikkilä Machine Vision Group Infotech Oulu and Department of Electrical and Information
How To Filter Spam Image From A Picture By Color Or Color
Image Content-Based Email Spam Image Filtering Jianyi Wang and Kazuki Katagishi Abstract With the population of Internet around the world, email has become one of the main methods of communication among
PHYSIOLOGICALLY-BASED DETECTION OF COMPUTER GENERATED FACES IN VIDEO
PHYSIOLOGICALLY-BASED DETECTION OF COMPUTER GENERATED FACES IN VIDEO V. Conotter, E. Bodnari, G. Boato H. Farid Department of Information Engineering and Computer Science University of Trento, Trento (ITALY)
Real-Time Tracking of Pedestrians and Vehicles
Real-Time Tracking of Pedestrians and Vehicles N.T. Siebel and S.J. Maybank. Computational Vision Group Department of Computer Science The University of Reading Reading RG6 6AY, England Abstract We present
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network
Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal
THE MS KINECT USE FOR 3D MODELLING AND GAIT ANALYSIS IN THE MATLAB ENVIRONMENT
THE MS KINECT USE FOR 3D MODELLING AND GAIT ANALYSIS IN THE MATLAB ENVIRONMENT A. Procházka 1,O.Vyšata 1,2,M.Vališ 1,2, M. Yadollahi 1 1 Institute of Chemical Technology, Department of Computing and Control
Free Fall: Observing and Analyzing the Free Fall Motion of a Bouncing Ping-Pong Ball and Calculating the Free Fall Acceleration (Teacher s Guide)
Free Fall: Observing and Analyzing the Free Fall Motion of a Bouncing Ping-Pong Ball and Calculating the Free Fall Acceleration (Teacher s Guide) 2012 WARD S Science v.11/12 OVERVIEW Students will measure
Fall Detection using Kinect Sensor and Fall Energy Image
Fall Detection using Kinect Sensor and Fall Energy Image Bogdan Kwolek 1 and Michal Kepski 2 1 AGH University of Science and Technology, 30 Mickiewicza Av., 30-059 Krakow, Poland [email protected] 2 University
Multivariate data visualization using shadow
Proceedings of the IIEEJ Ima and Visual Computing Wor Kuching, Malaysia, Novembe Multivariate data visualization using shadow Zhongxiang ZHENG Suguru SAITO Tokyo Institute of Technology ABSTRACT When visualizing
A New Approach to Cutting Tetrahedral Meshes
A New Approach to Cutting Tetrahedral Meshes Menion Croll August 9, 2007 1 Introduction Volumetric models provide a realistic representation of three dimensional objects above and beyond what traditional
Computer Animation and Visualisation. Lecture 1. Introduction
Computer Animation and Visualisation Lecture 1 Introduction 1 Today s topics Overview of the lecture Introduction to Computer Animation Introduction to Visualisation 2 Introduction (PhD in Tokyo, 2000,
Interactive person re-identification in TV series
Interactive person re-identification in TV series Mika Fischer Hazım Kemal Ekenel Rainer Stiefelhagen CV:HCI lab, Karlsruhe Institute of Technology Adenauerring 2, 76131 Karlsruhe, Germany E-mail: {mika.fischer,ekenel,rainer.stiefelhagen}@kit.edu
CCTV - Video Analytics for Traffic Management
CCTV - Video Analytics for Traffic Management Index Purpose Description Relevance for Large Scale Events Technologies Impacts Integration potential Implementation Best Cases and Examples 1 of 12 Purpose
