Fall Detection using Kinect Sensor and Fall Energy Image
|
|
|
- Kory Blair
- 10 years ago
- Views:
Transcription
1 Fall Detection using Kinect Sensor and Fall Energy Image Bogdan Kwolek 1 and Michal Kepski 2 1 AGH University of Science and Technology, 30 Mickiewicza Av., Krakow, Poland [email protected] 2 University of Rzeszow, 16c Rejtana Av., Rzeszów, Poland [email protected] Abstract. One of the main reasons for low acceptance by seniors the available technology for automatic fall detection is that the existing devices generate too much false alarms. Additionally, the camera-based devices do not preserve the privacy adequately. In our approach an accelerometer is utilized to indicate a potential fall. A fall hypothesis is then verified in the second stage in which we employ a depth image, which was shot at the moment of the potential fall. A detector that was trained in advance on features extracted both from depth images and points cloud is responsible for verification whether a person is lying on the floor. After all, to reliably distinguish the fall from fall-like activities we perform final verification, in which we employ the proposed fall energy image. The fall energy image expresses the distribution of the person s motion in the set of images preceding the fall. Keywords: Depth image and point cloud processing; fall detection. 1 Introduction Falls are a major health risk and a significant obstacle to independent living of the seniors [13]. In response to the demand for fall detection technology, plenty of research has been done in the recent years to develop unobtrusive fall detection systems for enhancing the functional ability of the elderly and patients [16]. However, despite many efforts made to obtain reliable and unobtrusive person fall detection, current technology does not meet the seniors needs. One of the main reasons for non-acceptance of the currently available technology by elderly is that the existing devices generate too much false alarms, which in turn lead to considerable frustration of the seniors. Additionally, the existing devices do not preserve the privacy and unobtrusiveness adequately. In recent years, a lot of research has been done on detecting falls using a wide range of sensor types. Mubashir et al. [16] done a survey of methods used in the existing systems. Single CCD camera [18], multiple cameras [6], specialized omnidirectional ones [15] and stereo-pair cameras [9] are widely used in the vision systems for fall detection. Most of the currently available techniques for fall
2 detection are based on wearable sensors. Accelerometers or both accelerometers and gyroscopes are the most frequently used sensors in devices responsible for fall monitoring [17]. However, on the basis of inertial sensors it is not easy to separate real falls from fall-like activities [2] and in consequence the devices that are built on only such sensors typically trigger significant amount of false alarms. The reason is that the characteristic motion patterns of fall also exist in many actions. For instance, the crouch also demonstrates a rapid downward motion. Recently, Kinect sensor was used in prototype systems for fall detection [10, 11, 14]. It is the world s first low-cost device that combines an RGB camera and a depth sensor. Unlike 2D cameras, it allows 3D tracking of the body movements. Thus, if only depth images are used it preserves the person s privacy. Because depth images are extracted with the support of an active light source, they are largely independent of external light conditions. Thanks to the use of the infrared light the Kinect is capable of extracting the depth images in dark rooms. In this work we demonstrate an approach to reduce the number of false positives alarms in fall detection through the use of an accelerometer and the depth images. The accelerometer is utilized to indicate a potential fall. A fall hypothesis is then verified in the second stage in which we employ a depth image, which was shot at the time of the potential fall of the person. A detector that was trained in advance on features extracted both from depth images and points cloud is responsible for verification whether a person is lying on the floor. After all, to reliably distinguish the fall from fall-like activities we perform final verification, in which we employ the proposed fall energy image. The fall energy image expresses the distribution of the person s motion in a collection of the images, acquired in a certain period of time before the potential fall alert. The contribution of this work is twofold: firstly we propose fall energy images (FEI) as an effective spatiotemporal representation of the human fall. Secondly, we show how to extract such energy fall images on the basis of the depth images and then how to utilize them to achieve reliable fall detection. Shape modeling using spatiotemporal features provides crucial information about human activities. In [7], a method for fall detection that is based on a combination of the eigenspace and integrated time motion images (ITMI) was developed. ITMI contain motion information and time stamps of motion occurrence. Multilayer perceptron neural network was utilized for classification of motions and detection of the fall event. In [19], a mobile human airbag system was designed for fall protection for the elderly. A Micro Inertial Measurement Unit consisting of three dimensional accelerometers, gyroscopes, a Bluetooth module and a Micro Controller Unit (MCU) is utilized to record human motion information. Through analysis of images acquired by a high-speed camera, a lateral fall can be determined on the basis of a gyro threshold. The classification of falls is performed by a support vector machine (SVM) classifier. The majority of vision based systems for fall detection do not take into account the motion information. In this work we demonstrate how to extract fall energy images using accelerometer and depth images as well as how to process them. The accelerometer helps us to extract the representative segment of the images as a representation of the fall event.
3 2 Person Detection in Depth Images Depth is very useful cue to achieve reliable person detection because humans may not have consistent color and texture but have to occupy an integrated region in space. The Kinect combines structured light with two classic computer vision techniques, namely depth from focus and depth from stereo. It is equipped with infrared laser-based IR emitter, an infrared camera and a RGB camera. The IR camera and the IR projector compose a stereo pair with a baseline of approximately 75 mm. A known pattern of dots is projected from the IR laser emitter. These specs are captured by the IR camera and then compared to the known pattern. Since there is the distance between laser and sensor, the images correspond to different camera positions, and that in turn allows to use stereo triangulation to calculate each spec depth. The field of view of the system is 57 horizontally and 43 vertically, the minimum measurement range is about 0.6 m, whereas the maximum range is somewhere between 4-5 m. It captures the depth and color images simultaneously at a frame rate of about 30 fps. The default RGB video stream has size and 8-bit for each channel, whereas the depth stream is resolution and with 11-bit depth. The software called NITE from PrimeSense offers skeleton tracking on the basis of depth images. However, this software is targeted for supporting the human-computer interaction, and not for detecting the person fall. Thus, in many circumstances it can have difficulties in extracting and tracking the person s skeleton. Therefore, we employ a person detection method [11], which reliably extracts the subject including situations when he/she is lying on the floor. The person was delineated on the basis of a scene reference image, which was extracted in advance and then updated on-line. In the depth reference image each pixel assumes the median value of several pixels values from the past images. In the setup phase we collect a number of the depth images, and for each pixel we assemble a list of the pixel values from the former images, which is then sorted in order to extract the median. Given the sorted lists of pixels the depth reference image can be updated quickly by removing the oldest pixels and updating the sorted lists with the pixels from the current depth image and then extracting the median value. We found that for typical human motions, good results can be obtained using 13 depth images [11]. For Kinect acquiring the images at 25 Hz we take every fifteenth image. In the detection mode the foreground objects are extracted through differencing the current depth image from such a depth reference map. Afterwards, the person is delineated through extracting the largest connected component in the thresholded difference between the current map and the reference map. 3 V-disparity Based Ground Plane Extraction Given a depth map provided by the Kinect sensor, the disparity d can be determined in the following manner: d = b f z (1)
4 where z is the depth (in meters), b is the horizontal baseline between the cameras (in meters), f is the (common) focal length of the cameras (in pixels). The IR camera and the IR projector form a stereo pair with a baseline of approximately b = 7.5 cm, whereas the focal length f is equal to 580 pixels. Let H be a function of the disparities d such that H(d) = I d. The I d is the v-disparity image and H accumulates the pixels with the same disparity from a given line of the disparity image. Thus, in the v-disparity image each point in the line i represents the number of points with the same disparity occurring in the i-th line of the disparity image. In [12] the v-disparity maps between two stereo images were used to achieve reliable obstacle detection. In our work the v-disparity maps are extracted using depth images determined by Kinect. The line corresponding to the floor pixels in the v-disparity map was extracted using the Hough transform operating on v-disparity values and a predefined range of parameters. The accumulator was incremented by v-disparity values [11]. Assuming that the Kinect is placed at height about 1 m from the floor, the line representing the floor should begin in the disparities ranging from 21 to 25 depending on the tilt angle of the sensor. Given the extracted line in such a way, the pixels belonging to the floor areas were determined [11]. Due to the measurement inaccuracies, pixels falling into some disparity extent d t were also considered as belonging to the ground. Assuming that d y is a disparity in the line y, which represents the pixels belonging to the ground plane, we take into account the disparities from the range d (d y d t, d y + d t ) as a representation of the ground plane. After the transformation of the pixels representing the floor to the 3D points cloud, the plane described by the equation ax+by+cx+d has been recovered [11]. The parameters a, b, c and d were estimated using the RANSAC algorithm. The distance to the ground plane from the 3D centroid of points cloud corresponding to the segmented person was determined on the basis of the following equation: D = ax c + by c + cz c + d a2 + b 2 + c 2 (2) where X c, Y c, Z c stand for the coordinates of the person s centroid. 4 Lying Pose Recognition The recognition of lying pose was achieved using a classier trained on features representing the extracted person both in depth images and in point clouds. A data-set consisting of images with normal activities like walking, sitting down, crouching down and lying has been composed in order to train a classifier responsible for testing whether a person is lying on the floor and to evaluate its performance. Thirty five volunteers with age under 28 years attended in preparation of the data-set. The image sequences were recorded using two Kinect devices. The first Kinect was placed at a height of about one meter to the floor, whereas the second one was placed at a ceiling corner of the room. Figure 1 shows example depth images seen from such two different views.
5 Fig. 1: Person in depth images seen from two different views. In total 312 images representing typical human actions were selected and then utilized to extract the following features: h/w - a ratio of width to height of the person s bounding box, calculated in the points cloud h/h max - a ratio expressing the height of the person s surrounding box in the current frame to the height of the person dist - the distance of the person centroid to the floor, expressed in millimeters max(σ x, σ z ) - standard deviation from the centroid for the abscissa and the depth, respectively. Figure 2 depicts a scatterplot matrix for the employed attributes, in which a collection of scatterplots is organized in a two-dimensional matrix simultaneously to provide correlation information among the attributes. In a single scatterplot two h/w h/h max dist max(σ x,σ z ) h/w h/h max dist max(σ x,σ z ) Fig. 2: Multivariate classification scatter plot for features used in lying pose recognition.
6 attributes are projected along the x-y axes of the Cartesian coordinates. As we can observe, the overlaps in the attribute space are not too significant. We considered also another attributes, for instance, a filling ratio of the rectangles making up the person s bounding box. The worth of the considered features was evaluated on the basis of the information gain [4], which measures the dependence between the feature and the class label. In the assessment of the discrimination power of the considered features and selecting the most discriminative ones we utilized the InfoGainAttributeEval procedure from the Weka [5], which is a collection of machine learning algorithms. 5 Fall Energy Image Several motion features have been proposed until now to represent people activities, such as Motion History Image (MHI) [1]. Usually, the MHI is generated on the basis of binary images, where the person silhouette sequence is condensed into gray scale images as a weighted combination of all motion images. The result of such a motion condensation is a scalar-valued image in which more recently moving pixels are brighter. One of the advantages of the MHI representation is that a range of action images may be encoded in a single motion-shape. Typically, in action recognition phase such a static shape pattern is compared with pre-stored action prototypes. The Fall Energy Image is an average of all silhouette images of a single fall. Such a spatiotemporal energy map spans the time scale of person fall. The energy map is calculated using a number of binary silhouette images before the fall. The images are scaled according to the distance of the person to the camera. We assume that a fall occurs if the signal upper peak value from the accelerometer is greater than 3g. Figure 3 illustrates example fall energy images with the corresponding plots of signal upper peak value (UPV) vs. time. As we can observe, both actions have quite similar characteristics in the acceleration domain, but totally different fall energy maps. The weighted average (moment) of the fall energy expressed by pixel intensities was computed using moments as follows: x y xp (x, y) x c = x y P (x, y) (3) y c = x y x y yp (x, y) P (x, y) where x, y are pixel coordinates. The major length and width (eigenvalues) of the fall energy has been calculated in the following manner [8]: l = (a + c) + b 2 + (a c) 2 w = (a + c) (4) b 2 + (a c) 2
7 Fig. 3: Fall energy images for a forward fall (left) and sitting on a chair (right) with corresponding plots of signal upper peak value vs. time (bottom row). where a = M 20 M 00 x 2 c, b = 2( M 11 M 00 x c y c ), c = M 02 M 00 M 11 = x y xyp (x, y), M 20 = x yc 2, M 00 = x y x2 P (x, y), M 02 = x P (x, y), y y2 P (x, y). We calculated also the average fall energy, i.e. the mean value of non-zero pixel values in the fall energy image P (x, y) as well as the Euclidean distance d E between the weighted average location of the fall energy (y c, x c ) and the geometrical centroid of the thresholded energy map. Figure 4 depicts the scatter plot matrix for such energy features. The features were extracted on the basis of 30 image sequences in which half of them contained person falls. The remaining sequences contained person activities, which were very similar to fall. The activities were performed close to the floor and contained actions consisting in sitting on the floor, laying down on the floor, for instance to raise an object, etc. The features were extracted on the basis of 30 depth images just before the human fall, which in turn was signaled by a procedure processing data from the accelerometer. That means that the FEI image expresses the fall energy in about 1 sec. As we can observe, on the basis of such a set of features the person fall can be distinguished from the non-fall activities. We considered also energy features extracted on the basis of the bank of Log-Gabor filters. Their worth was evaluated on the basis of the information gain and then compared to the discrimination power of the above discussed features. The experimental results showed that their worth is not worse in comparison to Gabor filter based energy features and therefore we decided to use them in the evaluation of the whole system. It is worth to note that they can be extracted in considerably shorter time. y
8 avg(i) l w d E avg(i) w d E l Fig. 4: Multivariate classification scatter plot for features extracted on fall energy images. 6 Experimental Results Thirty five young health volunteers with age under 28 years attended in preparation of the data-sets and in the evaluation of the fall detection accuracy. To show the resistance of the system to the placement of the camera the images were acquired by two Kinect devices. The motion data were acquired by a wearable smart device (Sony PlayStation Move) containing accelerometer and gyroscope sensors. Data from the device were transmitted wirelessly via Bluetooth and received by a laptop computer. In all, 312 images acquired by two Kinect devices were selected and then used to evaluate the k-nn classifier responsible for checking whether the person is lying on the floor. The number of images with a fall was equal to 110. We evaluated also KStar [3], SVM and multilayer perceptron (MLP) classifiers. The KStar and MLP classified all falls correctly, whereas the remaining algorithms incorrectly classified two instances. A k-nn based motion classifier was trained on 30 image sequences of which 15 contained fall events. Its accuracy was evaluated in 10-fold cross-validation and one fall was classified incorrectly. The SVM and KStar classified all falls correctly. The complete system for fall detection was tested with simulated-falls performed by young volunteers under supervised conditions onto crash mats. The accelerometer was worn near the pelvis. Five volunteers attended in the tests and evaluations of our system. Intentional falls were performed in home towards a carpet with thickness of about 2 cm. Each individual performed ADLs like walk-
9 ing, sitting, crouching down, leaning down/picking up objects from the floor, lying on a bed. As expected, using only the accelerometer the number of false alarms was considerable. Experimental results demonstrated that most of them can be ignored owing to the use of our recognition module of the lying pose. This operation is done at low computational cost as the verification of the fall is performed if the module processing the data from the accelerometer triggers the alarm. Moreover, on the basis of the accelerometer based alarm the system obtains information which image should be processed to decide if an event consisting in person lying on the floor takes place. All person activities that have been considered in the previous work [10] were classified correctly. During the evaluation of the system the volunteers found several fall-like actions, which were not considered in the previous work and for which the two-stage algorithm triggered false alarms. The experimental results obtained on the system with three modules, i.e. accelerometer, lying pose recognition and fall energy analysis demonstrated that the fall energy features are very useful in further reduction of the false alarm ratio. A comprehensive evaluation showed that the system has high accuracy of fall detection and very low level of false alarms. It demonstrated that the placement of the cameras does not have an influence on the classification accuracy. The depth images were acquired by the Kinect sensors using OpenNI. The system was implemented in C/C++ and runs at 25 fps on 2.4 GHz I7 notebook. The most computationally demanding operation is extraction of the depth reference image of the scene. For images of size the computation time needed for extraction of the depth reference image is about 9 milliseconds. In order to reduce the computational overload the depth reference images were only updated if on the image acquired in the moment of the fall, two or more blobs had been detected. In practice, we examined the thresholded difference between the current depth map and the reference map in terms of the number of blobs. 7 Conclusions In this work we demonstrated how to achieve reliable fall detection with low false positives number. Given the alarm trigger obtained on the basis of data from wireless accelerometer, the system extracts the person features from the corresponding depth image and point clouds. The system uses them in a k-nn classifier to examine if the person is lying on the floor. In order to further reduce the false alarm ratio the system extracts fall energy images from a sequence of images up to the fall and then employs the energy features in a k-nn classifier. Experimental results demonstrated that this leads to considerable reduction of false alarms and high detection ratio. The system preserves the privacy of the user and works in poor lighting conditions. Acknowledgment This work has been supported by the National Science Centre (NCN) within the project N N
10 References 1. Ahad, M.A.R., Tan, J.K., Kim, H., Ishikawa, S.: Motion history image: its variants and applications. Mach. Vision Appl. 23(2), (Mar 2012) 2. Bourke, A., O Brien, J., Lyons, G.: Evaluation of a threshold-based tri-axial accelerometer fall detection algorithm. Gait & Posture 26(2), (2007) 3. Cleary, J., Trigg, L.: An instance-based learner using an entropic distance measure. In: Int. Conf. on Machine Learning. pp (1995) 4. Cover, T.M., Thomas, J.A.: Elements of Information Theory. Wiley (1992) 5. Cover, T.M., Thomas, J.A.: Data Mining: Practical machine learning tools and techniques. Morgan Kaufmann, San Francisco, 2nd edn. (2005) 6. Cucchiara, R., Prati, A., Vezzani, R.: A multi-camera vision system for fall detection and alarm generation. Expert Systems 24(5), (2007) 7. Foroughi, H., Naseri, A., Saberi, A., Yazdi, H.: An eigenspace-based approach for human fall detection using integrated time motion image and neural network. In: 9th Int. Conf. on Signal Processing. pp (2008) 8. Horn, B.: Robot Vision. The MIT Press, Cambridge, MA (1986) 9. Jansen, B., Deklerck, R.: Context aware inactivity recognition for visual fall detection. In: Proc. IEEE Pervasive Health Conference and Workshops. pp. 1 4 (2006) 10. Kepski, M., Kwolek, B., Austvoll, I.: Fuzzy inference-based reliable fall detection using Kinect and accelerometer. In: The 11th Int. Conf. on Artificial Intelligence and Soft Computing. pp LNCS, vol. 7267, Springer-Verlag (May 2012) 11. Kepski, M., Kwolek, B.: Human fall detection using Kinect sensor. In: Proc. of the 8th Int. Conf. on Computer Recognition Systems, Advances in Intelligent Systems and Computing, vol. 226, pp Springer (2013) 12. Labayrade, R., Aubert, D., Tarel, J.P.: Real time obstacle detection in stereovision on non flat road geometry through v-disparity representation. In: Intelligent Vehicle Symposium, IEEE. vol. 2, pp vol. 2 (june 2002) 13. Marshall, S.W., Runyan, C.W., Yang, J., Coyne-Beasley, T., Waller, A.E., Johnson, R.M., Perkis, D.: Prevalence of selected risk and protective factors for falls in the home. American J. of Preventive Medicine 8(1), (2005) 14. Mastorakis, G., Makris, D.: Fall detection system using Kinect s infrared sensor. J. of Real-Time Image Processing pp (2012) 15. Miaou, S.G., Sung, P.H., Huang, C.Y.: A customized human fall detection system using omni-camera images and personal information. Distributed Diagnosis and Home Healthcare pp (2006) 16. Mubashir, M., Shao, L., Seed, L.: A survey on fall detection: Principles and approaches. Neurocomputing 100, (2013), special issue: Behaviours in video 17. Noury, N., Fleury, A., Rumeau, P., Bourke, A., ÓLaighin, G., Rialle, V., Lundy, J.: Fall detection - principles and methods. In: Annual Int. Conf. of the IEEE Engineering in Medicine and Biology Society. pp (2007) 18. Rougier, C., Meunier, J., St-Arnaud, A., Rousseau, J.: Monocular 3D head tracking to detect falls of elderly people. In: Annual Int. Conf. of the IEEE Engineering in Medicine and Biology Society. pp (2006) 19. Shi, G., Chan, C.S., Li, W.J., Leung, K.S., Zou, Y., Jin, Y.: Mobile human airbag system for fall protection using MEMS sensors and embedded SVM classifier. Sensors Journal, IEEE 9(5), (2009)
Human fall detection on embedded platform using depth maps and wireless accelerometer
Kwolek and Kepski Human fall detection on embedded platform using depth maps and wireless accelerometer Bogdan Kwolek 1* and Michal Kepski 2 Abstract Since falls are a major public health problem in an
Automated Monitoring System for Fall Detection in the Elderly
Automated Monitoring System for Fall Detection in the Elderly Shadi Khawandi University of Angers Angers, 49000, France [email protected] Bassam Daya Lebanese University Saida, 813, Lebanon Pierre
An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network
Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal
Privacy Preserving Automatic Fall Detection for Elderly Using RGBD Cameras
Privacy Preserving Automatic Fall Detection for Elderly Using RGBD Cameras Chenyang Zhang 1, Yingli Tian 1, and Elizabeth Capezuti 2 1 Media Lab, The City University of New York (CUNY), City College New
Vision based approach to human fall detection
Vision based approach to human fall detection Pooja Shukla, Arti Tiwari CSVTU University Chhattisgarh, [email protected] 9754102116 Abstract Day by the count of elderly people living alone at home
Journal of Industrial Engineering Research. Adaptive sequence of Key Pose Detection for Human Action Recognition
IWNEST PUBLISHER Journal of Industrial Engineering Research (ISSN: 2077-4559) Journal home page: http://www.iwnest.com/aace/ Adaptive sequence of Key Pose Detection for Human Action Recognition 1 T. Sindhu
Neural Network based Vehicle Classification for Intelligent Traffic Control
Neural Network based Vehicle Classification for Intelligent Traffic Control Saeid Fazli 1, Shahram Mohammadi 2, Morteza Rahmani 3 1,2,3 Electrical Engineering Department, Zanjan University, Zanjan, IRAN
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering
Static Environment Recognition Using Omni-camera from a Moving Vehicle
Static Environment Recognition Using Omni-camera from a Moving Vehicle Teruko Yata, Chuck Thorpe Frank Dellaert The Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 USA College of Computing
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA N. Zarrinpanjeh a, F. Dadrassjavan b, H. Fattahi c * a Islamic Azad University of Qazvin - [email protected]
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
3D Scanner using Line Laser. 1. Introduction. 2. Theory
. Introduction 3D Scanner using Line Laser Di Lu Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute The goal of 3D reconstruction is to recover the 3D properties of a geometric
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Fall Detection System based on Kinect Sensor using Novel Detection and Posture Recognition Algorithm
Fall Detection System based on Kinect Sensor using Novel Detection and Posture Recognition Algorithm Choon Kiat Lee 1, Vwen Yen Lee 2 1 Hwa Chong Institution, Singapore [email protected] 2 Institute
Fall detection in the elderly by head tracking
Loughborough University Institutional Repository Fall detection in the elderly by head tracking This item was submitted to Loughborough University's Institutional Repository by the/an author. Citation:
Automatic Labeling of Lane Markings for Autonomous Vehicles
Automatic Labeling of Lane Markings for Autonomous Vehicles Jeffrey Kiske Stanford University 450 Serra Mall, Stanford, CA 94305 [email protected] 1. Introduction As autonomous vehicles become more popular,
Distributed Vision-Based Reasoning for Smart Home Care
Distributed Vision-Based Reasoning for Smart Home Care Arezou Keshavarz Stanford, CA 9435 [email protected] Ali Maleki Tabar Stanford, CA 9435 [email protected] Hamid Aghajan Stanford, CA 9435 [email protected]
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS Norbert Buch 1, Mark Cracknell 2, James Orwell 1 and Sergio A. Velastin 1 1. Kingston University, Penrhyn Road, Kingston upon Thames, KT1 2EE,
Vision-Based Blind Spot Detection Using Optical Flow
Vision-Based Blind Spot Detection Using Optical Flow M.A. Sotelo 1, J. Barriga 1, D. Fernández 1, I. Parra 1, J.E. Naranjo 2, M. Marrón 1, S. Alvarez 1, and M. Gavilán 1 1 Department of Electronics, University
Recognition of Day Night Activity Using Accelerometer Principals with Aurdino Development Board
Recognition of Day Night Activity Using Accelerometer Principals with Aurdino Development Board Swapna Shivaji Arote R.S. Bhosale S.E. Pawar Dept. of Information Technology, Dept. of Information Technology,
ECE 533 Project Report Ashish Dhawan Aditi R. Ganesan
Handwritten Signature Verification ECE 533 Project Report by Ashish Dhawan Aditi R. Ganesan Contents 1. Abstract 3. 2. Introduction 4. 3. Approach 6. 4. Pre-processing 8. 5. Feature Extraction 9. 6. Verification
A System for Capturing High Resolution Images
A System for Capturing High Resolution Images G.Voyatzis, G.Angelopoulos, A.Bors and I.Pitas Department of Informatics University of Thessaloniki BOX 451, 54006 Thessaloniki GREECE e-mail: [email protected]
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals Chun Lei He, Ping Zhang, Jianxiong Dong, Ching Y. Suen, Tien D. Bui Centre for Pattern Recognition and Machine Intelligence,
Low-resolution Character Recognition by Video-based Super-resolution
2009 10th International Conference on Document Analysis and Recognition Low-resolution Character Recognition by Video-based Super-resolution Ataru Ohkura 1, Daisuke Deguchi 1, Tomokazu Takahashi 2, Ichiro
Terrain Traversability Analysis using Organized Point Cloud, Superpixel Surface Normals-based segmentation and PCA-based Classification
Terrain Traversability Analysis using Organized Point Cloud, Superpixel Surface Normals-based segmentation and PCA-based Classification Aras Dargazany 1 and Karsten Berns 2 Abstract In this paper, an stereo-based
A DECISION TREE BASED PEDOMETER AND ITS IMPLEMENTATION ON THE ANDROID PLATFORM
A DECISION TREE BASED PEDOMETER AND ITS IMPLEMENTATION ON THE ANDROID PLATFORM ABSTRACT Juanying Lin, Leanne Chan and Hong Yan Department of Electronic Engineering, City University of Hong Kong, Hong Kong,
Activity recognition in ADL settings. Ben Kröse [email protected]
Activity recognition in ADL settings Ben Kröse [email protected] Content Why sensor monitoring for health and wellbeing? Activity monitoring from simple sensors Cameras Co-design and privacy issues Necessity
HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT
International Journal of Scientific and Research Publications, Volume 2, Issue 4, April 2012 1 HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT Akhil Gupta, Akash Rathi, Dr. Y. Radhika
How does the Kinect work? John MacCormick
How does the Kinect work? John MacCormick Xbox demo Laptop demo The Kinect uses structured light and machine learning Inferring body position is a two-stage process: first compute a depth map (using structured
Speed Performance Improvement of Vehicle Blob Tracking System
Speed Performance Improvement of Vehicle Blob Tracking System Sung Chun Lee and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA [email protected], [email protected] Abstract. A speed
Accurate, Fast Fall Detection Using Gyroscopes and Accelerometer-Derived Posture Information
Accurate, Fast Fall Detection Using Gyroscopes and Accelerometer-Derived Posture Information Qiang Li, John A. Stankovic, Mark Hanson, Adam Barth, John Lach University of Virginia {lq7c, stankovic, mah6s,
A Survey on Vision-based Fall Detection
A Survey on Vision-based Fall Detection Zhong Zhang, Christopher Conly, and Vassilis Athitsos Department of Computer Science and Engineering University of Texas at Arlington Arlington, Texas, USA [email protected],
Template-based Eye and Mouth Detection for 3D Video Conferencing
Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY V. Knyaz a, *, Yu. Visilter, S. Zheltov a State Research Institute for Aviation System (GosNIIAS), 7, Victorenko str., Moscow, Russia
Model-Based 3D Human Motion Capture Using Global-Local Particle Swarm Optimizations
Model-Based 3D Human Motion Capture Using Global-Local Particle Swarm Optimizations Tomasz Krzeszowski, Bogdan Kwolek, and Konrad Wojciechowski Abstract. We present an approach for tracking the articulated
Removing Moving Objects from Point Cloud Scenes
1 Removing Moving Objects from Point Cloud Scenes Krystof Litomisky [email protected] Abstract. Three-dimensional simultaneous localization and mapping is a topic of significant interest in the research
Real time vehicle detection and tracking on multiple lanes
Real time vehicle detection and tracking on multiple lanes Kristian Kovačić Edouard Ivanjko Hrvoje Gold Department of Intelligent Transportation Systems Faculty of Transport and Traffic Sciences University
Novelty Detection in image recognition using IRF Neural Networks properties
Novelty Detection in image recognition using IRF Neural Networks properties Philippe Smagghe, Jean-Luc Buessler, Jean-Philippe Urban Université de Haute-Alsace MIPS 4, rue des Frères Lumière, 68093 Mulhouse,
THE MS KINECT USE FOR 3D MODELLING AND GAIT ANALYSIS IN THE MATLAB ENVIRONMENT
THE MS KINECT USE FOR 3D MODELLING AND GAIT ANALYSIS IN THE MATLAB ENVIRONMENT A. Procházka 1,O.Vyšata 1,2,M.Vališ 1,2, M. Yadollahi 1 1 Institute of Chemical Technology, Department of Computing and Control
Human Activities Recognition in Android Smartphone Using Support Vector Machine
2016 7th International Conference on Intelligent Systems, Modelling and Simulation Human Activities Recognition in Android Smartphone Using Support Vector Machine Duc Ngoc Tran Computer Engineer Faculty
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences Byoung-moon You 1, Kyung-tack Jung 2, Sang-kook Kim 2, and Doo-sung Hwang 3 1 L&Y Vision Technologies, Inc., Daejeon,
T-REDSPEED White paper
T-REDSPEED White paper Index Index...2 Introduction...3 Specifications...4 Innovation...6 Technology added values...7 Introduction T-REDSPEED is an international patent pending technology for traffic violation
The Visual Internet of Things System Based on Depth Camera
The Visual Internet of Things System Based on Depth Camera Xucong Zhang 1, Xiaoyun Wang and Yingmin Jia Abstract The Visual Internet of Things is an important part of information technology. It is proposed
Multi-Person Pose Recognition Using a Zigbee Sensor Network
Proceedings of the 7th World Congress The International Federation of Automatic Control Seoul, Korea, July 6, 8 Multi-Person Pose Recognition Using a Zigbee Sensor Network Kai-Tai Song and Chun-Wei Chen
Poker Vision: Playing Cards and Chips Identification based on Image Processing
Poker Vision: Playing Cards and Chips Identification based on Image Processing Paulo Martins 1, Luís Paulo Reis 2, and Luís Teófilo 2 1 DEEC Electrical Engineering Department 2 LIACC Artificial Intelligence
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
A smartphone based real-time daily activity monitoring system. Shumei Zhang Paul McCullagh Jing Zhang Tiezhong Yu
A smartphone based real-time daily activity monitoring system Shumei Zhang Paul McCullagh Jing Zhang Tiezhong Yu Outline Contribution Background Methodology Experiments Contribution This paper proposes
Automatic Traffic Estimation Using Image Processing
Automatic Traffic Estimation Using Image Processing Pejman Niksaz Science &Research Branch, Azad University of Yazd, Iran [email protected] Abstract As we know the population of city and number of
THE proportion of the elderly population is rising rapidly
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS Fall Detection Based on Body Part Tracking Using a Depth Camera Zhen-Peng Bian, Student Member, IEEE, Junhui Hou, Student Member, IEEE, Lap-Pui Chau, Senior
How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm
IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode
Accurate, Fast Fall Detection Using Gyroscopes and Accelerometer-Derived Posture Information
Accurate, Fast Fall Detection Using Gyroscopes and Accelerometer-Derived Posture Information Qiang Li, John A. Stankovic, Mark A. Hanson, Adam T. Barth, John Lach University of Virginia Email: {lq7c, stankovic,
Data quality in Accounting Information Systems
Data quality in Accounting Information Systems Comparing Several Data Mining Techniques Erjon Zoto Department of Statistics and Applied Informatics Faculty of Economy, University of Tirana Tirana, Albania
VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS
VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS Aswin C Sankaranayanan, Qinfen Zheng, Rama Chellappa University of Maryland College Park, MD - 277 {aswch, qinfen, rama}@cfar.umd.edu Volkan Cevher, James
Support Vector Machine-Based Human Behavior Classification in Crowd through Projection and Star Skeletonization
Journal of Computer Science 6 (9): 1008-1013, 2010 ISSN 1549-3636 2010 Science Publications Support Vector Machine-Based Human Behavior Classification in Crowd through Projection and Star Skeletonization
International Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014
Efficient Attendance Management System Using Face Detection and Recognition Arun.A.V, Bhatath.S, Chethan.N, Manmohan.C.M, Hamsaveni M Department of Computer Science and Engineering, Vidya Vardhaka College
Face detection is a process of localizing and extracting the face region from the
Chapter 4 FACE NORMALIZATION 4.1 INTRODUCTION Face detection is a process of localizing and extracting the face region from the background. The detected face varies in rotation, brightness, size, etc.
Video Surveillance System for Security Applications
Video Surveillance System for Security Applications Vidya A.S. Department of CSE National Institute of Technology Calicut, Kerala, India V. K. Govindan Department of CSE National Institute of Technology
Shape Measurement of a Sewer Pipe. Using a Mobile Robot with Computer Vision
International Journal of Advanced Robotic Systems ARTICLE Shape Measurement of a Sewer Pipe Using a Mobile Robot with Computer Vision Regular Paper Kikuhito Kawasue 1,* and Takayuki Komatsu 1 1 Department
Mouse Control using a Web Camera based on Colour Detection
Mouse Control using a Web Camera based on Colour Detection Abhik Banerjee 1, Abhirup Ghosh 2, Koustuvmoni Bharadwaj 3, Hemanta Saikia 4 1, 2, 3, 4 Department of Electronics & Communication Engineering,
Tracking performance evaluation on PETS 2015 Challenge datasets
Tracking performance evaluation on PETS 2015 Challenge datasets Tahir Nawaz, Jonathan Boyle, Longzhen Li and James Ferryman Computational Vision Group, School of Systems Engineering University of Reading,
A Reliability Point and Kalman Filter-based Vehicle Tracking Technique
A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video
Face Recognition For Remote Database Backup System
Face Recognition For Remote Database Backup System Aniza Mohamed Din, Faudziah Ahmad, Mohamad Farhan Mohamad Mohsin, Ku Ruhana Ku-Mahamud, Mustafa Mufawak Theab 2 Graduate Department of Computer Science,UUM
An Active Head Tracking System for Distance Education and Videoconferencing Applications
An Active Head Tracking System for Distance Education and Videoconferencing Applications Sami Huttunen and Janne Heikkilä Machine Vision Group Infotech Oulu and Department of Electrical and Information
Geometric Camera Parameters
Geometric Camera Parameters What assumptions have we made so far? -All equations we have derived for far are written in the camera reference frames. -These equations are valid only when: () all distances
USING THE XBOX KINECT TO DETECT FEATURES OF THE FLOOR SURFACE
USING THE XBOX KINECT TO DETECT FEATURES OF THE FLOOR SURFACE By STEPHANIE COCKRELL Submitted in partial fulfillment of the requirements For the degree of Master of Science Thesis Advisor: Gregory Lee
A Real-Time Fall Detection System in Elderly Care Using Mobile Robot and Kinect Sensor
International Journal of Materials, Mechanics and Manufacturing, Vol. 2, No. 2, May 201 A Real-Time Fall Detection System in Elderly Care Using Mobile Robot and Kinect Sensor Zaid A. Mundher and Jiaofei
Synthetic Aperture Radar: Principles and Applications of AI in Automatic Target Recognition
Synthetic Aperture Radar: Principles and Applications of AI in Automatic Target Recognition Paulo Marques 1 Instituto Superior de Engenharia de Lisboa / Instituto de Telecomunicações R. Conselheiro Emídio
Real Time Target Tracking with Pan Tilt Zoom Camera
2009 Digital Image Computing: Techniques and Applications Real Time Target Tracking with Pan Tilt Zoom Camera Pankaj Kumar, Anthony Dick School of Computer Science The University of Adelaide Adelaide,
Automatic Fall Detector based on Sliding Window Principle
Automatic Fall Detector based on Sliding Window Principle J. Rodriguez 1, M. Mercuri 2, P. Karsmakers 3,4, P.J. Soh 2, P. Leroux 3,5, and D. Schreurs 2 1 UPC, Div. EETAC-TSC, Castelldefels, Spain 2 KU
Navigation Aid And Label Reading With Voice Communication For Visually Impaired People
Navigation Aid And Label Reading With Voice Communication For Visually Impaired People A.Manikandan 1, R.Madhuranthi 2 1 M.Kumarasamy College of Engineering, [email protected],karur,india 2 M.Kumarasamy
3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving
3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving AIT Austrian Institute of Technology Safety & Security Department Manfred Gruber Safe and Autonomous Systems
SmartMonitor An Intelligent Security System for the Protection of Individuals and Small Properties with the Possibility of Home Automation
Sensors 2014, 14, 9922-9948; doi:10.3390/s140609922 OPEN ACCESS sensors ISSN 1424-8220 www.mdpi.com/journal/sensors Article SmartMonitor An Intelligent Security System for the Protection of Individuals
VOLUMNECT - Measuring Volumes with Kinect T M
VOLUMNECT - Measuring Volumes with Kinect T M Beatriz Quintino Ferreira a, Miguel Griné a, Duarte Gameiro a, João Paulo Costeira a,b and Beatriz Sousa Santos c,d a DEEC, Instituto Superior Técnico, Lisboa,
Multimodal Biometric Recognition Security System
Multimodal Biometric Recognition Security System Anju.M.I, G.Sheeba, G.Sivakami, Monica.J, Savithri.M Department of ECE, New Prince Shri Bhavani College of Engg. & Tech., Chennai, India ABSTRACT: Security
3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving
3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving AIT Austrian Institute of Technology Safety & Security Department Christian Zinner Safe and Autonomous Systems
Keywords image processing, signature verification, false acceptance rate, false rejection rate, forgeries, feature vectors, support vector machines.
International Journal of Computer Application and Engineering Technology Volume 3-Issue2, Apr 2014.Pp. 188-192 www.ijcaet.net OFFLINE SIGNATURE VERIFICATION SYSTEM -A REVIEW Pooja Department of Computer
Integrated Monitoring System for Fall Detection in Elderly
Integrated Monitoring System for Fall Detection in Elderly 1 S. KHAWANDI, 2 B. DAYA, 3 P. CHAUVET 1 Phd student, University of Angers, Angers 49000-France 2 Prof., Lebanese University, 813-Lebanon 3 Prof.,
Interactive person re-identification in TV series
Interactive person re-identification in TV series Mika Fischer Hazım Kemal Ekenel Rainer Stiefelhagen CV:HCI lab, Karlsruhe Institute of Technology Adenauerring 2, 76131 Karlsruhe, Germany E-mail: {mika.fischer,ekenel,rainer.stiefelhagen}@kit.edu
SIGNATURE VERIFICATION
SIGNATURE VERIFICATION Dr. H.B.Kekre, Dr. Dhirendra Mishra, Ms. Shilpa Buddhadev, Ms. Bhagyashree Mall, Mr. Gaurav Jangid, Ms. Nikita Lakhotia Computer engineering Department, MPSTME, NMIMS University
A Method of Caption Detection in News Video
3rd International Conference on Multimedia Technology(ICMT 3) A Method of Caption Detection in News Video He HUANG, Ping SHI Abstract. News video is one of the most important media for people to get information.
Azure Machine Learning, SQL Data Mining and R
Azure Machine Learning, SQL Data Mining and R Day-by-day Agenda Prerequisites No formal prerequisites. Basic knowledge of SQL Server Data Tools, Excel and any analytical experience helps. Best of all:
Open-Set Face Recognition-based Visitor Interface System
Open-Set Face Recognition-based Visitor Interface System Hazım K. Ekenel, Lorant Szasz-Toth, and Rainer Stiefelhagen Computer Science Department, Universität Karlsruhe (TH) Am Fasanengarten 5, Karlsruhe
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: [email protected]
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
Projection Center Calibration for a Co-located Projector Camera System
Projection Center Calibration for a Co-located Camera System Toshiyuki Amano Department of Computer and Communication Science Faculty of Systems Engineering, Wakayama University Sakaedani 930, Wakayama,
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor Dominik Wȩgrzyn and Luís A. Alexandre IT - Instituto de Telecomunicações Dept. of Computer Science, Univ. Beira Interior, 6200-001 Covilhã, Portugal
Tracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey [email protected] b Department of Computer
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
E27 SPRING 2013 ZUCKER PROJECT 2 PROJECT 2 AUGMENTED REALITY GAMING SYSTEM
PROJECT 2 AUGMENTED REALITY GAMING SYSTEM OVERVIEW For this project, you will implement the augmented reality gaming system that you began to design during Exam 1. The system consists of a computer, projector,
LIST OF CONTENTS CHAPTER CONTENT PAGE DECLARATION DEDICATION ACKNOWLEDGEMENTS ABSTRACT ABSTRAK
vii LIST OF CONTENTS CHAPTER CONTENT PAGE DECLARATION DEDICATION ACKNOWLEDGEMENTS ABSTRACT ABSTRAK LIST OF CONTENTS LIST OF TABLES LIST OF FIGURES LIST OF NOTATIONS LIST OF ABBREVIATIONS LIST OF APPENDICES
Robot Perception Continued
Robot Perception Continued 1 Visual Perception Visual Odometry Reconstruction Recognition CS 685 11 Range Sensing strategies Active range sensors Ultrasound Laser range sensor Slides adopted from Siegwart
How To Filter Spam Image From A Picture By Color Or Color
Image Content-Based Email Spam Image Filtering Jianyi Wang and Kazuki Katagishi Abstract With the population of Internet around the world, email has become one of the main methods of communication among
