User, Device and Orientation Independent Human Activity Recognition on Mobile Phones: Challenges and a Proposal

Similar documents
People centric sensing Leveraging mobile technologies to infer human activities. Applications. History of Sensing Platforms

Vibration-Based Surface Recognition for Smartphones

Human Activities Recognition in Android Smartphone Using Support Vector Machine

Inactivity Recognition: Separating Moving Phones from Stationary Users

International Journal of Scientific & Engineering Research, Volume 4, Issue 8, August ISSN

Sensor Fusion Mobile Platform Challenges and Future Directions Jim Steele VP of Engineering, Sensor Platforms, Inc.

Effective Use of Android Sensors Based on Visualization of Sensor Information

Basic Principles of Inertial Navigation. Seminar on inertial navigation systems Tampere University of Technology

A smartphone based real-time daily activity monitoring system. Shumei Zhang Paul McCullagh Jing Zhang Tiezhong Yu

Enhancing survey methods and instruments to better capture personal travel behaviour

Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm

A Demonstration of a Robust Context Classification System (CCS) and its Context ToolChain (CTC)

A DECISION TREE BASED PEDOMETER AND ITS IMPLEMENTATION ON THE ANDROID PLATFORM

Social Data Mining through Distributed Mobile Sensing: A Position Paper

Accelerometer-Based Transportation Mode Detection on Smartphones

Cell Phone based Activity Detection using Markov Logic Network

International Journal of Advanced Research in Computer Science and Software Engineering

Motion Sensing with mcube igyro Delivering New Experiences for Motion Gaming and Augmented Reality for Android Mobile Devices

A Context Server to Allow Peripheral Interaction

Indoor Positioning using Sensor-fusion in Android Devices

VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS

Social Media Mining. Data Mining Essentials

Brewing Quality in Android Robots

Information regarding the Lockheed F-104 Starfighter F-104 LN-3. An article published in the Zipper Magazine #48. December Theo N.M.M.

Sensors. Marco Ronchetti Università degli Studi di Trento

Keywords - Accelerometer; Activity Recognition ; Clustered KNN; KNN; Smartphone. I. INTRODUCTION

HP TouchPad Sensor Setup for Android

Fitness Motion Recognition

STMicroelectronics is pleased to present the. SENSational. Attend a FREE One-Day Technical Seminar Near YOU!

Your AP Knows How You Move: Fine-grained Device Motion Recognition through WiFi

Speech Signal Processing: An Overview

Journal of Industrial Engineering Research. Adaptive sequence of Key Pose Detection for Human Action Recognition

BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES

Tracking and Recognition in Sports Videos

AN APPLYING OF ACCELEROMETER IN ANDROID PLATFORM FOR CONTROLLING WEIGHT

Kathy Au Billy Yi Fan Zhou Department of Electrical and Computer Engineering University of Toronto { kathy.au, billy.zhou }@utoronto.

Rotation: Moment of Inertia and Torque

E0-245: ASP. Lecture 16+17: Physical Sensors. Dipanjan Gope

Classifying Manipulation Primitives from Visual Data

VIRTUAL REALITY GAME CONTROLLED WITH USER S HEAD AND BODY MOVEMENT DETECTION USING SMARTPHONE SENSORS

An inertial haptic interface for robotic applications

How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm

Detecting User Activities using the Accelerometer on Android Smartphones

Automatic Labeling of Lane Markings for Autonomous Vehicles

Module 1: Sensor Data Acquisition and Processing in Android

Standardized Scalable Relocatable Context-Aware Middleware for Mobile applications (SCAMMP)

New Insights into WiFi-based Device-free Localization

A wireless sensor network for traffic surveillance

An Analysis of the Transitions between Mobile Application Usages based on Markov Chains

PeakVue Analysis for Antifriction Bearing Fault Detection

Multi-Person Pose Recognition Using a Zigbee Sensor Network

This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.

Charith Pereral, Arkady Zaslavsky, Peter Christen, Ali Salehi and Dimitrios Georgakopoulos (IEEE 2012) Presented By- Anusha Sekar

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

Magnetic Field Feature Analysis of Smartphone Application Activities Using Android MI Sensors

Vibrations can have an adverse effect on the accuracy of the end effector of a

Whitepaper. Image stabilization improving camera usability

Tracking devices. Important features. 6 Degrees of freedom. Mechanical devices. Types. Virtual Reality Technology and Programming

Frequently Asked Questions (FAQs)

Application Note IMU Visualization Software

USER STATE TRACKING USING SMARTPHONES

EnergyLens: Combining Smartphones with Electricity Meter for Accurate Activity Detection and User Annotation

Force/position control of a robotic system for transcranial magnetic stimulation

Multimodal Biometric Recognition Security System

Physics 11 Assignment KEY Dynamics Chapters 4 & 5

Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior

Anomaly Detection and Predictive Maintenance

A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation

NAVIGATING SCIENTIFIC LITERATURE A HOLISTIC PERSPECTIVE. Venu Govindaraju

ANDROID APPS DEVELOPMENT FOR MOBILE AND TABLET DEVICE (LEVEL II)

Context and Activity Recognition (aka Sensors + Inference ) Varun Manjunatha CMSC 818G

Comparing Methods to Identify Defect Reports in a Change Management Database

Data Mining - Evaluation of Classifiers

ANDROID APPLICATION DEVELOPMENT FOR ENVIRONMENT MONITORING USING SMART PHONES

Development of Integrated Management System based on Mobile and Cloud Service for Preventing Various Hazards

Tracking of Small Unmanned Aerial Vehicles

Matlab Based Interactive Simulation Program for 2D Multisegment Mechanical Systems

CHAPTER 5 PREDICTIVE MODELING STUDIES TO DETERMINE THE CONVEYING VELOCITY OF PARTS ON VIBRATORY FEEDER

Analyzing Driver Behavior using Smartphone Sensors: A Survey

The Calculation of G rms

Using Received Signal Strength Variation for Surveillance In Residential Areas

Development of Integrated Management System based on Mobile and Cloud service for preventing various dangerous situations

Advantages of Auto-tuning for Servo-motors

Sensors and Cellphones

PRODUCT DATASHEET. J1939 Vehicle Inertia Monitor. Advanced Vehicle Inertial Measurement and Vibration Monitoring Device. fleet-genius.

The accelerometer designed and realized so far is intended for an. aerospace application. Detailed testing and analysis needs to be

Robust Indoor Localization on a Commercial Smart-Phone

Context-aware Library Management System using Augmented Reality

Development of a Service Robot System for a Remote Child Monitoring Platform

An Introduction to Applied Mathematics: An Iterative Process

Position Estimation and Hospital Guidance Systems

Modeling and Performance Analysis of Hybrid Localization Using Inertial Sensor, RFID and Wi-Fi Signal

Towards the Counter Free Store: Requirements for Mobile Sales Assistants

Analecta Vol. 8, No. 2 ISSN

Predict the Popularity of YouTube Videos Using Early View Data

Physics 9e/Cutnell. correlated to the. College Board AP Physics 1 Course Objectives

Seventh IEEE Workshop on Embedded Computer Vision. Ego-Motion Compensated Face Detection on a Mobile Device

Automatic Fall Detector based on Sliding Window Principle

A secure face tracking system

Time Domain and Frequency Domain Techniques For Multi Shaker Time Waveform Replication

Transcription:

User, Device and Orientation Independent Human Activity Recognition on Mobile Phones: Challenges and a Proposal Yunus Emre Ustev NETLAB, Department of Computer Engineering Bogazici University, Bebek 34342 Istanbul/Turkey emre.ustev@boun.edu.tr Cem Ersoy NETLAB, Department of Computer Engineering Bogazici University, Bebek 34342 Istanbul/Turkey ersoy@boun.edu.tr Ozlem Durmaz Incel Peralab, Department of Computer Engineering Galatasaray University Besiktas 34349 Istanbul/Turkey odincel@gsu.edu.tr Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from permissions@acm.org UbiComp 13 Adjunct, September 8 12, 2013, Zurich, Switzerland. Copyright 2013 978-1-4503-2215-7/13/09 $15.00. Abstract Smart phones equipped with a rich set of sensors are explored as alternative platforms for human activity recognition in the ubiquitous computing domain. However, there exist challenges that should be tackled before the successful acceptance of such systems by the masses. In this paper, we particularly focus on the challenges arising from the differences in user behavior and in the hardware. To investigate the impact of these factors on the recognition accuracy, we performed tests with 20 different users focusing on the recognition of basic locomotion activities using the accelerometer, gyroscope and magnetic field sensors. We investigated the effect of feature types, to represent the raw data, and the use of linear acceleration for user, device and orientation-independent activity recognition. Author Keywords Human Activity Recognition, Mobile Phone Sensing ACM Classification Keywords I.5.2 [Design Methodology]: Feature evaluation and selection 1427

I. Introduction Human activity recognition systems using different sensing modalities, such as cameras or wearable inertial sensors, have been an active field of research in the domain of ubiquitous computing. Considering their potential to be applied in various application areas, including ambient assisted living, health and wellbeing monitoring, targeted advertisement, human activity recognition systems are becoming a part of our daily lives. Recently, smart phones, equipped with a rich set of sensors, are explored as alternative platforms for human activity recognition [1-5]. Besides the inclusion of sensors, such as accelerometer, compass, gyroscope, proximity, light, GPS, microphone, camera, the ubiquity, and unobtrusiveness of the phones and the availability of different wireless interfaces, such as Wi-Fi, 3G and Bluetooth, make them an attractive platform for human activity recognition. However, in spite of its advantages, human activity recognition on mobile phones also faces some challenges due to hardware limitations including battery, processing and storage constraints compared to more powerful stations. Due to the battery limitation, it is challenging to support continuous sensing applications as addressed in [6]. Considering the processing limitations, since human activity recognition requires running classification algorithms, originating from statistical machine learning techniques, it may be challenging to run resource-intensive classifiers on the phones [1, 6]. Other challenges stem from the different use of the phones by different people and the differences in the way people perform activities. For instance, the phone context problem [6] arises from the human behavior when the phone is carried in an inappropriate position relative to the event being sensed. Especially with human activity recognition using inertial sensors, location where the phone is carried, such as in the pocket or in the bag, and the orientation of the phone affect the activity recognition performance. The orientation of the phone is also a serious challenge in this area of research. A similar solution to our methodology was proposed in [9]. Gyroscope was used to detect the orientation of the mobile device. Also in [10], a filtering approach was used to estimate and remove the gravity from each axis of the accelerometer. Besides these challenges, since the classification of activities is mostly based on the use of statistical machine learning techniques, a learning phase is required. Mostly, supervised or semi-supervised learning techniques are utilized and such techniques rely on labeled data, i.e., associated with a specific class or activity. Labeling the data in the training phase is usually a tedious and complex process. In most of the cases, the user is required to label the activities and this, in turn, increases the burden on the user. Hence, user-independent training and activity recognition are required to foster the use of human activity recognition systems where the system can use the training data from other users in classifying the activities of a new subject. Besides the userindependent recognition, it is also important that human activity recognition algorithms should work on different mobile phone platforms in a deviceindependent manner for the acceptance of such systems by masses. In a recent study [1], we provided a review of activity recognition systems using mobile phones. We observe 1428

that motion-based activity recognition, using inertial sensors is the dominating type of activity recognition on mobile phones, besides the systems using locationbased activity recognition and motion-based activity recognition using wireless transceivers and other sensors, such as GPS. In the context of motion-based activity recognition systems, 3-axis accelerometers are the mostly utilized sensors available on the phones and most of the studies focus on detecting the locomotion activities, such as walking, standing, or transportation modes, biking, traveling with a vehicle. In this paper, we specifically focus on the challenges of user, device and orientation-independent activity recognition on mobile phones to accelerate the acceptance of such systems in practical applications. In particular, we focus on the recognition of five basic locomotion activities, including running, walking, biking, standing and sitting, using fused data from the accelerometer, the magnetic field sensor (magnetometer) and the gyroscope available on most of the smart phone platforms. As the initial step, we collect data from twenty different users carrying different types of mobile phones at different orientations. First we process the data using simple features, including the mean, the variance and the standard deviation, where the features are extracted from the square sum of the three acceleration components on x, y and z -axes and classify the activities using the K-nearest neighbors (KNN) classifier. Considering the results of this first set, we observed that although the activities can be recognized with a high accuracy in a user and device independent way, the orientation of the mobile phone impacts the accuracy of the results and the accuracy is found to be around 83%. Hence, as the next step, we focused on the use of extended set of features, including fast Fourier transform (FFT) coefficients and autocorrelation based features, for orientation-independent activity recognition and this increased the accuracy to 85%. In the second group of tests, we focused on how to improve the performance further, and explored the use of linear acceleration, excluding the effect of gravitational force. With this modification, the accuracy was increased to 93%. As the final step, we focused on the use of earth coordinates in the classification process, which further improved the classification performance to 97% accuracy. Main contributions of this paper can be described as follows: We investigate the existing problems of activity recognition using smart phone from several perspectives and proposed a robust methodology to overcome these challenges. We used multiple sensors in order to remove the effect gravity on the accelerometer readings and provide complete orientation independency by converting accelerometer readings from body coordinate system to earth coordinate system. The rest of the paper is organized as follows. In Section 2, we present the activity recognition system. In Section 3, we explain the test scenarios and the results of the experiments. Section 4 concludes the paper. II. Process of Activity Recognition The objective of our study is to investigate a possible solution for activity recognition on smart phones that will enable classification of five different daily activities with high accuracy, providing not only independency of orientation but also device/model and user 1429

Domain Time Time Frequency Frequency Extracted From Magnitude x, y and z axes Magnitude Autocorrellation of Magnitude Table 1. List of features Feature Mean, Variance, Std. Dev. Mean, Variance, Std. Dev. FFT Coefficients 1-10 Zero Crossing Rate, Max. Correlation Value, Period (Index of Max. Correlation) independency, as well. The activity recognition can be simply defined as the process of how to interpret the raw sensor data to classify a set of activities [1]. As mentioned, in the literature, statistical machine learning techniques are used to infer information about the activities from raw sensor readings and this process usually includes a training phase and a testing phase. The training phase requires the collection of labeled data to learn the model parameters. Once the data is collected, preprocessing, segmentation, feature extraction and classification steps follow to build the training model. In the following subsections, we explain the details of the activity recognition process proposed in this study. a. Feature Extraction The selection of proper features extracted from the accelerometer data plays an important role in the performance of activity recognition systems. We extracted several time and frequency domain features from the square sum of three acceleration components on x, y and z-axes. Simple time domain features like mean, variance and standard deviation from the magnitude of accelerometer vector and all three axes separately were calculated. In the extended feature set, frequency domain features were added to this list. In particular, the first 10 FFT coefficients and autocorrelation-based features like the zero crossing rate, the maximum correlation value and the index of the maximum correlation were computed. The list of features is presented in Table 1. The reason for focusing on this specific set is the fact that with these features, we can detect the period of an activity and this can differentiate activities with repeating motion patterns (e.g. walking) from those without a repetition. In the data collection phase, sampling rate was chosen as 100Hz. During the calculation of features, data was segmented into 2.56 seconds length windows. Some of the features evaluated in this study are computed using FFT and this calculation requires data sizes that are a power of two in order to avoid zero padding. So each window will contain 256 samples with selected sampling rate. b. Classification One of the most important factors to determine the performance of an activity recognition system is certainly the classifier algorithm. We decided to use the KNN classifier in the process, since among different alternatives; the KNN was shown to provide remarkably high recognition rates in this area of research. [3,7]. c. Linear Acceleration and Earth Coordinate System The linear acceleration can be defined as the acceleration without the effect of gravity. It can be calculated by the subtraction of gravity components on x, y and z axes from the same three axes of raw accelerometer values. The gravity component affecting each axis of an accelerometer can change according to the orientation of the phone. If the orientation can be calculated correctly, it would be possible to find the corresponding gravity vector being applied on each axis respectively. The orientation angles could be measured by using both the smart phone s accelerometer and the magnetic field sensor. If the device is not subject to 1430

Figure 1: Phone and Earth Coordinate System [8] Figure 2: Activity Recognition Block any other acceleration (the device is stationary or moves with constant speed), the accelerometer measures the gravity acceleration that points toward the center of the Earth, so with this information the tilt angles pitch and roll can be determined. Also, the magnetic field sensor provides the magnetic vector in the three axes of the device s coordinate system in orthogonal directions. This sensor could be utilized to derive the device s azimuth angle. With the fusion of the data from the accelerometer and the magnetic field sensor, it is possible to detect the orientation of device. But this combination has some drawbacks like noisy output from accelerometer or low response time and inaccurate readings from the magnetic field sensor. In order to overcome these limitations, the gyroscope was integrated into the system. The gyroscope has high response time and provides smooth outputs but its frailness comes from the fact that it drifts over time. The gyroscope provides rotation speeds relative to the phone s own coordinate system and this data can be used to correct errors caused by the accelerometer and the magnetic field sensor. With the fusion of these sensors, using the rotation vector which contains the rotation angles along each axis, it is possible to calculate the orientation of the phone. So, the gravitational force affecting on each axis of the accelerometer can be excluded and pure accelerometer values (linear acceleration) can be obtained. The magnetic field sensor can also be utilized as the compass to convert the acceleration output of the smart phone from its body frame of reference to the earth frame of reference. If all acceleration outputs are generated according to a single reference point, similar measurements can be observed for the same set of activities regardless of the orientation of the phone. As illustrated in Figure 1, using the magnetic field vector, the coordinate system of the phone could be converted to the earth magnetic coordinate system [8]. So fused sensor system generates the same acceleration output independent of orientation of the phone during the motion. Figure 2 shows the block diagram of activity recognition steps. In the data collection step, embedded sensors are used to compute gravity component in order to obtain dynamic acceleration value. This value can either be read in phone coordinate system or converted into earth coordinate system with the help of magnetic field sensor. Afterwards, time and frequency domain features are calculated for each segment of the data. In the final step, the classifier algorithm is used for the recognition of activities. III. Performance Evaluation In this section, we explain the performance of the activity recognition process explained in Section II. First, we present the experiment set up and the system parameters, and next, we will present the results and improvements in each section. a. Android Application Before starting the experiments, in order to collect the movement data from the subjects, an Android application is developed which gathers data from the embedded accelerometer, gyroscope and magnetic field sensors of the smart phone. Since it is a user-friendly application and it does not require any expertise to use, before performing the activity, user selects the ground truth label from the list, and then puts the phone into the pocket and performs the activity. 1431

Figure 3: Recognition rates using simple and extended feature sets. b. Experiment Scenarios The experiments were performed in two batches. In the first stage, recognition accuracy of the system was measured by different dependency conditions, such as orientation, device and user dependency, using only the accelerometer sensor. In the second stage, acceleration readings are isolated from the gravitational force and the linear acceleration data was obtained by enabling the use of the gyroscope and the magnetic field sensor which made it possible to convert the phone coordinate system to the earth coordinate system. Hence, the same tests were performed again with the same set of subjects to collect linear acceleration readings both in the phone and earth coordinate systems together. In order to detect the body movements of a subject, pockets of the trousers are selected as the most suitable place to carry the phone, as in most of the previous work [1-3, 7]. In the experiments, tests were performed to measure the orientation, device and user dependency of the activity recognition system. Hence, we had 4 different test cases performed by each individual: Orientation tests: carry the same phone in vertical/horizontal orientations to investigate orientation dependency Device tests: carry the same model phones (Device A/B) in the same pocket to find the effect of the device difference, i.e. calibration, on recognition accuracy Device-model tests: carry different model phones (Model X/Y) in the same pocket to measure the effect of the phone model User tests: all participants carry the same device to evaluate the user dependency. In the data collection stage, 16 male and 4 female, total of 20 healthy participants between the ages 18 and 59 were asked to perform five locomotion activities. All participants were wearing trousers and phones were placed into their pockets before performing each activity. All activities were performed for 3 minutes and a total 15 minutes of movement data was collected from every participant in each step of the experiments. Two Samsung Galaxy W and one Samsung Galaxy S3 model smart phones were used. In the orientation experiments, first, participants were asked to perform the given activities while the phone was placed vertically in their pockets whereas in the second step phone was placed horizontally. In the device dependency experiments, first, subjects were asked to carry the same model devices in the same pocket and then different models while they were performing the related activities again. Finally in the user dependency experiments, the data was collected from different participants by carrying the same device. In the classification phase, collected data from all individuals was evaluated separately and classification results were calculated on a subject basis. In the orientation dependency tests, while the data collected in one orientation was given as the training set, the data coming in the other orientation was used as the test data in the classifier. Again in the device dependency test, the data collected in one device was used as the test set and the data collected with the other device as the training set. In the user dependency tests, the leave-one-out approach was used. The collected data was processed offline: first, 1432

Truth Session: Ubiquitous Mobile Instrumentation Figure 4: Recognition rates using linear acceleration according to different coordinate systems the features are extracted from the raw data and then the classification was applied using the KNN classifier available in the Matlab statistics toolbox. c. Activity Recognition with Simple Features In the first attempt, the recognition accuracy of the system was measured using the simple time domain features. The mean, the variance and the standard deviation of each window were calculated using the magnitude of the accelerometer axes and also each of the axes separately. As a result, a feature vector with a length of 12 was obtained. This feature vector was reduced to 3 features that are the variance, the standard deviation of the magnitude and the mean of Z-axis. This reduction was performed by selecting the features which improve the recognition rates most using the sequential-forward selection. In Figure 3, we see that when the same model phones are carried in the same pocket, 96% accuracy can be achieved while it is 95% if the phone model is different. Also in the user dependency tests 91% accuracy was achieved. However, the accuracy in the orientation test was lower compared to the other tests, resulting in a 83% recognition rate. d. Activity Recognition with Extended Features In order to improve the overall performance of the system, we expanded the feature set with the frequency domain features described in Section II-a. In Figure 3, we observe that using more sophisticated frequency domain features improves the classification by 1-2% in all cases. As these results suggest, we can achieve acceptable accuracy results with different devices, models and users. However, when the orientation of the phone changes, the accuracies decrease. Hence, in the next step, we focused on how to improve the accuracy for the cases with different orientations of the phones. Also Table 1 shows that mostly there is a misclassification between periodic activities such as walking, biking and running and stationary activities such as sitting and standing. Classification Running Standing Biking Sitting Walking Running 82.9% 1.3% 8.5% 2.6% 4.7% Standing 0.5% 87.2% 1.5% 9.3% 1.5% Biking 4.5% 2.1% 83.3% 1.2% 8.9% Sitting 0.6% 8.9% 0.5% 89.1% 0.9% Walking 5.6% 1.2% 8.6% 0.5% 84.1% Table 1: Confusion matrix of orientation tests using only accelerometer with extended feature set e. Activity Recognition with Linear Acceleration After expanding the feature set with frequency domain features, we still encountered results with low accuracy in the orientation tests. The orientation of a phone can be recognized by fusing sensor data from the accelerometer, the gyroscope and the magnetic field sensor, hence, the gravity component on each axis can be isolated to obtain the linear acceleration values. In order to measure the effect of the linear acceleration on the recognition rates, new tests are performed with 20 different participants again. In these tests, the linear acceleration data referenced by the phone and earth coordinate systems were collected at the same time. Only the extended feature set is used in the classification since it provided better results in the previous tests. 1433

Truth Truth Session: Ubiquitous Mobile Instrumentation Classification Running Standing Biking Sitting Walking Running 93.5% 0.1% 4.5% 0.2% 1.7% Standing 0.1% 97.2% 0.1% 2.3% 0.3% Biking 5.4% 0.3% 87.3% 0.1% 6.9% Sitting 0.2% 3.4% 0.1% 96.1% 0.2% Walking 1.2% 0.1% 5.4% 0.2% 93.1% Table 2: Confusion matrix of orientation tests using linear acceleration with phone coordinate system As illustrated in Figure 4, in the orientation tests the recognition accuracy increased to 93% with the new dataset using the linear acceleration values when the phone coordinate system was selected as the reference. Also in the user-dependency tests, the recognition accuracy increased to 95%. Confusion matrix in Table 2 shows that the recognition rate of biking activity is still below the acceptable values. It is mostly confused with walking and running according to speed of pedaling. Classification Running Standing Biking Sitting Walking Running 97.5% 0.1% 1.4% 0.2% 0.8% Standing 0.1% 97.4% 0.3% 2.1% 0.1% Biking 2.9% 0.1% 95.2% 0.2% 1.6% Sitting 0.3% 2.3% 0.1% 97.1% 0.2% Walking 0.9% 0.1% 1.9% 0.2% 96.9% Table 3: Confusion matrix of orientation tests using linear acceleration with earth coordinate system Before the design decision of fusing the accelerometer, the gyroscope and the magnetic field sensors, we expected that if the movement of the smart phone could be measured according to a single reference system, the orientation in the pocket would lose its meaning. So we converted the linear acceleration readings from the phone coordinate system to the earth coordinate system. Figure 4 confirms our assumption that results of the orientation dependent tests boosted up to 97%. Also similar increase is visible in the user dependency tests with 97% accuracy. Also Table 3 shows that, there is a significant increase in the recognition accuracy of biking activity compared to results obtained using phone coordinate system. IV. Discussion and Conclusion In this paper, we focused on the challenges of practical activity recognition on smart phones. We specifically focused on the challenges arising from the differences in user behaviors and device, model differences as well as the burden of the training phase when using statistical machine learning algorithms in the classification of activities. Using the accelerometer, the magnetic field sensor and the gyroscope on the phones, we explored the recognition of simple locomotion activities in a device, user and orientation-independent way. With the experiments, performed with 20 users, we showed that although the accuracies are quite high in user and device dependency test, the difference in the orientation of the phone decreases the accuracy. To improve this, we proposed to use the linear acceleration, excluding the effect of gravitational force using the earth coordinates with the help of the gyroscope and the magnetic sensor. With this modification, we showed that the accuracy results increased remarkably. As a future work, we will investigate the impact of the device position, i.e., where the phone is carried, such as in the pocket or in the bag, on the performance. Also, we will investigate 1434

the effect of using multiple sensors on the energy consumption of the mobile phones. Acknowledgements This work is supported partially by the Bogazici University Research Fund under the grant numbers 11A01P7 and 11A01D4, and by the Galatasaray University Research Fund under the grant number 13.401.002. References [1] Incel, O.D., Kose, M., Ersoy, C, A Review and Taxonomy of Activity Recognition on Mobile Phones, Springer BioNanoScience Journal, Volume 3, Issue 2 (2013), Page 145-171, 2013. [2] Reddy, S., Mun, M., Burke, J., Estrin, D., Hansen, M., Srivastava, M.: Using mobile phones to determine transportation modes. ACM Trans. Sen. Netw. 6(2), 1 27 (2010) [3] Siirtola, P., Roning, J.: Recognizing human activities userindependently on smartphones based on accelerometer data. International Journal of Interactive Multimedia and Artificial Intelligence 1(5), 38{45} (2012) [4] Lane, N.D., Xu, Y., Lu, H., Hu, S., Choudhury, T., Campbell, A.T., Zhao, F.: Enabling large-scale human activity inference on smartphones using community similarity networks (csn). In: Proceedings of UbiComp 11, pp. 355 364. ACM, USA (2011) [5] Jun-geun Park Ami Patel, D.C.J.L.S.T.: Online pose classification and walking speed estimation using handheld devices. In: Proceedings of UbiComp 2012, pp. 113 122 (2012) [6] Lane, N., Miluzzo, E., Lu, H., Peebles, D., Choudhury, T., Campbell, A.: A survey of mobile phone sensing. Communications Magazine, IEEE 48(9), 140 150 (2010) [7] Kose, M., Incel, O.D., Ersoy, C.: Online human activity recognition on smart phones. In: Workshop on Mobile Sensing: From Smartphones and Wearables to Big Data, pp. 11-15., (2012). [8] Sensor Manager Tutorial, http://developer.android.com, 2013. [9] Blanke, U., & Schiele, B.: Sensing location in the pocket. Ubicomp Poster Session, 2008. [10] Yang, Jun.: Toward physical activity diary: motion recognition using simple acceleration features with mobile phones. Proceedings of the 1st international workshop on Interactive multimedia for consumer electronics. ACM, 2009. 1435