Mathematical Problems of Computer Science 40, 44---54, 2013. 44 Time Domain Feature Extraction and SVM Processing for Activity Recognition Using Smartphone Signals Sahak I. Kaghyan1 and Hakob G. Sarukhanyan2 1Armenian-Russian (Slavonic) University 2Institute for Informatics and Automation Problems of NAS RA e-mail: sahak.kaghyan@gmail.com, hakop@ipia.sci.am Abstract Automatic classification of human movement is a feature that is desired for a multitude of applications and mobile phone technology continuously evolves and incorporates more and more sensors to enable advanced applications. Combining these two concepts we can deal with “an activity recognition via smartphone sensors” problem where sensors of these devices play a core role when we deal with personalized activity tracking systems. In this paper we give an overview of the recent work in the field of activity recognition from mobile devices that can be attached to different parts of the body (pocket, wrist, and forearm). We focus on the technique of feature extraction from raw acceleration signal sequences of smartphone (mean, standard deviation, minimal and maximal signal values, correlation, median crossing). Further processing of these data allowed to classify the activity performed by the user. The core classification stage of the current approach was based on the method of “learning with the teacher” where the features of signal sequences were analyzed using the support vector machines (SVM) learning method. Keywords: Smartphone, Accelerometer, Activity recognition, Time domain feature, SVM. 1. Introduction Nowadays the Internet and personal computer are the most common ways to connect people, allowing them to share information with each other. On the other hand, none of them is able to reach each person anywhere and anytime like the cell phone does. Moreover, concerning the mobile technologies in general, now they are becoming ubiquitous all over the world, changing the way we communicate, conduct trade, and provide care and services. Certainly, some of the most compelling benefits of mobile technologies are in the areas of disease prevention, chronic disease management and improving healthcare delivery. For all the advances that occur in mobile healthcare (mHealth), its full potential for a very large group of beneficiaries – the elderly and those who support them – is S. Kaghyan, H. Sarukhanyan 45 only starting to emerge. The ability to monitor the physical state of a person leads us to the concept of personalized healthcare system implementation. One of the ways to help people with diseases is to give the doctors an opportunity to remotely monitor their patients’ life activity via cellular phones and smartphones they care. Human activity recognition (HAR) has matured in recent years which will enable many health promotions and intervention applications. There are no standardized performance evaluation strategies. Recent efforts on designing public datasets might be one of the approaches to address this problem. Generally, activity recognition (AR) aims to identify the actions taken by a person. Three main classes of activity recognition are considered including coarse location tracking, video stream analysis and inertial navigation systems (INS) such as accelerometers. Sensor data are typically communicated from sensors to servers for further processing. Alternatively signal processing can be performed in mobile devices such as smart-phones. Many authors usually don’t use standard tests for accuracy rate checks and validity of most re ported results depends on testing specifics. There is no consensus even on a standard list of activities, but most of the reports include “walking”, “sitting”, “jogging” and “standing” patterns. Recognition can be accomplished, for example, by exploiting the information retrieved from inertial sensors such as accelerometers [4]. In some smartphones these sensors are embedded by default and we benefit from this to classify a set of physical activities (standing, walking, laying, , walking upstairs and walking downstairs) by processing inertial body signals through a supervised Machine Learning (ML) algorithm for hardware with limited resources. So, in general, activity recognition algorithms can be divided into two major categories. The first one is based on supervised and unsupervised machine learning methods. Supervised learning requires the use of labeled data upon which an algorithm is trained. 2. Related Works Recognizing a predefined set of activities is a recognition (classification) task: features are extracted from the space-time information collected by sensors and then used for classification. Feature representations are used to map the data to another representation space with the intention of making the classification problem easier to solve. In most cases, a model of classification is used that relates the activity to sensor patterns. Learning of such models is usually done in a supervised manner (human labeling) and requires a large annotated datasets recorded in different settings. Smart phones include various sensors such as gyroscopes, accelerometers, proximity sensors and have become affordable and ubiquitous. Convenient user interfaces make them attractive for all population groups. Oner et al. in [1] presented an early work on a pedometer mobile application that was coupled with e-mail to notify medical assistants or family members. Their purpose was to use a mobile smart phone to detect the fall event regardless of the phone position or orientation. The algorithm that was introduced in the article was based on the acceleration peak detection and was tested for different conditions. Das et al in [3] introduced an attempt to recognize the activity using Motorola Droid smartphone. Activity classification was done through several stages: data acquisition, signal processing, feature extraction and classification. Using the nearest neighbor classifier the program could predict patterns or activities with 93% accuracy after it had been calibrated for a particular user. One of our early works [4] introduced a general method of classification that used the nearest neighbor method and showed 80% of accuracy. Time Domain Feature Extraction and SVM Processing for Activity Recognition Using Smartphone Signals46 3. Feature Extraction Concept There are many terms, used almost interchangeably, or the process of extracting the important features from a set of data, including “feature extraction,” “feature (subset) selection,” “data mining,” “information content” and “(feature) dimension reduction.” Whatever the term used, the process of feature extraction can be defined as “a process of identifying valid, useful and understandable patterns in data”. The large amount of time-series data that is generated in a gait analysis study is called “high dimensional” data. High dimensional data can be thought of simply as lots of data (there are lots of dimensions to the data). Fast developing gait measuring techniques and methodologies generate more and more data. This in turn results in what is known as the “curse of dimensionality” (more and more data to manage, more dimensions). The objective of feature extraction is to keep all the useful features of the data and discard all the redundant parts of the data. There are two required outcomes to this process: (1) reduce the amount of data to a manageable level (dimension reduction), and (2) keep the most important features of the data and eliminate all the redundant features of the data (feature selection). The idea is to provide a “summary” which can be used to give a meaningful interpretation of the data. The objective of the first step towards the feature selection is a dimension reduction, to reduce the search space to a lower, more manageable dimensionality and this has to be achieved in a way that retains relevant features of the data and removes irrelevant features. That is, the feature selection selects “m” relevant features from the entire set of “n” features such that m