Defining a Roadmap Towards Comparative Research in Online Activity
Recognition on Mobile Phones
Muhammad Shoaib
1
, Stephan Bosch
1
, Ozlem Durmaz Incel
2
, Hans Scholten
1
and Paul J. M. Havinga
1
1
Pervasive Systems Group, Department of Computer Science, University of Twente, Enschede, The Netherlands
2
Department of Computer Science, Galatasary University, Istanbul, Turkey
Keywords:
Online Activity Recognition, Real-time, Mobile Phones, Context-awareness, Accelerometer, Smartphones.
Abstract:
Many context-aware applications based on activity recognition are currently using mobile phones. Most of this
work is done in an offline way. However, there is a shift towards an online approach in recent studies, where
activity recognition systems are implemented on mobile phones. Unfortunately, most of these studies lack
proper reproducibility, resource consumption analysis, validation, position-independence, and personalization.
Moreover, they are hard to compare in various aspects due to different experimental setups. In this paper,
we present a short overview of the current research on online activity recognition using mobile phones, and
highlight their limitations. We discuss these studies in terms of various aspects, such as their experimental
setups, position-independence, resource consumption analysis, performance evaluation, and validation. Based
on this analysis, we define a roadmap towards a better comparative research on online activity recognition
using mobile phones.
1 INTRODUCTION
Activity recognition is being used in various re-
search areas (Incel et al., 2013; Marin-Perianu et al.,
2008). Especially, mobile phones have enabled many
context-aware applications based on activity recog-
nition, because they are equipped with various on-
board sensors, such as an accelerometer, gyroscope,
and GPS (Lara and Labrador, 2013). Moreover, they
are used by most people in their daily life these days.
Research on activity recognition using mobile phones
is done mainly in two ways. The first is the offline
approach, where the data from various mobile phone
sensors is collected using data collection tools. This
data is analyzed later on using machine learning tools
such as WEKA (Hall et al., 2009; Shoaib et al., 2013).
The second is the online approach, where data collec-
tion is done on a mobile phone and data analysis is
done online either locally on the mobile phone (An-
jum and Ilyas, 2013; Lane et al., 2011) or on a remote
server in real time (Gil et al., 2011; Lau et al., 2010).
Most of this research has been done in an offline
way (Incel et al., 2013; Shoaib et al., 2014). How-
ever, for the practical usability of such activity recog-
nition systems, it is important to evaluate them on-
line on mobile phones, which is a challenging task.
In the current research, we see a shift towards an on-
line approach. There are ample surveys which discuss
studies using the offline approach (Lara and Labrador,
2013; Incel et al., 2013; Liang et al., 2013). How-
ever, studies using the online approach are yet to be
surveyed or analyzed. Moreover, to be able to eval-
uate and analyze online activity recognition methods,
research should be aligned in terms of defining test
setups, metrics, and platforms. In this work, we fill
this gap by presenting a short overview of the cur-
rent research on online activity recognition on mobile
phones, and highlighting its limitations. We discuss
various aspects of these studies, such as their exper-
imental setups, position-independence, performance
evaluation, resource consumption analysis, and vali-
dation. Unfortunately, most of the studies using an
online approach lack important aspects, such as re-
producibility, proper validation, personalization, clas-
sifiers’ comparison, and resource consumption analy-
sis. Based on our analysis, we present a roadmap with
various recommendations for future studies on online
activity recognition. We believe this work will help in
a better comparative research in this area.
The rest of the paper is organized as follows. In
Section 2, we describe various aspects online activity
recognition studies using mobile phones. In Section
3, we present a roadmap for future studies. Finally,
we conclude this paper in Section 4.
154
Shoaib M., Bosch S., Durmaz Incel O., Scholten H. and J. M. Havinga P..
Defining a Roadmap Towards Comparative Research in Online Activity Recognition on Mobile Phones.
DOI: 10.5220/0005324401540159
In Proceedings of the 5th International Conference on Pervasive and Embedded Computing and Communication Systems (PECCS-2015), pages
154-159
ISBN: 978-989-758-084-0
Copyright
c
2015 SCITEPRESS (Science and Technology Publications, Lda.)
2 ONLINE ACTIVITY
RECOGNITION STUDIES
Online physical activity recognition can be imple-
mented in two main ways on mobile phones. In first
case, the main steps in the activity recognition pro-
cess are performed locally on a mobile device, which
is termed as local recognition. The main advantage
of this approach is that no internet connection is re-
quired. For example, this type approach is adapted in
(Anjum and Ilyas, 2013; Lane et al., 2011). In the sec-
ond case, these steps are either fully or partially done
on a remote server or in a cloud, which is termed as
remote recognition. This method may result in less
processing burden on the mobile device, however,it
will require an internet connection at all times. This
type of approach is adapted in (Gil et al., 2011; Lau
et al., 2010).
We target only mobile phones as the main entity in
the activity recognition process. Therefore, we only
consider the studies utilizing strictly local recognition
in this paper. Moreover, we also considered the fol-
lowing criteria in selecting these studies:
They use only mobile phone sensors, where mo-
tion sensors are used in a lead role in the recog-
nition process with other on-board sensors in the
supporting role. We do not consider video-based
activity recognition.
They are able to recognize different physical ac-
tivities.
We found 29 publications based on these selec-
tion criteria (Anjum and Ilyas, 2013; Berchtold et al.,
2010; Das et al., 2012; Das et al., 2010; Frank et al.,
2011; Gomes et al., 2012; Guiry et al., 2012; Khan
et al., 2013; Khan et al., 2014; Kim et al., 2013;
Kose, Mustafa et al., 2012; Lane et al., 2011; Lara and
Labrador, 2012; Liang et al., 2012; Lu et al., 2010;
Martin et al., 2013; Miluzzo et al., 2008; Ouchi and
Doi, 2012; Reddy et al., 2010; Ryder et al., 2009;
Schindhelm, 2012; Siirtola and Roning, 2013; Siir-
tola, 2012; Stewart et al., 2012; Thiemjarus et al.,
2013; Vo et al., 2013; Wang et al., 2009; Yan et al.,
2012; Zhao et al., 2013). In terms of time line, they
range from 2008 to 2014. We analyze these studies
in various aspects which are listed as follows: train-
ing and classification, experimental setups, position-
independence, resource consumption analysis, perfor-
mance evaluation, and validation.
2.1 Training and Classification
Classification is the core part of the activity recogni-
tion process. A classification step typically involves
two steps: training and testing (classifying new ac-
tivities based on the training). In the analyzed 29
studies, various classifiers are implemented. They are
presented in Figure 1 in relation to how many stud-
ies implemented each of the mentioned classifiers in
the figure. Based on this figure, decision tree, support
vector machines, naive bayes, and KNN are the most
commonly implemented classifiers. Decision tree is
implemented in 11 studies because it is a light-weight
classifier, which make it suitable for running on mo-
bile phones.
These classifiers can be trained in two ways: on-
line or offline. Most of these studies have trained
their classifiers offline, which makes them static, and
they cannot be trained in real-time for new activi-
ties. Moreover, these classifiers cannot be dynami-
cally personalized for individual users. This causes a
lack of personalization in such solutions. We found
that only 17% of all the studies have the capability to
train their classifiers online on the mobile phones as
shown in Figure 2. Online training is an important
feature, as it gives the users the flexibility to train the
systems as per their needs in real time, which will lead
to better personalization in such solutions. However,
83% of the studies are still missing the online training
capability. There is only one study (Berchtold et al.,
2010) in this category that uses online training, but
it does so on a remote server and not locally on the
mobile phone.
1
1
1
1
1
3
4
5
5
11
0 5 10 15
Artificial Neural Networks
Rule-based Classifier
Quadratic Discriminant Analysis
Decision Table
Fuzzy Classification
Multilayer Classifiers
K-Nearest Neighbor
SVM
Naïve Bayes
Decision Tree
Classifiers
Implemented Classifiers
number of Studies
Figure 1: Number of studies vs implemented classifiers.
17%
83%
Classifiers
Training Process
Studies with
online training
Studies with
offline training
Figure 2: Offline vs Online training process.
DefiningaRoadmapTowardsComparativeResearchinOnlineActivityRecognitiononMobilePhones
155
2.2 Experimental Setup
There are four common operating system platforms
used for developing activity recognition solutions in
the analyzed studies. These are: Android, iOS, Sym-
bian, and Debian Linux. Most of these studies use
Android. Initially, Symbian was used in a few stud-
ies, but recently the focus is more on Android because
it is the popular operating system among users these
days. For example, we found Symbian in almost all
studies in between 2008 and 2010, but from 2011 and
onwards, Android gained popularity. The distribution
of these platforms among all 29 studies is given in
Figure 3.
These studies use different sensors for activity
recognition, either alone or in various combinations.
Usually different sensors are combined for better ac-
tivity recognition. A large portion of these studies use
only the accelerometer. However, in some cases, it is
used in combination with other sensors such as GPS,
gyroscope, microphone, magnetometer, and pressure
sensor. The accelerometer is used alone in 76% of
the analyzed studies and its combination with other
sensors counts for 24% of the analyzed studies.
22
8
3
1
0
5
10
15
20
25
Android Symbian iOS Debian
Linux
Mobile Operating Systems
Mobile Platforms vs Number Of Studies
Number of
studies
Figure 3: Platforms used for activity recognition systems.
Mobile Phone Sensors
Studies with accelerometer
in combination with other
sensors (GPS,
Magnetometer, Pressure
sensor, Gyroscope,
Microphone,
Linear Acceleration Sensor,
Gravity Sensor)
24%
Combination
76% Alone
Figure 4: Sensors used in activity recognition systems.
2.3 Position-independence
In most of the existing studies, when data is collected
for physical activity recognition, the mobile device is
kept on a fixed body position. Therefore, the trained
classifier will only work well for that specific position
and can produce poor results if used with other posi-
tions. This is because some sensors like accelerom-
eter and gyroscope are position-sensitive. There are
three ways to counter this problem (Thiemjarus et al.,
2013). First way is to use position-insensitive data
features. The second solution is to train classifiers
for all possible positions and then dynamically de-
tect the position first before classifying an activity.
This is called a position-aware classifier. In the third
way, various signal transformations are used to mini-
mize the effects of changing positions. Solutions that
use a fixed position, such as a pants pocket, may not
be practical. Users cannot be forced to always keep
the mobile device in a specific position. It is still a
challenging problem to have a completely position-
independent solution while using position-sensitive
sensors like accelerometer and gyroscope. However,
there are some studies which have tried to solve this
problem to some extent. We highlight these studies in
Figure 5. There were some studies where there was
no information available about the position. We be-
lieve these studies assume a fixed position, because
position-independence is an important contribution
and should have been mentioned explicitly in these
studies if considered. Figure 5 shows that only 34%
of the studies considered the position-independence
property, therefore future studies should consider this
feature for a practical and user-friendly activity recog-
nition system. It is important to note that orientation-
independence is also an important aspect for activity
recognition but we do not consider it in this study.
34%
4
%
7%
Stu
ies ith Position n epen ence
Position-Independent
Fixed-Position
No Information
available
Figure 5: Studies with the position-independence.
2.4 Validation
To validate various aspects of a physical activity
recognition system, it should be tested with a rea-
sonable duration and with a reasonable number of
users. These aspects can be memory, CPU, and bat-
tery usage and its tradeoff with recognition perfor-
mance, such as accuracy. The reasonable duration of
the testing and the number of users can vary in dif-
ferent applications and should be decided by the re-
searchers accordingly. We evaluated all 29 studies in
terms of the duration of the testing and the number of
users involved. It is clear from Figure 6 that 18 stud-
ies are missing the testing duration information, out
of which 9 studies have no information on how many
users were involved in the testing process. For the
PECCS2015-5thInternationalConferenceonPervasiveandEmbeddedComputingandCommunicationSystems
156
Figure 6: Testing duration for implemented systems.
rest of the studies the number of users ranges from 1
to 21. The lack of proper validation in these could be
caused by the fact that many of them are still a work
in progress.
2.5 Resource Consumption Analysis
One of the important aspects of evaluating online ac-
tivity recognition systems is to measure the resource
consumption beside the recognition performance, be-
cause it enables us to compare different systems in a
fair way. We found three resources which were eval-
uated in general in these studies. These are: mem-
ory, CPU and battery usage, and they are measured in
MBs (memory), usage percentage (CPU), and number
of hours or watt-hour per hour (battery). The battery,
memory, and CPU usage is measured in 38%, 21%,
and 31% of all the studies, respectively.
Battery consumption is an important criterion for
the feasibility of activity recognition solutions us-
ing mobile phones. It is still a challenge to propose
energy-efficient solutions which do not drain the mo-
bile phone’s battery quickly. Some of these studies
(Guiry et al., 2012; Lane et al., 2011; Liang et al.,
2012) use the number of hours for which the battery
lasts while running the activity recognition applica-
tion as a resource consumption metric. This can be
misleading because of the different battery capacities
in various mobile phones as well as how old these bat-
teries are. A relatively better metric can be the watt-
hour per hour which shows the amount of watt-hours
used by the activity recognition application in specific
time duration. This metric has been used in some of
the studies (Miluzzo et al., 2008). However, this is
also not platform-independent, because the efficiency
of the hardwaresuch as CPU and the operating system
is also an important factor for the battery life. These
factors can vary wildly in various phones. There is a
still need for a metric which can be used across multi-
ple platforms like Android, iOS and Symbian. These
aspects need further research.
2.6 Performance Evaluation
Most of these studies use classification accuracy as a
performance metric for evaluation. Various physical
activities were used for evaluating the classification
performance of different classifiers. The most com-
monly used activities are: walking, running, walking
upstairs and downstairs, sitting, standing, and biking.
We do not present the accuracy values in this paper
for comparative analysis, because it is hard to com-
pare them due to different experimental setups. For
example, these studies use different sensor combina-
tions, classifiers, data features, sampling rates for data
collection, and sets of physical activities.
3 ROADMAP FOR FUTURE
STUDIES
There has been significant progress in the shift from
the offline towards the online approach. This work
is still in its infancy and further research needs to be
done to explore this area in detail. Moreover,it is hard
to compare various studies because of their differ-
ent experimental setups and validation methods. We
present the following recommendationswhich need to
be taken into account while conducting future studies
on online activity recognition.
Reproducibly: For reproducibility, it is necessary
that important details should be clearly presented in
the paper. For example, some information were miss-
ing in some of the reviewed studies such as the sam-
pling rate , the position-independence property, and
validation details. There were five studies (Frank
et al., 2011; Gomes et al., 2012; Lane et al., 2011;
Miluzzo et al., 2008; Wang et al., 2009), where we
did not find information about the sampling rate for
sensor data collection.
Proper Validation. These systems should be prop-
erly validated with a reasonable number of users and
for a reasonable amount of time. This reasonable
value can be defined in the context of each specific
study and is to be explored.
Resource Consumption Analysis. For a fair com-
parison of various classifiers, a tradeoff between the
recognition results and the resources consumed in that
process are both important. That is why resources
consumption analysis should be considered in future
studies. Moreover, there is a need for a platform-
and device-independent performance metric to the re-
sources consumed by such systems, especially for
battery consumption.
Personalization. To make a step towards person-
alization, providing online training support should be
DefiningaRoadmapTowardsComparativeResearchinOnlineActivityRecognitiononMobilePhones
157
considered in the future studies, because it is impor-
tant for its practical usability. Online training support
was missing in 85% of the reviewed studies, which
means future studies should concentrate on this aspect
besides online testing.
Position-independence. More efforts should be
done towards activity recognition systems that are not
dependent on a specific body position, thereby giving
users more freedom. Moreover, it should be explic-
itly mentioned if this feature was considered or not
for clarity unlike some of the existing studies.
Comparing Classifiers. Most of the existing stud-
ies implement one or two classifiers with a specific
experimental setup. This makes it hard to compare
multiple classifiers because of the different experi-
mental setups and validation methods. There is a need
for studies where multiple classifiers are validated in
the same experimental environments to have a fair
comparison about their performance results and re-
source consumption. Such type of studies can play a
role in defining a benchmark for future online activity
recognition research.
4 CONCLUSIONS
We reviewed the current research on online activity
recognition using mobile devices which only uses on-
board sensors and do local processing. Though there
has been a significant amount of progress done in this
direction, there is still room for improvements. We
discussed various aspects of the existing studies such
as experimental setups, resource consumption, and
validation. We presented an overview of the current
research, and a roadmap for future studies with vari-
ous recommendations. In future, we intend to explore
these aspects further in depth.
ACKNOWLEDGEMENTS
This work is supported by Dutch National Program
COMMIT in the context of SWELL project, by the
Galatasaray UniversityResearch Fund under the grant
number 13.401.002 and by Tubitak under the grant
number 113E271.
REFERENCES
Anjum, A. and Ilyas, M. (2013). Activity recognition using
smartphone sensors. In 2013 IEEE Consumer Com-
munications and Networking Conference (CCNC),
pages 914–919.
Berchtold, M., Budde, M., Gordon, D., Schmidtke, H., and
Beigl, M. (2010). ActiServ: Activity recognition ser-
vice for mobile phones. In 2010 International Sympo-
sium on Wearable Computers (ISWC), pages 1–8.
Das, B., Seelye, A., Thomas, B., Cook, D., Holder, L.,
and Schmitter-Edgecombe, M. (2012). Using smart
phones for context-aware prompting in smart environ-
ments. In 2012 IEEE Consumer Communications and
Networking Conference (CCNC), pages 399–403.
Das, S., Green, L., Perez, B., Murphy, M., and Perring,
A. (2010). Detecting user activities using the ac-
celerometer on android smartphones. The Team for
Research in Ubiquitous Secure Technology, TRUST-
REU Carnefie Mellon University.
Frank, J., Mannor, S., and Precup, D. (2011). Activity
recognition with mobile phones. In Machine Learn-
ing and Knowledge Discovery in Databases, number
6913 in Lecture Notes in Computer Science, pages
630–633. Springer Berlin Heidelberg.
Gil, G. B., Berlanga, A., and Molina, J. M. (2011). Physical
actions architecture: Context-aware activity recogni-
tion in mobile devices. In User-Centric Technologies
and Applications, pages 19–27. Springer.
Gomes, J., Krishnaswamy, S., Gaber, M., Sousa, P., and
Menasalvas, E. (2012). MARS: A personalised mo-
bile activity recognition system. In 2012 IEEE 13th
International Conference on Mobile Data Manage-
ment (MDM), pages 316–319.
Guiry, J. J., Ven, P. v. d., and Nelson, J. (2012). Orienta-
tion independent human mobility monitoring with an
android smartphone. ACTAPRESS.
Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann,
P., and Witten, I. H. (2009). The WEKA data min-
ing software: An update. SIGKDD Explor. Newsl.,
11(1):10–18.
Incel, O. D., Kose, M., and Ersoy, C. (2013). A review and
taxonomy of activity recognition on mobile phones.
BioNanoScience, 3(2):145–171.
Khan, A. M., Siddiqi, M. H., and Lee, S.-W. (2013).
Exploratory data analysis of acceleration signals to
select light-weight and accurate features for real-
time activity recognition on smartphones. Sensors,
13(10):13099–13122.
Khan, A. M., Tufail, A., Khattak, A. M., and Laine,
T. H. (2014). Activity recognition on smartphones via
sensor-fusion and KDA-based SVMs. International
Journal of Distributed Sensor Networks, 2014.
Kim, T.-S., Cho, J.-H., and Kim, J. T. (2013). Mobile
motion sensor-based human activity recognition and
energy expenditure estimation in building environ-
ments. In Sustainability in Energy and Buildings,
number 22 in Smart Innovation, Systems and Tech-
nologies, pages 987–993. Springer Berlin Heidelberg.
Kose, Mustafa, Incel, O. D., and Ersoy, C. (2012). On-
line human activity recognition on smart phones. In
Workshop on Mobile Sensing: From Smartphones and
Wearables to Big Data,, pages 11–15.
Lane, N., Mohammod, M., Lin, M., Yang, X., Lu, H., Ali,
S., Doryab, A., Berke, E., Choudhury, T., and Camp-
PECCS2015-5thInternationalConferenceonPervasiveandEmbeddedComputingandCommunicationSystems
158
bell, A. (2011). BeWell: A smartphone application to
monitor, model and promote wellbeing. IEEE.
Lara, O. and Labrador, M. (2012). A mobile platform for
real-time human activity recognition. In 2012 IEEE
Consumer Communications and Networking Confer-
ence (CCNC), pages 667–671.
Lara, O. D. and Labrador, M. A. (2013). A survey on human
activity recognition using wearable sensors. Com-
munications Surveys & Tutorials, IEEE, 15(3):1192–
1209.
Lau, S. L., Konig, I., David, K., Parandian, B., Carius-
Dussel, C., and Schultz, M. (2010). Supporting patient
monitoring using activity recognition with a smart-
phone. In Wireless Communication Systems (ISWCS),
2010 7th International Symposium on, pages 810–
814. IEEE.
Liang, Y., Zhou, X., Yu, Z., and Guo, B. (2013). Energy-
efficient motion related activity recognition on mobile
devices for pervasive healthcare. Mobile Networks
and Applications, pages 1–15.
Liang, Y., Zhou, X., Yu, Z., Guo, B., and Yang, Y. (2012).
Energy efficient activity recognition based on low res-
olution accelerometer in smart phones. In Advances
in Grid and Pervasive Computing, number 7296 in
Lecture Notes in Computer Science, pages 122–136.
Springer Berlin Heidelberg.
Lu, H., Yang, J., Liu, Z., Lane, N. D., Choudhury, T., and
Campbell, A. T. (2010). The jigsaw continuous sens-
ing engine for mobile phone applications. In Pro-
ceedings of the 8th ACM Conference on Embedded
Networked Sensor Systems, SenSys ’10, pages 71–84,
New York, NY, USA. ACM.
Marin-Perianu, M., Lombriser, C., Amft, O., Havinga, P.,
and Tr¨oster, G. (2008). Distributed activity recogni-
tion with fuzzy-enabled wireless sensor networks. In
Distributed Computing in Sensor Systems, pages 296–
313. Springer.
Martin, H., Bernardos, A. M., Iglesias, J., and Casar, J. R.
(2013). Activity logging using lightweight classifica-
tion techniques in mobile devices. Personal and Ubiq-
uitous Computing, 17(4):675–695.
Miluzzo, E., Lane, N. D., Fodor, K., Peterson, R., Lu, H.,
Musolesi, M., Eisenman, S. B., Zheng, X., and Camp-
bell, A. T. (2008). Sensing meets mobile social net-
works: The design, implementation and evaluation of
the CenceMe application. In Proceedings of the 6th
ACM Conference on Embedded Network Sensor Sys-
tems, SenSys ’08, pages 337–350, New York, NY,
USA. ACM.
Ouchi, K. and Doi, M. (2012). Indoor-outdoor activ-
ity recognition by a smartphone. In Proceedings of
the 2012 ACM Conference on Ubiquitous Computing,
UbiComp ’12, pages 600–601, New York, NY, USA.
ACM.
Reddy, S., Mun, M., Burke, J., Estrin, D., Hansen, M., and
Srivastava, M. (2010). Using mobile phones to deter-
mine transportation modes. ACM Trans. Sen. Netw.,
6(2):1–27.
Ryder, J., Longstaff, B., Reddy, S., and Estrin, D. (2009).
Ambulation: A tool for monitoring mobility patterns
over time using mobile phones. In International Con-
ference on Computational Science and Engineering,
2009. CSE ’09, volume 4, pages 927–931.
Schindhelm, C. (2012). Activity recognition and step de-
tection with smartphones: Towards terminal based in-
door positioning system. In 2012 IEEE 23rd Inter-
national Symposium on Personal Indoor and Mobile
Radio Communications (PIMRC), pages 2454–2459.
Shoaib, M., Bosch, S., Incel, O. D., Scholten, H., and
Havinga, P. J. M. (2014). Fusion of smartphone mo-
tion sensors for physical activity recognition. Sensors,
14(6):10146–10176.
Shoaib, M., Scholten, H., and Havinga, P. (2013). To-
wards physical activity recognition using smartphone
sensors. In Ubiquitous Intelligence and Computing,
2013 IEEE 10th International Conference on and 10th
International Conference on Autonomic and Trusted
Computing (UIC/ATC), pages 80–87.
Siirtola, P. (2012). Recognizing human activities user-
independently on smartphones based on accelerome-
ter data. International Journal of Interactive Multime-
dia and Artificial Intelligence, 1(Special Issue on Dis-
tributedComputing and Artificial Intelligence):38–45.
Siirtola, P. and Roning, J. (2013). Ready-to-use activity
recognition for smartphones. In 2013 IEEE Sympo-
sium on Computational Intelligence and Data Mining
(CIDM), pages 59–64.
Stewart, V., Ferguson, S., Peng, J.-X., and Rafferty, K.
(2012). Practical automated activity recognition us-
ing standard smartphones. In Pervasive Computing
and Communications Workshops, IEEE International
Conference on, volume 0, pages 229–234, Los Alami-
tos, CA, USA. IEEE Computer Society.
Thiemjarus, S., Henpraserttae, A., and Marukatat, S.
(2013). A study on instance-based learning
with reduced training prototypes for device-context-
independent activity recognition on a mobile phone.
In 2013 IEEE International Conference on Body Sen-
sor Networks (BSN), pages 1–6.
Vo, Q. V., Hoang, M. T., and Choi, D. (2013). Personal-
ization in mobile activity recognition system using k-
medoids clustering algorithm. International Journal
of Distributed Sensor Networks, 2013.
Wang, Y., Lin, J., Annavaram, M., Jacobson, Q. A., Hong,
J., Krishnamachari, B., and Sadeh, N. (2009). A
framework of energy efficient mobile sensing for auto-
matic user state recognition. In Proceedings of the 7th
International Conference on Mobile Systems, Appli-
cations, and Services, MobiSys ’09, pages 179–192,
New York, NY, USA. ACM.
Yan, Z., Subbaraju, V., Chakraborty, D., Misra, A., and
Aberer, K. (2012). Energy-efcient continuous ac-
tivity recognition on mobile phones: An activity-
adaptive approach. In 2012 16th International Sympo-
sium on Wearable Computers (ISWC), pages 17–24.
Zhao, K., Du, J., Li, C., Zhang, C., Liu, H., and Xu, C.
(2013). Healthy: A diary system based on activity
recognition using smartphone. In 2013 IEEE 10th In-
ternational Conference on Mobile Ad-Hoc and Sensor
Systems (MASS), pages 290–294.
DefiningaRoadmapTowardsComparativeResearchinOnlineActivityRecognitiononMobilePhones
159