one of the most commonly used clinical tests to
assess concussion, and measures standing posture
and balance related impairments. This test requires
the clinician to count balance “errors” that include
eye opening, arm movements, trunk leaning and
stepping while simultaneously protecting the patient
against balance loss. This multitasking on the part of
the clinician can increase the chances of human
errors in counting the balance errors. A means to
administer the BESS automatically may reduce the
error in scoring by individual clinicians and also
allow them to focus on patient safety during the test.
The purpose of this research is to replace the
manual administration of the BESS test with a fully
automated version by using emerging motion
capturing and image processing technologies in
combination with custom software. This would not
only increase the reliability of the test results, but
would also provide an easier way of conducting
different tests and recording the results into the
database for future comparisons. Resulting research
will not only aid in detecting post concussive
symptoms, but also help in preventing the risks of
multiple concussions by improving the reliability of
return-to play decisions.
The concept of administrating the BESS using
machine vision can be achieved using cameras
capable of measuring depth such as laser-based
time-of-flight cameras, structured light systems and
camera-based triangulation systems which may cost
~$100k USD. Alternatively, we explore the use of
emerging gaming technology such as the Kinect for
Windows which costs less than $300 USD, opening
up the use of depth cameras in a wide range of
applications (Choppin, 2013). As a research tool, the
Kinect can be controlled and accessed through
computer and driver software easily (Kinect for
Windows Programming Guide, 2013).
2 METHODS
The purpose of this research is to create a system for
inexpensively and accurately quantifying post-
concussive symptoms by administering a computer
automated version of the standard BESS test.
Whereas the standard BESS test is scored by a
highly trained human clinician, our system will use
the Microsoft Kinect, a commodity motion capture
system, to track patient movement and to score the
exam. This system will be valuable because it will
facilitate the measurement of concussion especially
in situations where a trained clinician is not readily
available such as at amateur sporting events or in
active military environments. By improving the
determination of concussion symptoms, our system
will facilitate return-to-play and return-to-duty
decisions, thereby improving clinical outcomes for
patients.
2.1 Overview
Our system is comprised of just two hardware
elements, both of which are readily available
commodity items requiring no physical alteration or
modification. The Kinect (Microsoft, Redmond,
WA, USA) is a relatively inexpensive motion
capture system originally developed for gaming
applications. A built-in software layer tracks human
body movement in real-time, expressed as x-y
coordinates for 20 key body joints. The second
hardware element is a standard Window-based
personal computer. Using an open-source software
development kit, custom software is written for the
PC that can quantify the relevant measurements of
the BESS test (trunk angle, foot lift, etc) using the
skeleton coordinates returned from the Kinect. The
Kinect is also used to detect eyeblinks. For ease of
development, software is written using Matlab
(Mathworks, Natick MA, USA), although an
eventual production version of the system would be
coded in C/C++ or Java. The system is self-
contained, portable, and can be easily administered
by a technician with no medical training.
2.2 Microsoft Kinect
Launched in November 2010, Microsoft Kinect is a
sensor suite based around the PrimeSense design,
which allows it to provide depth, RGB, infrared and
audio information to the end user of the product
(Boulos, 2011).
The sensor has an RGB (red-green-
blue) camera for color video, and an infrared emitter
and camera that measure depth (in millimeter).
Through its depth camera it is able to capture point
cloud data at 30Hz, effectively scanning a surface as
it does so. Proprietary algorithms developed by
PrimeSense and Microsoft are not only able to use
the depth cloud to recognize human users within the
field of view but also to calculate joint positions and
segment angles for the purposes of gesture
recognition and command (Choppin, 2013).
The Kinect can see a usable range from zero to
five meters in front of the sensor. The field of view
is 57
◦
horizontal and 43
◦
vertical. The motorized tilt
of the sensor allows for ±28
◦
of movement in the
vertical axis. Image data is captured at 1280x1024
but the algorithm operating within the Kinect
BIODEVICES2014-InternationalConferenceonBiomedicalElectronicsandDevices
330