
one of the most commonly used clinical tests to 
assess concussion, and measures standing posture 
and balance related impairments. This test requires 
the clinician to count balance “errors” that include 
eye opening, arm movements, trunk leaning and 
stepping while simultaneously protecting the patient 
against balance loss. This multitasking on the part of 
the clinician can increase the chances of human 
errors in counting the balance errors. A means to 
administer the BESS automatically may reduce the 
error in scoring by individual clinicians and also 
allow them to focus on patient safety during the test. 
The purpose of this research is to replace the 
manual administration of the BESS test with a fully 
automated version by using emerging motion 
capturing and image processing technologies in 
combination with custom software. This would not 
only increase the reliability of the test results, but 
would also provide an easier way of conducting 
different tests and recording the results into the 
database for future comparisons. Resulting research 
will not only aid in detecting post concussive 
symptoms, but also help in preventing the risks of 
multiple concussions by improving the reliability of 
return-to play decisions.  
The concept of administrating the BESS using 
machine vision can be achieved using cameras 
capable of measuring depth such as laser-based 
time-of-flight cameras, structured light systems and 
camera-based triangulation systems which may cost 
~$100k USD. Alternatively, we explore the use of 
emerging gaming technology such as the Kinect for 
Windows which costs less than $300 USD, opening 
up the use of depth cameras in a wide range of 
applications (Choppin, 2013). As a research tool, the 
Kinect can be controlled and accessed through 
computer and driver software easily (Kinect for 
Windows Programming Guide, 2013). 
2 METHODS 
The purpose of this research is to create a system for 
inexpensively and accurately quantifying post-
concussive symptoms by administering a computer 
automated version of the standard BESS test. 
Whereas the standard BESS test is scored by a 
highly trained human clinician, our system will use 
the Microsoft Kinect, a commodity motion capture 
system, to track patient movement and to score the 
exam. This system will be valuable because it will 
facilitate the measurement of concussion especially 
in situations where a trained clinician is not readily 
available such as at amateur sporting events or in 
active military environments. By improving the 
determination of concussion symptoms, our system 
will facilitate return-to-play and return-to-duty 
decisions, thereby improving clinical outcomes for 
patients. 
2.1 Overview 
Our system is comprised of just two hardware 
elements, both of which are readily available 
commodity items requiring no physical alteration or 
modification. The Kinect (Microsoft, Redmond, 
WA, USA) is a relatively inexpensive motion 
capture system originally developed for gaming 
applications. A built-in software layer tracks human 
body movement in real-time, expressed as x-y 
coordinates for 20 key body joints. The second 
hardware element is a standard Window-based 
personal computer. Using an open-source software 
development kit, custom software is written for the 
PC that can quantify the relevant measurements of 
the BESS test (trunk angle, foot lift, etc) using the 
skeleton coordinates returned from the Kinect. The 
Kinect is also used to detect eyeblinks. For ease of 
development, software is written using Matlab 
(Mathworks, Natick MA, USA), although an 
eventual production version of the system would be 
coded in C/C++ or Java. The system is self-
contained, portable, and can be easily administered 
by a technician with no medical training. 
2.2 Microsoft Kinect 
Launched in November 2010, Microsoft Kinect is a 
sensor suite based around the PrimeSense design, 
which allows it to provide depth, RGB, infrared and 
audio information to the end user of the product 
(Boulos, 2011).
 The sensor has an RGB (red-green-
blue) camera for color video, and an infrared emitter 
and camera that measure depth (in millimeter). 
Through its depth camera it is able to capture point 
cloud data at 30Hz, effectively scanning a surface as 
it does so. Proprietary algorithms developed by 
PrimeSense and Microsoft are not only able to use 
the depth cloud to recognize human users within the 
field of view but also to calculate joint positions and 
segment angles for the purposes of gesture 
recognition and command (Choppin, 2013). 
The Kinect can see a usable range from zero to 
five meters in front of the sensor. The field of view 
is 57
◦
 horizontal and 43
◦
 vertical. The motorized tilt 
of the sensor allows for ±28
◦ 
of movement in the 
vertical axis. Image data is captured at 1280x1024 
but the algorithm operating within the Kinect 
BIODEVICES2014-InternationalConferenceonBiomedicalElectronicsandDevices
330