
 
virtual objects by holding account of the objects’ 
properties and theirs positions (left, right, up, down) 
is allowed by this system. 
Secondly, we could quote the ARTHUR project 
Aish, F., Broll, W., Stoerring, M., Fatah, A., 
Mottram, C., 2004. Its goal has been for a planning 
urban on a table. To see the mixed environment, 
each user had worn a Head-Mounted Display 
(HMD) with a high resolution. Interaction tools are 
placeholder, and, in order to change the user 
interface gesture recognition has been used. The 
database of virtual object was duplicated in each 
node and synchronization was made for each state’s 
change.  
The next project is MARE Raphaël G., Jean-
Dominique G., 2002. It has designed an AR system 
on tabletop that has allowed several users 
collaborating. This system has used a stereo optical 
see-through HMD as visualization tools. It is 
attached with magnetic or optical tracker, its 
workspace is divided in two areas: a public one is 
reserved for collaborative interaction, and, a private 
one is dedicated to put the real tools (mouse, 
keyboard, PDA …) and virtual tools used by the 
users. The virtual menu, which is used by users for 
choosing a virtual object, has been a part of these 
tools. The system can support several types of 
interaction like navigation, selection and 
manipulation. It is used to control access manager 
and personal viewer of user.  
Then, the EMMIE Andreas B., Tobias H., Steven 
F., Blair M., Clifford B., 1999 framework introduces 
a hybrid user interface for AR systems that enables 
information management using a wide range of 
hardware devices. EMMIE’s environment manager 
component addresses the needs of Ubicomp by 
providing techniques such as mixed reality 
interaction and privacy management to organize 
virtual information on several displays shared by 
multiple users.  
Mooser, J., Lu, W., Suya, Y., Neumann, U., 2007 
Proposed an AR user interface framework 
specifically designed to expose disparate data 
sources through a single application server. It uses a 
multi-tier architecture to separate back-end data 
retrieval from front-end graphical presentation and 
UI event handling. For target recognition and pose 
estimation, the system uses TriCode fiducials. The 
recognition and pose estimation can run a mobile 
client in real time, with the identity of a newly 
detected TriCode then passed to the server as a 
single numerical value. Data can exist in many 
forms, such as relational databases, sensor readings, 
or locations of other users. Once the user interface 
structures have been built, the application server 
serializes them into a simple XML stream, which is 
sent back to the mobile client. Having received the 
XML stream, the client has been built its own local 
version of the user interface structure. The user has 
been shown the mixed content through a Sony Vaio 
UX Micro PC display. 
The acclaim MARS (Mobile Augmented Reality 
Systems) project Höllerer, T., Feiner, S., Terauchi, 
T., Rashid, G., and Hallaway, D., 1999 presented in 
1999 by Columbia University was one of the first 
truly mobile augmented reality setups which allowed 
the user to freely walk around while having all 
necessary equipment mounted onto his back.  
The AR Phone project Mark, A., David, J., 
Daniel, C., Adam, H., 2003 was an AR system 
where the tracking task was dedicated to an AR 
server. Cell phone has been used as thin client and 
an access point was relayed the exchange between 
the client and the AR server. The tracking task is 
dedicated to ARToolKit Billinghurst, M., Kato, H., 
Weghorst, S., Furness, T. A., 1999 library running 
on the AR server. 
 More recent works have used lightweight 
wearable devices such as PDA and Smartphone as 
display device.  
In Daniel, W., Thomas, P., Florian, L., Dieter, S., 
2005, Daniel Wagner presented an AR system 
collaborative running on a PDA. The users have 
been able to interact with the virtual objects using a 
stylus. The tracking task is dedicated to Opentracker 
Reitmayr, G., Schmalstieg, D., 2001 library and is 
run on the PDA itself. The platform is based on 
studierstube Schmalstieg, D., Fuhrmann, A., Hesina, 
G., Szalav´ari, Z., Encarna, c., L.M., Gervautz, M., 
Purgathofer, W., 2002 framework optimized for a 
mobile light device. The system used ACE (the 
Adaptive Communication Environment) for network 
communication abstraction.  
AR tennis Henrysson, A., Billinghurst, M., 
Ollila, M., 2006 proposed a system AR on a table 
running on a Smartphone. This last one used as 
tangible interaction and visual tools. It has used the 
ARToolkit’s version revised by Henrysson as 
tracking tools and OpenGL ES (Open Graphics 
Library for Embedded System) as render engine. It 
has allowed two users to collaborate. The system 
also provided audio feedback when the ball touches 
the racket.  
The sections that follow expand our system 
overview.  
 
A NEW ARCHITECTURE FOR A MULTIPLATFORM AUGMENTED REALITY SYSTEM
337