III/IV B.TECH (E.C.E), III/IV B.TECH (E.C.E), AITAM, Tekkali. AITAM, Tekkali. sateesh.4a1@gmail.com ngrreddy@gmail.com Abstract: everywhere. This resources at any paper describes location and at Augmented Introduction our work to build any time. AR is Reality (AR), a mobile and related often used as an enhancing a Augmented work : user interface user’s perception Reality system technique in of the real world Augmented that supports true wearable with computer Reality (AR), stereoscopic 3D computing generated annotating the graphics, a pen because it entities, and real world with and pad inter- provides an mobile computer face and direct information computing, generated interaction with space which is allowing users to entities, is a virtual objects. continuously and access and powerful user The system is transparently manipulate in- interface assembled from accessible . formation paradigm off-the-shelf Information can anytime and allowing users to hardware be accessed independent of interact with components and hands-free, and location, are two serves as a basic the user’s view of emerging user test bed for user the real-world is interface interface computers in a not interrupted, a technologies that experiments natural way. requirement for show great related to Mobilizing such continuous use. promise. The computer an interface by combination of If these supported deploying both into a single technologies are collaborative wearable system makes the combined with work in computers is a power of position tracking, Augmented logical extension computer location aware Reality. It also as the body of enhanced applications are describes some related research interaction and possible. The applications we shows. communication computer are developing in in the real world Wearable transparently the area of accessible computing allows changes its location based anytime and the user to access behavior based computing. on the for building our GeForce2Go environment own. On one video chip. The without the The mobile hand this allows device has a user’s AR setup : us to quickly 1GHZ processor intervention. An upgrade old and runs under impressive While the devices or add Windows demonstrator for computational new ones and to mobile location power for change the aware AR using stereoscopic configuration 2000. We also both a head- rendering and easily. On the added a wireless mounted and a computer vision other hand we do LAN network hand-held display is becoming not obtain the adapter to the is Columbia’s available in smallest and note-book to Touring Machine mobile computer lightest system enable [3] which was systems, the size possible. communication used to create a and weight of with our campus such systems is Hardware : stationary setup information still not optimal. or a future Nevertheless, our The most system and second mobile setup is solely powerful portable situated unit. It is carried build from off- graphics solution documentaries by the user in a the-shelf available [4]. backpack. hardware currently is a PC
components to notebook As an output
avoid the effort equipped with a device, we use an and time required NVidia i-glasses see- through provide more with virtual stationary work- stereoscopic accurate tracking objects or with space, our mobile color HMD. The on the pad itself. user interface setup with body- display is fixed to Figure 1 gives an elements stabilized display a helmet worn by overview of the registered with allows to arrange the user. setup. and displayed on 3D information Moreover, an the pad. in a wearable InterSense workspace that InterTrax2 User interface travels along orientation sensor with a user. management and a web Applications stay software : camera for where they are fiducial tracking As our software put relative to the of interaction platform we use user, and can props are Studierstube 2.1 easily be mounted on the [5], a user accessed helmet. interface anytime, aided by management proprioception system for AR and spatial The main user Applications are based on but not memory. Figure interface is a pen implemented as limited to 2 shows a simple and pad setup runtime loadable stereoscopic 3D painting using a Wacom objects executing graphics. It application. graphics tablet in designated provides a multi- volumetric and its pen. Both user, multi- containers, a kind devices are application of 3D window optically tracked environment, and equivalent. While by the camera supports a variety the original using markers. of display stationary The 2D position devices including Studierstube of the pen stereoscopic environment (provided by the HMDs. It also allowed a user to Wacom tablet) is provides the arrange multiple incorporated into means of 6DOF application in a the processing to interaction, either user’s locale will passes through a data flow Figure 2. A user be unaffected, series of steps. It network of the interacting with but the second is generated by transformations. the paint user will be able tracking The framework's application . to see movement hardware, read design is based The view of the of the graphical by device drivers, on XML, taking user. objects contained transformed to fit full advantage of in the first user’s the requirements this new Our user locale. For of the application technology by interface effective and send over allowing the use management collaboration, it network of standard XML system is also will in most cases connections to tools for capable of be necessary to other hosts. development, managing add a third These tasks are configuration and multiple locales, stationary locale handled by a documentation. which can that contains library called OpenTracker contain any graphical OpenTracker [6], uses a vision number of applications that an open software tracking library graphical objects. both users should architecture for called ARToolkit Locales are work with. the different [7] to implement important for tasks involved in the tracking of multi-user or tracking input the fiducial multi-display devices and Tracking : markers on the operation. For processing interaction props. example, each multimodal input Mobile AR It analyses the mobile user will data. requires video images require a separate significantly delivered by the wear-able The main more complex web camera workspace that concept behind tracking than a mounted to the defines a distinct OpenTracker is traditional VR helmet and locale to break up the application. In a establishes the (coordinate whole data typical VR or AR position of the system). As one manipulation into application pen and pad user moves these individual tracking data about, a second steps and build a relative to the Location when the user users head. based AR looks at them and again displays applications : the correct Location Building on the location of the tracking : mobile platform book in the described above library shelves. A similar we are currently technique is used A bookshelf was developing a to track the user’s fit out with number of position within fiducial markers prototype the environment. used for tracking. location based Our laboratory Then the Augmented and neighboring bookshelf’s Reality rooms are rigged position can be applications. with larger computed by the These markers along tracking library. applications are the walls. The Dedicated books based on the locations of these were rigged as location tracking markers are well with these described in the measured and markers, so that last section. incorporated in a the system model of the A simple location recognizes such a building. based application book when the Together with the is the AR library. user is looking at tracking It performs two it. In the information basic tasks : prototype delivered by the Firstly, it shows a application, the fiducial tracking user the location markers are the system of a requested attached to the Figure 3. The computes the book in the vast wall instead of a correct location users position bookshelves of a real shelf. Figure of a detected within these library. And 3 shows both book is rooms from the secondly, it modes. displayed. A detected markers. recognizes books selected book is shown in the of the environment on within this real shelf. environment as the tablet. The environment. The well as a means user’s location way finding to track the user’s and current room application will Another typical location within are highlighted. be ex-tended to scenario for the environment. She can select a encompass a part mobile AR As described destination by of our building to systems is a way above we clicking into the allow the user to finding prepared the room she wants roam in a larger application. The environment to to go to. Then the environment. The aim is to guide a allow the system system computes integration of user along a path to compute this. the shortest path both applications to a selected For each room a to this room and is straightforward destination. This set of markers highlights the because of the is accomplished was set up and rooms she needs multi application using two means: their locations to cross. features of the a world in measured. The Additionally the Studierstube miniature model tracking can now doors that she system. This will of the establish the needs to take are allow the user to environment with user’s location augmented in the find her way to the users location and the direction user’s view to the library and and she is looking in. guide her along then use the pathhighlighted Thus the system the path to the library and augmenting can continuously destination. application in the user’s view display place. with navigation navigation guides such as information Future work : Conclusion: arrows, high- registered to the This paper lighted doors and The prototype real world. describes our lines along the applications are work to develop desired path. In the application not finished yet. a mobile AR itself the user is We plan to platform that Such an presented with a augment a real allows location- application miniature model library and test based computing. requires a model of the the application While most related work Azuma focuses on R.: A providing Survey of information as Augment text or 2D ed overlays, we Reality. concentrate on pp. 355- 3D information 385, that the user can August interact with. 1997. First we describe Starner the mobile setup T., S. itself consisting Mann, B. of the hardware Rhodes, J. used and the Levine, J. software system Healey, developed. Then D. Kirsch, we describe two R. Picard, prototype Presence, applications we Vol. 6, are currently No. 4, pp. developing to 386-398, demonstrate the August abilities of the 1997. platform.