Beruflich Dokumente
Kultur Dokumente
WELCOME
1
NISHAD KV No :10 S7 IT
ABSTRACT
TranslatAR is a mobile augmented reality (AR) translation system, using a Smartphone's camera and touch screen, that requires the user to simply tap on the word of interest once in order to produce a translation, presented as an AR overlay.
CONTENTS
1) 2) 3) 4) 5) 6) 7) 8) 9)
Introduction What is Augmented Reality? Overview of the system Implementation Details Advantages Disadvantages Future Enhancements Conclusion References
4
INTRODUCTION
Written text is one of the most common methods for conveying information in our daily lives. However, when written text is encountered in a language unfamiliar to an individual, the information cannot be conveyed. Using a Smartphone with touch screen and camera as the physical device, we present a system for automatic translation of visual text that has an efficient and easy-to-use input and a natural and compelling form of presentation. The use of our system, TranslatAR, is shown in Fig. 1.
5
nishadkv shm engg college kollam
FIG.1
FIG 2
1. TEXT DETECTION
Given a point c onto which the user tapped, the system first finds the bounding box around the word, then the exact location and orientation of the text within.
1) 2)
nishadkv shm engg college kollam
10
11
12
Pixels within bounding box are considered. Lines that cross the vertical line through c at an
angle are considered. A voting scheme is assigned for the task of finding text baselines. The resulting quadrilateral region of interest is warped into a rectangle, correcting any perspective distortion and showing the text as if seen in.(fig 3.c, fig 3.d)
13
14
15
2.2 OCR(OPTICAL CHARACTER RECOGNITION) With the rectified image of the word, we rely on a
standard OCR system recognition of the letters. for extraction and
nishadkv shm engg college kollam
We
Tesseract
is an OCR engine developed by HP from 1985 to 1995 until the project was continued and released as open source by Google in 2005.
16
OCR (Cont)
Tesseract
It is open source, compiles easily on any platform, and performs all that is necessary for TranslatAR's purpose.
17
The
Levenshtein distance to the found string is computed for each dictionary word within 2 of the length of the found string.
18
The
19
20
of the word of interest in the live video stream and to present the translation in a live, AR-style overlay. Several circumstances make tracking in our application easier than it is in the general case:
1) We may assume that the text is displayed on a near-to-planar surface. 2) As the region of interest consists of text, it is automatically well-textured and contains features with high contrast, which is important for tracking.
21
22
computed by the tracker, a graphical augmentation is rendered onto the live video screen.
23
First a placeholder (please wait...) is displayed while the text is being translated.
24
25
FIG 2
26
N900, which runs on Linux based OS- Maemo. The code was developed in C++. The graphical augmentation was implemented in OpenGL ES 2. HTTP requests to and responses from Googles online translation service are handled with the curl library.
27
in a Linux platform, The application TranslatAR is easy to extend and improve by other developers compared to systems on other platforms. This application is reasonably simple case for tracking compared to other tracking applications. Processing is done on the mobile device itself, providing immediate feedback. No need for other peripheral devices for translation of text.
28
DISADVANTAGES
of control over the viewfinder's focus currently limits translatAR to relatively large fonts. The frame rate has to be significantly improved, the most obvious way is to make use of hardware acceleration in many of the image processing & rendering tasks. Real world imaging problems like illumination variance, glare, and dirt are found to cause significant problems for both text detection and the Tesseract OCR module.
Lack
29
would absorb some of the OCR errors. Text to voice translation and vice versa. Translation of text without the use of online services. Implementation of similar systems in mobile devices other than smart phones.
FUTURE ENHANCEMENTS Translation of texts in all font formats. Translation of texts at faster frame rates. Integration of spell checkers before translation
30
CONCLUSION
In this work, we presented a prototype for a real-time, mobile, visual translation system, which requires only a single tap on the word of interest and presents its result as a live AR rendering The application is capable of overlaying an automatically translated text over a region of interest which is extracted and tracked in the video stream of the camera.
nishadkv shm engg college kollam
31
REFERENCES
T. N. Dinh, J. Park, and G. Lee. Low-complexity text extraction in Korean signboards for mobile applications. In Computer and Information Technology, 2008. CIT 2008. 8th IEEE Intl. Conf. on, pages 333 337, 8-11 2008. http://www.wikipedia.org/ B. Epshtein, E. Ofek, and Y.Wexler. Detecting text in natural scenes with stroke width transform. In Proc. IEEE CVPR 2010, July 2010. http://ieeexplore.ieee.org http://ilab.cs.ucsb.edu/index.php/component/content/article/10/14
32
THANK YOU
33