• Title/Summary/Keyword: 이미지 변환 엔진

Search Result 16, Processing Time 0.017 seconds

A Text Processing Method for Devanagari Scripts in Andriod (안드로이드에서 힌디어 텍스트 처리 방법)

  • Kim, Jae-Hyeok;Maeng, Seung-Ryol
    • The Journal of the Korea Contents Association
    • /
    • v.11 no.12
    • /
    • pp.560-569
    • /
    • 2011
  • In this paper, we propose a text processing method for Hindi characters, Devanagari scripts, in the Android. The key points of the text processing are to device automata, which define the combining rules of alphabets into a set of syllables, and to implement a font rendering engine, which retrieves and displays the glyph images corresponding to specific characters. In general, an automaton depends on the type and the number of characters. For the soft-keyboard, we designed the automata with 14 consonants and 34 vowels based on Unicode. Finally, a combined syllable is converted into a glyph index using the mapping table, used as a handle to load its glyph image. According to the multi-lingual framework of Freetype font engine, Dvanagari scripts can be supported in the system level by appending the implementation of our method to the font engine as the Hindi module. The proposed method is verified through a simple message system.

Study on Extracting Filming Location Information in Movies Using OCR for Developing Customized Travel Content (맞춤형 여행 콘텐츠 개발을 위한 OCR 기법을 활용한 영화 속 촬영지 정보 추출 방안 제시)

  • Park, Eunbi;Shin, Yubin;Kang, Juyoung
    • The Journal of Bigdata
    • /
    • v.5 no.1
    • /
    • pp.29-39
    • /
    • 2020
  • Purpose The atmosphere of respect for individual tastes that have spread throughout society has changed the consumption trend. As a result, the travel industry is also seeing customized travel as a new trend that reflects consumers' personal tastes. In particular, there is a growing interest in 'film-induced tourism', one of the areas of travel industry. We hope to satisfy the individual's motivation for traveling while watching movies with customized travel proposals, which we expect to be a catalyst for the continued development of the 'film-induced tourism industry'. Design/methodology/approach In this study, we implemented a methodology through 'OCR' of extracting and suggesting film location information that viewers want to visit. First, we extract a scene from a movie selected by a user by using 'OpenCV', a real-time image processing library. In addition, we detected the location of characters in the scene image by using 'EAST model', a deep learning-based text area detection model. The detected images are preprocessed by using 'OpenCV built-in function' to increase recognition accuracy. Finally, after converting characters in images into recognizable text using 'Tesseract', an optical character recognition engine, the 'Google Map API' returns actual location information. Significance This research is significant in that it provides personalized tourism content using fourth industrial technology, in addition to existing film tourism. This could be used in the development of film-induced tourism packages with travel agencies in the future. It also implies the possibility of being used for inflow from abroad as well as to abroad.

Design and Implementation of Real-time High Performance Face Detection Engine (고성능 실시간 얼굴 검출 엔진의 설계 및 구현)

  • Han, Dong-Il;Cho, Hyun-Jong;Choi, Jong-Ho;Cho, Jae-Il
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.47 no.2
    • /
    • pp.33-44
    • /
    • 2010
  • This paper propose the structure of real-time face detection hardware architecture for robot vision processing applications. The proposed architecture is robust against illumination changes and operates at no less than 60 frames per second. It uses Modified Census Transform to obtain face characteristics robust against illumination changes. And the AdaBoost algorithm is adopted to learn and generate the characteristics of the face data, and finally detected the face using this data. This paper describes the face detection hardware structure composed of Memory Interface, Image Scaler, MCT Generator, Candidate Detector, Confidence Comparator, Position Resizer, Data Grouper, and Detected Result Display, and verification Result of Hardware Implementation with using Virtex5 LX330 FPGA of Xilinx. Verification result with using the images from a camera showed that maximum 32 faces per one frame can be detected at the speed of maximum 149 frame per second.

Trends in Deep Learning-based Medical Optical Character Recognition (딥러닝 기반의 의료 OCR 기술 동향)

  • Sungyeon Yoon;Arin Choi;Chaewon Kim;Sumin Oh;Seoyoung Sohn;Jiyeon Kim;Hyunhee Lee;Myeongeun Han;Minseo Park
    • The Journal of the Convergence on Culture Technology
    • /
    • v.10 no.2
    • /
    • pp.453-458
    • /
    • 2024
  • Optical Character Recognition is the technology that recognizes text in images and converts them into digital format. Deep learning-based OCR is being used in many industries with large quantities of recorded data due to its high recognition performance. To improve medical services, deep learning-based OCR was actively introduced by the medical industry. In this paper, we discussed trends in OCR engines and medical OCR and provided a roadmap for development of medical OCR. By using natural language processing on detected text data, current medical OCR has improved its recognition performance. However, there are limits to the recognition performance, especially for non-standard handwriting and modified text. To develop advanced medical OCR, databaseization of medical data, image pre-processing, and natural language processing are necessary.

Characteristics of Meteorological and Marine Environments for the Red Tide Occurrence of Mid-South Sea in Korea (한국 남해중부해역의 적조발생에 관한 기상 및 해양환경 특성)

  • 윤홍주;김승철;박일흠
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.7 no.4
    • /
    • pp.845-852
    • /
    • 2003
  • This study deals with the relationship between the red tide occurrence and the meteorological and marine factors, the prediction of areas where the red tide is likely to occur based on the information, and the satellite monitoring for the red tide in mid-South Sea of Korea. From 1990 to 2001, the red tide was observed every year and the number of occurrences increased as well. The red tide mostly occurred in July, August, and September. The most important meteorological factor governing the mechanisms of the increase in the number of red tide occurrences is found to be a heavy precipitation. It was found that the favorable marine environmental conditions for the red tide formation are some of marine factors such as the warm water temperature, the low salinity, the high suspended solid, the low phosphorus, and the low nitrogen. The necessary conditions for the red tide occurrence are found to be the heavy precipitation (23.4-54.5 mm) for 2∼4 days, the warm temperature $(24.6∼25.9^{\circ}C)$, proper sunshine (2∼10.3 h), and light winds (2∼4.6 m/s & SW) for the day in red tide occurrence. It was possible to monitor the spatial distributions and concentration of the red tide using the satellite images. It was found that the likely areas for red tide occurrence in August 2000 were Yosu - Dolsan coast, Gamak bay, Namhae coast, Marado coast, Goheung coast, and Deukryang bay.

A Study on CSS3 Stylesheet Extension and Emulator for Representation of Stereo Web Content in 3DTV (3DTV에서 스테레오 웹 콘텐츠 표현을 위한 CSS3 사양 확장 및 에뮬레이터 연구)

  • Lee, Hee-Jin;Yim, Hyun-Jeong;Lim, Soon-Bum
    • Journal of the Korea Computer Graphics Society
    • /
    • v.19 no.4
    • /
    • pp.1-11
    • /
    • 2013
  • In this paper, we represented the HTML5 webpage in 3D space for the 3D stereoscopic display by using CSS3 stylesheet; browser-based declarative contents, and proposed CSS Stereo 3DTV Profile. First, we suggested various webpage separation methods for reconstructing webpages in the 3D space effectively. Next, we suggested 3D view volume setting method by using extended CSS3 modules. And then, pre-processor converts sample contents which is written extended CSS stylesheet into the present CSS stylesheet for displaying in the webkit based browser. For the resulting stereoscopic images, we developed a rendering engine emulator which is implemented in JavaScript for simple display in the web browser, which produced dual images from virtual left and right-eyed cameras. And we have checked the sample contents displayed on the 3DTV.