• Title/Summary/Keyword: 멀티 모달

Search Result 264, Processing Time 0.017 seconds

Efficient Uncertainty Analysis of TOPMODEL Using Particle Swarm Optimization (입자군집최적화 알고리듬을 이용한 효율적인 TOPMODEL의 불확실도 분석)

  • Cho, Huidae;Kim, Dongkyun;Lee, Kanghee
    • Journal of Korea Water Resources Association
    • /
    • v.47 no.3
    • /
    • pp.285-295
    • /
    • 2014
  • We applied the ISPSO-GLUE method, which integrates the Isolated-Speciation-based Particle Swarm Optimization (ISPSO) with the Generalized Likelihood Uncertainty Estimation (GLUE) method, to the uncertainty analysis of the Topography Model (TOPMODEL) and compared its performance with that of the GLUE method. When we performed the same number of model runs for the both methods, we were able to identify the point where the performance of ISPSO-GLUE exceeded that of GLUE, after which ISPSOGLUE kept improving its performance steadily while GLUE did not. When we compared the 95% uncertainty bounds of the two methods, their general shapes and trends were very similar, but those of ISPSO-GLUE enclosed about 5.4 times more observed values than those of GLUE did. What it means is that ISPSOGLUE requires much less number of parameter samples to generate better performing uncertainty bounds. When compared to ISPSO-GLUE, GLUE overestimated uncertainty in the recession limb following the maximum peak streamflow. For this recession period, GLUE requires to find more behavioral models to reduce the uncertainty. ISPSO-GLUE can be a promising alternative to GLUE because the uncertainty bounds of the method were quantitatively superior to those of GLUE and, especially, computationally expensive hydrologic models are expected to greatly take advantage of the feature.

Multi-modal Image Processing for Improving Recognition Accuracy of Text Data in Images (이미지 내의 텍스트 데이터 인식 정확도 향상을 위한 멀티 모달 이미지 처리 프로세스)

  • Park, Jungeun;Joo, Gyeongdon;Kim, Chulyun
    • Database Research
    • /
    • v.34 no.3
    • /
    • pp.148-158
    • /
    • 2018
  • The optical character recognition (OCR) is a technique to extract and recognize texts from images. It is an important preprocessing step in data analysis since most actual text information is embedded in images. Many OCR engines have high recognition accuracy for images where texts are clearly separable from background, such as white background and black lettering. However, they have low recognition accuracy for images where texts are not easily separable from complex background. To improve this low accuracy problem with complex images, it is necessary to transform the input image to make texts more noticeable. In this paper, we propose a method to segment an input image into text lines to enable OCR engines to recognize each line more efficiently, and to determine the final output by comparing the recognition rates of CLAHE module and Two-step module which distinguish texts from background regions based on image processing techniques. Through thorough experiments comparing with well-known OCR engines, Tesseract and Abbyy, we show that our proposed method have the best recognition accuracy with complex background images.

Artificial Intelligence for Assistance of Facial Expression Practice Using Emotion Classification (감정 분류를 이용한 표정 연습 보조 인공지능)

  • Dong-Kyu, Kim;So Hwa, Lee;Jae Hwan, Bong
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.17 no.6
    • /
    • pp.1137-1144
    • /
    • 2022
  • In this study, an artificial intelligence(AI) was developed to help with facial expression practice in order to express emotions. The developed AI used multimodal inputs consisting of sentences and facial images for deep neural networks (DNNs). The DNNs calculated similarities between the emotions predicted by the sentences and the emotions predicted by facial images. The user practiced facial expressions based on the situation given by sentences, and the AI provided the user with numerical feedback based on the similarity between the emotion predicted by sentence and the emotion predicted by facial expression. ResNet34 structure was trained on FER2013 public data to predict emotions from facial images. To predict emotions in sentences, KoBERT model was trained in transfer learning manner using the conversational speech dataset for emotion classification opened to the public by AIHub. The DNN that predicts emotions from the facial images demonstrated 65% accuracy, which is comparable to human emotional classification ability. The DNN that predicts emotions from the sentences achieved 90% accuracy. The performance of the developed AI was evaluated through experiments with changing facial expressions in which an ordinary person was participated.

A Study on Biometric Model for Information Security (정보보안을 위한 생체 인식 모델에 관한 연구)

  • Jun-Yeong Kim;Se-Hoon Jung;Chun-Bo Sim
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.19 no.1
    • /
    • pp.317-326
    • /
    • 2024
  • Biometric recognition is a technology that determines whether a person is identified by extracting information on a person's biometric and behavioral characteristics with a specific device. Cyber threats such as forgery, duplication, and hacking of biometric characteristics are increasing in the field of biometrics. In response, the security system is strengthened and complex, and it is becoming difficult for individuals to use. To this end, multiple biometric models are being studied. Existing studies have suggested feature fusion methods, but comparisons between feature fusion methods are insufficient. Therefore, in this paper, we compared and evaluated the fusion method of multiple biometric models using fingerprint, face, and iris images. VGG-16, ResNet-50, EfficientNet-B1, EfficientNet-B4, EfficientNet-B7, and Inception-v3 were used for feature extraction, and the fusion methods of 'Sensor-Level', 'Feature-Level', 'Score-Level', and 'Rank-Level' were compared and evaluated for feature fusion. As a result of the comparative evaluation, the EfficientNet-B7 model showed 98.51% accuracy and high stability in the 'Feature-Level' fusion method. However, because the EfficietnNet-B7 model is large in size, model lightweight studies are needed for biocharacteristic fusion.