• Title/Summary/Keyword: learning through the image

Search Result 931, Processing Time 0.022 seconds

Implementation of Face Recognition Pipeline Model using Caffe (Caffe를 이용한 얼굴 인식 파이프라인 모델 구현)

  • Park, Jin-Hwan;Kim, Chang-Bok
    • Journal of Advanced Navigation Technology
    • /
    • v.24 no.5
    • /
    • pp.430-437
    • /
    • 2020
  • The proposed model implements a model that improves the face prediction rate and recognition rate through learning with an artificial neural network using face detection, landmark and face recognition algorithms. After landmarking in the face images of a specific person, the proposed model use the previously learned Caffe model to extract face detection and embedding vector 128D. The learning is learned by building machine learning algorithms such as support vector machine (SVM) and deep neural network (DNN). Face recognition is tested with a face image different from the learned figure using the learned model. As a result of the experiment, the result of learning with DNN rather than SVM showed better prediction rate and recognition rate. However, when the hidden layer of DNN is increased, the prediction rate increases but the recognition rate decreases. This is judged as overfitting caused by a small number of objects to be recognized. As a result of learning by adding a clear face image to the proposed model, it is confirmed that the result of high prediction rate and recognition rate can be obtained. This research will be able to obtain better recognition and prediction rates through effective deep learning establishment by utilizing more face image data.

Semantic Indoor Image Segmentation using Spatial Class Simplification (공간 클래스 단순화를 이용한 의미론적 실내 영상 분할)

  • Kim, Jung-hwan;Choi, Hyung-il
    • Journal of Internet Computing and Services
    • /
    • v.20 no.3
    • /
    • pp.33-41
    • /
    • 2019
  • In this paper, we propose a method to learn the redesigned class with background and object for semantic segmentation of indoor scene image. Semantic image segmentation is a technique that divides meaningful parts of an image, such as walls and beds, into pixels. Previous work of semantic image segmentation has proposed methods of learning various object classes of images through neural networks, and it has been pointed out that there is insufficient accuracy compared to long learning time. However, in the problem of separating objects and backgrounds, there is no need to learn various object classes. So we concentrate on separating objects and backgrounds, and propose method to learn after class simplification. The accuracy of the proposed learning method is about 5 ~ 12% higher than the existing methods. In addition, the learning time is reduced by about 14 ~ 60 minutes when the class is configured differently In the same environment, and it shows that it is possible to efficiently learn about the problem of separating the object and the background.

Character Recognition Algorithm in Low-Quality Legacy Contents Based on Alternative End-to-End Learning (대안적 통째학습 기반 저품질 레거시 콘텐츠에서의 문자 인식 알고리즘)

  • Lee, Sung-Jin;Yun, Jun-Seok;Park, Seon-hoo;Yoo, Seok Bong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.11
    • /
    • pp.1486-1494
    • /
    • 2021
  • Character recognition is a technology required in various platforms, such as smart parking and text to speech, and many studies are being conducted to improve its performance through new attempts. However, with low-quality image used for character recognition, a difference in resolution of the training image and test image for character recognition occurs, resulting in poor accuracy. To solve this problem, this paper designed an end-to-end learning neural network that combines image super-resolution and character recognition so that the character recognition model performance is robust against various quality data, and implemented an alternative whole learning algorithm to learn the whole neural network. An alternative end-to-end learning and recognition performance test was conducted using the license plate image among various text images, and the effectiveness of the proposed algorithm was verified with the performance test.

Space Design for Enhancing Learning Ability with Children's Character Type - Through Analyzed Enneagram Tool - (어린이 성격유형별 학습능력 향상을 위한 공간디자인 구축 방안 - 에니어그램 성격 특성 분석을 통하여 -)

  • Kim, Kook-Sun
    • Journal of the Korea Furniture Society
    • /
    • v.24 no.1
    • /
    • pp.42-50
    • /
    • 2013
  • The objective of this study is to explore basic type of character of humans and to suggest a design method of establishing a spatial construction environment for developing effective learning ability based on such type of character. As a range of research, spatial formative language was deduced and space design strategy for the children was suggested through an analysis of spatial requirements by exploring connectivity depending on features of 9 types of character through Enneagram. As a method of research, a process of suggesting a concrete method after defining an element of spatial construction and deducing a formative language for developing and strengthening effective learning ability for each type of character. As a result of research, the methods of children space design strategy for enhancing learning ability for leadership in a future specific fields were suggested through 9 different type of character with image of case study.

  • PDF

An Intelligent Fire Learning and Detection System Using Convolutional Neural Networks (컨볼루션 신경망을 이용한 지능형 화재 학습 및 탐지 시스템)

  • Cheoi, Kyungjoo;Jeon, Minseong
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.11
    • /
    • pp.607-614
    • /
    • 2016
  • In this paper, we propose an intelligent fire learning and detection system using convolutional neural networks (CNN). Through the convolutional layer of the CNN, various features of flame and smoke images are automatically extracted, and these extracted features are learned to classify them into flame or smoke or no fire. In order to detect fire in the image, candidate fire regions are first extracted from the image and extracted candidate regions are passed through CNN. Experimental results on various image shows that our system has better performances over previous work.

A Development of Façade Dataset Construction Technology Using Deep Learning-based Automatic Image Labeling (딥러닝 기반 이미지 자동 레이블링을 활용한 건축물 파사드 데이터세트 구축 기술 개발)

  • Gu, Hyeong-Mo;Seo, Ji-Hyo;Choo, Seung-Yeon
    • Journal of the Architectural Institute of Korea Planning & Design
    • /
    • v.35 no.12
    • /
    • pp.43-53
    • /
    • 2019
  • The construction industry has made great strides in the past decades by utilizing computer programs including CAD. However, compared to other manufacturing sectors, labor productivity is low due to the high proportion of workers' knowledge-based task in addition to simple repetitive task. Therefore, the knowledge-based task efficiency of workers should be improved by recognizing the visual information of computers. A computer needs a lot of training data, such as the ImageNet project, to recognize visual information. This study, aim at proposing building facade datasets that is efficiently constructed by quickly collecting building facade data through portal site road view and automatically labeling using deep learning as part of construction of image dataset for visual recognition construction by the computer. As a method proposed in this study, we constructed a dataset for a part of Dongseong-ro, Daegu Metropolitan City and analyzed the utility and reliability of the dataset. Through this, it was confirmed that the computer could extract the significant facade information of the portal site road view by recognizing the visual information of the building facade image. Additionally, In contribution to verifying the feasibility of building construction image datasets. this study suggests the possibility of securing quantitative and qualitative facade design knowledge by extracting the facade design knowledge from any facade all over the world.

Estimation of two-dimensional position of soybean crop for developing weeding robot (제초로봇 개발을 위한 2차원 콩 작물 위치 자동검출)

  • SooHyun Cho;ChungYeol Lee;HeeJong Jeong;SeungWoo Kang;DaeHyun Lee
    • Journal of Drive and Control
    • /
    • v.20 no.2
    • /
    • pp.15-23
    • /
    • 2023
  • In this study, two-dimensional location of crops for auto weeding was detected using deep learning. To construct a dataset for soybean detection, an image-capturing system was developed using a mono camera and single-board computer and the system was mounted on a weeding robot to collect soybean images. A dataset was constructed by extracting RoI (region of interest) from the raw image and each sample was labeled with soybean and the background for classification learning. The deep learning model consisted of four convolutional layers and was trained with a weakly supervised learning method that can provide object localization only using image-level labeling. Localization of the soybean area can be visualized via CAM and the two-dimensional position of the soybean was estimated by clustering the pixels associated with the soybean area and transforming the pixel coordinates to world coordinates. The actual position, which is determined manually as pixel coordinates in the image was evaluated and performances were 6.6(X-axis), 5.1(Y-axis) and 1.2(X-axis), 2.2(Y-axis) for MSE and RMSE about world coordinates, respectively. From the results, we confirmed that the center position of the soybean area derived through deep learning was sufficient for use in automatic weeding systems.

Effect of Input Data Video Interval and Input Data Image Similarity on Learning Accuracy in 3D-CNN

  • Kim, Heeil;Chung, Yeongjee
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.13 no.2
    • /
    • pp.208-217
    • /
    • 2021
  • 3D-CNN is one of the deep learning techniques for learning time series data. However, these three-dimensional learning can generate many parameters, requiring high performance or having a significant impact on learning speed. We will use these 3D-CNNs to learn hand gesture and find the parameters that showed the highest accuracy, and then analyze how the accuracy of 3D-CNN varies through input data changes without any structural changes in 3D-CNN. First, choose the interval of the input data. This adjusts the ratio of the stop interval to the gesture interval. Secondly, the corresponding interframe mean value is obtained by measuring and normalizing the similarity of images through interclass 2D cross correlation analysis. This experiment demonstrates that changes in input data affect learning accuracy without structural changes in 3D-CNN. In this paper, we proposed two methods for changing input data. Experimental results show that input data can affect the accuracy of the model.

3D Medical Image Data Augmentation for CT Image Segmentation (CT 이미지 세그멘테이션을 위한 3D 의료 영상 데이터 증강 기법)

  • Seonghyeon Ko;Huigyu Yang;Moonseong Kim;Hyunseung Choo
    • Journal of Internet Computing and Services
    • /
    • v.24 no.4
    • /
    • pp.85-92
    • /
    • 2023
  • Deep learning applications are increasingly being leveraged for disease detection tasks in medical imaging modalities such as X-ray, Computed Tomography (CT), and Magnetic Resonance Imaging (MRI). Most data-centric deep learning challenges necessitate the use of supervised learning methodologies to attain high accuracy and to facilitate performance evaluation through comparison with the ground truth. Supervised learning mandates a substantial amount of image and label sets, however, procuring an adequate volume of medical imaging data for training is a formidable task. Various data augmentation strategies can mitigate the underfitting issue inherent in supervised learning-based models that are trained on limited medical image and label sets. This research investigates the enhancement of a deep learning-based rib fracture segmentation model and the efficacy of data augmentation techniques such as left-right flipping, rotation, and scaling. Augmented dataset with L/R flipping and rotations(30°, 60°) increased model performance, however, dataset with rotation(90°) and ⨯0.5 rescaling decreased model performance. This indicates the usage of appropriate data augmentation methods depending on datasets and tasks.

Crosswalk Detection Model for Visually impaired Using Deep Learning (딥러닝을 이용한 시각장애인용 횡단보도 탐지 모델 연구)

  • Junsoo Kim;Hyuk Lee
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.17 no.1
    • /
    • pp.67-75
    • /
    • 2024
  • Crosswalks play an important role for the safe movement of pedestrians in a complex urban environment. However, for the visually impaired, crosswalks can be a big risk factor. Although assistive tools such as braille blocks and acoustic traffic lights exist for safe walking, poor management can sometimes act as a hindrance to safety. This paper proposes a method to improve accuracy in a deep learning-based real-time crosswalk detection model that can be used in applications for pedestrian assistance for the disabled at the beginning. The image was binarized by utilizing the characteristic that the white line of the crosswalk image contrasts with the road surface, and through this, the crosswalk could be better recognized and the location of the crosswalk could be more accurately identified by using two models that learned the whole and the middle part of the crosswalk, respectively. In addition, it was intended to increase accuracy by creating a boundary box that recognizes crosswalks in two stages: whole and part. Through this method, additional frames that the detection model did not detect in RGB image learning from the crosswalk image could be detected.