• 제목/요약/키워드: Image data training

검색결과 673건 처리시간 0.022초

Document Image Binarization by GAN with Unpaired Data Training

  • Dang, Quang-Vinh;Lee, Guee-Sang
    • International Journal of Contents
    • /
    • 제16권2호
    • /
    • pp.8-18
    • /
    • 2020
  • Data is critical in deep learning but the scarcity of data often occurs in research, especially in the preparation of the paired training data. In this paper, document image binarization with unpaired data is studied by introducing adversarial learning, excluding the need for supervised or labeled datasets. However, the simple extension of the previous unpaired training to binarization inevitably leads to poor performance compared to paired data training. Thus, a new deep learning approach is proposed by introducing a multi-diversity of higher quality generated images. In this paper, a two-stage model is proposed that comprises the generative adversarial network (GAN) followed by the U-net network. In the first stage, the GAN uses the unpaired image data to create paired image data. With the second stage, the generated paired image data are passed through the U-net network for binarization. Thus, the trained U-net becomes the binarization model during the testing. The proposed model has been evaluated over the publicly available DIBCO dataset and it outperforms other techniques on unpaired training data. The paper shows the potential of using unpaired data for binarization, for the first time in the literature, which can be further improved to replace paired data training for binarization in the future.

Reference Map을 이용한 시계열 image data의 자동분류법 (Automatic Classification Method for Time-Series Image Data using Reference Map)

  • 홍선표
    • 한국음향학회지
    • /
    • 제16권2호
    • /
    • pp.58-65
    • /
    • 1997
  • 본 논문에서는 시계열 image data를 안정되고 높은 정확도로 분류할 수 있는 자동분류법을 제안하였다. 제안한 방법은 대상 영역에 관한 분류도가 기존재하던 가, 아니면 최소한 시계열 image data 중 어느 한 image data가 분류되어 있다고 하는 전제조건에 그 기초를 두고 있다. 분류도는 training area를 선정하기 위라여 사용하는 기준주제도로 사용되어진다. 제안한 방법은 1)기준주제도를 사용한 training data의 추출, 2)taining data의 균질성에 의거한 변화화소의 검출, 3)검출된 변화화소에 대한 clustering, 4)training data의 재구성, 5)maximum likelihood classifier와 같은 판별법에 의한 분류 등 5개의 단계로 구성된다. 제안한 방법의 성능을 정량적으로 평가하기 위하여 4개의 시계열 Landsat TM image data를 제안한 방법과 숙련된 operator가 필요한 기존의 방법으로 각각 분류하여 비교 검토하였다. 그 결과, 기존의 방법으로는 숙련된 operator가 필요하고, 분류도를 얻기까지 수일이 소요되는 데 반하여, 제안한 방법으로는 숙련된 operator 없이, 신뢰성 있는 분류도를 수 시간 내에 자동으로 얻을 수 있었다.

  • PDF

Automated Training from Landsat Image for Classification of SPOT-5 and QuickBird Images

  • Kim, Yong-Min;Kim, Yong-Il;Park, Wan-Yong;Eo, Yang-Dam
    • 대한원격탐사학회지
    • /
    • 제26권3호
    • /
    • pp.317-324
    • /
    • 2010
  • In recent years, many automatic classification approaches have been employed. An automatic classification method can be effective, time-saving and can produce objective results due to the exclusion of operator intervention. This paper proposes a classification method based on automated training for high resolution multispectral images using ancillary data. Generally, it is problematic to automatically classify high resolution images using ancillary data, because of the scale difference between the high resolution image and the ancillary data. In order to overcome this problem, the proposed method utilizes the classification results of a Landsat image as a medium for automatic classification. For the classification of a Landsat image, a maximum likelihood classification is applied to the image, and the attributes of ancillary data are entered as the training data. In the case of a high resolution image, a K-means clustering algorithm, an unsupervised classification, was conducted and the result was compared to the classification results of the Landsat image. Subsequently, the training data of the high resolution image was automatically extracted using regular rules based on a RELATIONAL matrix that shows the relation between the two results. Finally, a high resolution image was classified and updated using the extracted training data. The proposed method was applied to QuickBird and SPOT-5 images of non-accessible areas. The result showed good performance in accuracy assessments. Therefore, we expect that the method can be effectively used to automatically construct thematic maps for non-accessible areas and update areas that do not have any attributes in geographic information system.

군 로봇의 장소 분류 정확도 향상을 위한 적외선 이미지 데이터 결합 학습 방법 연구 (A Study on the Training Methodology of Combining Infrared Image Data for Improving Place Classification Accuracy of Military Robots)

  • 최동규;도승원;이창은
    • 로봇학회논문지
    • /
    • 제18권3호
    • /
    • pp.293-298
    • /
    • 2023
  • The military is facing a continuous decrease in personnel, and in order to cope with potential accidents and challenges in operations, efforts are being made to reduce the direct involvement of personnel by utilizing the latest technologies. Recently, the use of various sensors related to Manned-Unmanned Teaming and artificial intelligence technologies has gained attention, emphasizing the need for flexible utilization methods. In this paper, we propose four dataset construction methods that can be used for effective training of robots that can be deployed in military operations, utilizing not only RGB image data but also data acquired from IR image sensors. Since there is no publicly available dataset that combines RGB and IR image data, we directly acquired the dataset within buildings. The input values were constructed by combining RGB and IR image sensor data, taking into account the field of view, resolution, and channel values of both sensors. We compared the proposed method with conventional RGB image data classification training using the same learning model. By employing the proposed image data fusion method, we observed improved stability in training loss and approximately 3% higher accuracy.

회전한 상표 이미지의 진위 결정을 위한 기계 학습 데이터 확장 방법 (Machine Learning Data Extension Way for Confirming Genuine of Trademark Image which is Rotated)

  • 구본근
    • Journal of Platform Technology
    • /
    • 제8권1호
    • /
    • pp.16-23
    • /
    • 2020
  • 상표권 보호를 위한 상표 이미지의 진위 결정에 심층 신경망인 합성곱 신경망을 이용할 수 있다. 이를 위해, 상표로 등록되어 있는 한 장의 상표 이미지를 반복적으로 학습하는 것은 기계학습의 성능을 감소시키는 원인이 된다. 따라서, 이러한 응용에서 학습 데이터는 다양한 방법으로 생성된다. 하지만 대상 이미지가 회전되어 있으면 원본이라 하더라도 인식하지 못하거나 위조 상표로 분류되기도 한다. 본 논문에서는 회전한 상표 이미지의 진위 결정을 위한 기계학습 데이터의 확장 방법을 제안한다. 본 논문에서 제안하는 학습 데이터 확장 방법은 기울어진 이미지를 생성하고 이를 학습 데이터로 사용하는 것이다. 본 논문에서 제안하는 학습 데이터 확장 방법의 유효성 검증을 위해 대학의 로고를 대상으로 학습 데이터를 생성하였으며, 이를 활용하여 합성곱 신경망을 학습시킨 후 검증용 데이터를 이용하여 정확도를 평가하였다. 정확도 평가 결과에 따르면 본 논문에서 제안한 방법으로 생성한 학습 데이터를 활용하면 회전한 상표를 대상으로 한 진위 여부 결정에 합성곱 신경망을 활용할 수 있다.

  • PDF

A Study on Improving the Accuracy of Medical Images Classification Using Data Augmentation

  • Cheon-Ho Park;Min-Guan Kim;Seung-Zoon Lee;Jeongil Choi
    • 한국컴퓨터정보학회논문지
    • /
    • 제28권12호
    • /
    • pp.167-174
    • /
    • 2023
  • 본 연구는 합성곱 신경망 모델에서 이미지 데이터 증강을 통하여 대장암 진단 모델의 정확도를 개선하고자 하였다. 이미지 데이터 증강은 기초 이미지 조작 방법을 이용하여 뒤집기, 회전, 이동, 밀림, 주밍을 사용하였다. 본 연구에서는 실험설계를 위해 보유하고 있는 5000개의 이미지 데이터에 대해 훈련 데이터와 평가 데이터로 각각 4000개와 1000개로 나누었으며, 훈련 데이터 4000개에 대해 이미지 데이터 증강 기법으로 4000개와 8000개의 이미지를 추가하여 모델을 학습시켰다. 평가 결과는 훈련 데이터 4000개, 8000개, 12000개에 대한 분류 정확도가 각각 85.1%, 87.0%, 90.2%로 나왔으며 이미지 데이터 증강에 따른 개선 효과를 확인하였다.

애완동물 분류를 위한 딥러닝 (Deep Learning for Pet Image Classification)

  • 신광성;신성윤
    • 한국정보통신학회:학술대회논문집
    • /
    • 한국정보통신학회 2019년도 춘계학술대회
    • /
    • pp.151-152
    • /
    • 2019
  • 본 논문에서는 동물 이미지 분류를위한 작은 데이터 세트를 기반으로 개선 된 심층 학습 방법을 제안한다. 첫째, CNN은 소규모 데이터 세트에 대한 교육 모델을 작성하고 데이터 세트를 사용하여 교육 세트의 데이터 세트를 확장하는 데 사용된다. 둘째, VGG16과 같은 대규모 데이터 세트에 사전 훈련 된 네트워크를 사용하여 작은 데이터 세트의 병목을 추출하여 새로운 교육 데이터 세트 및 테스트 데이터 세트로 두 개의 NumPy 파일에 저장하고, 마지막으로 완전히 연결된 네트워크를 새로운 데이터 세트로 학습한다.

  • PDF

Development of ResNet-based WBC Classification Algorithm Using Super-pixel Image Segmentation

  • Lee, Kyu-Man;Kang, Soon-Ah
    • 한국컴퓨터정보학회논문지
    • /
    • 제23권4호
    • /
    • pp.147-153
    • /
    • 2018
  • In this paper, we propose an efficient WBC 14-Diff classification which performs using the WBC-ResNet-152, a type of CNN model. The main point of view is to use Super-pixel for the segmentation of the image of WBC, and to use ResNet for the classification of WBC. A total of 136,164 blood image samples (224x224) were grouped for image segmentation, training, training verification, and final test performance analysis. Image segmentation using super-pixels have different number of images for each classes, so weighted average was applied and therefore image segmentation error was low at 7.23%. Using the training data-set for training 50 times, and using soft-max classifier, TPR average of 80.3% for the training set of 8,827 images was achieved. Based on this, using verification data-set of 21,437 images, 14-Diff classification TPR average of normal WBCs were at 93.4% and TPR average of abnormal WBCs were at 83.3%. The result and methodology of this research demonstrates the usefulness of artificial intelligence technology in the blood cell image classification field. WBC-ResNet-152 based morphology approach is shown to be meaningful and worthwhile method. And based on stored medical data, in-depth diagnosis and early detection of curable diseases is expected to improve the quality of treatment.

전이학습 기반 사출 성형품 burr 이미지 검출 시스템 개발 (Development of a transfer learning based detection system for burr image of injection molded products)

  • 양동철;김종선
    • Design & Manufacturing
    • /
    • 제15권3호
    • /
    • pp.1-6
    • /
    • 2021
  • An artificial neural network model based on a deep learning algorithm is known to be more accurate than humans in image classification, but there is still a limit in the sense that there needs to be a lot of training data that can be called big data. Therefore, various techniques are being studied to build an artificial neural network model with high precision, even with small data. The transfer learning technique is assessed as an excellent alternative. As a result, the purpose of this study is to develop an artificial neural network system that can classify burr images of light guide plate products with 99% accuracy using transfer learning technique. Specifically, for the light guide plate product, 150 images of the normal product and the burr were taken at various angles, heights, positions, etc., respectively. Then, after the preprocessing of images such as thresholding and image augmentation, for a total of 3,300 images were generated. 2,970 images were separated for training, while the remaining 330 images were separated for model accuracy testing. For the transfer learning, a base model was developed using the NASNet-Large model that pre-trained 14 million ImageNet data. According to the final model accuracy test, the 99% accuracy in the image classification for training and test images was confirmed. Consequently, based on the results of this study, it is expected to help develop an integrated AI production management system by training not only the burr but also various defective images.

Deep survey using deep learning: generative adversarial network

  • Park, Youngjun;Choi, Yun-Young;Moon, Yong-Jae;Park, Eunsu;Lim, Beomdu;Kim, Taeyoung
    • 천문학회보
    • /
    • 제44권2호
    • /
    • pp.78.1-78.1
    • /
    • 2019
  • There are a huge number of faint objects that have not been observed due to the lack of large and deep surveys. In this study, we demonstrate that a deep learning approach can produce a better quality deep image from a single pass imaging so that could be an alternative of conventional image stacking technique or the expensive large and deep surveys. Using data from the Sloan Digital Sky Survey (SDSS) stripe 82 which provide repeatedly scanned imaging data, a training data set is constructed: g-, r-, and i-band images of single pass data as an input and r-band co-added image as a target. Out of 151 SDSS fields that have been repeatedly scanned 34 times, 120 fields were used for training and 31 fields for validation. The size of a frame selected for the training is 1k by 1k pixel scale. To avoid possible problems caused by the small number of training sets, frames are randomly selected within that field each iteration of training. Every 5000 iterations of training, the performance were evaluated with RMSE, peak signal-to-noise ratio which is given on logarithmic scale, structural symmetry index (SSIM) and difference in SSIM. We continued the training until a GAN model with the best performance is found. We apply the best GAN-model to NGC0941 located in SDSS stripe 82. By comparing the radial surface brightness and photometry error of images, we found the possibility that this technique could generate a deep image with statistics close to the stacked image from a single-pass image.

  • PDF