• 제목/요약/키워드: Image machine learning

검색결과 592건 처리시간 0.036초

A Review of Computer Vision Methods for Purpose on Computer-Aided Diagnosis

  • Song, Hyewon;Nguyen, Anh-Duc;Gong, Myoungsik;Lee, Sanghoon
    • Journal of International Society for Simulation Surgery
    • /
    • 제3권1호
    • /
    • pp.1-8
    • /
    • 2016
  • In the field of Radiology, the Computer Aided Diagnosis is the technology which gives valuable information for surgical purpose. For its importance, several computer vison methods are processed to obtain useful information of images acquired from the imaging devices such as X-ray, Magnetic Resonance Imaging (MRI) and Computed Tomography (CT). These methods, called pattern recognition, extract features from images and feed them to some machine learning algorithm to find out meaningful patterns. Then the learned machine is then used for exploring patterns from unseen images. The radiologist can therefore easily find the information used for surgical planning or diagnosis of a patient through the Computer Aided Diagnosis. In this paper, we present a review on three widely-used methods applied to Computer Aided Diagnosis. The first one is the image processing methods which enhance meaningful information such as edge and remove the noise. Based on the improved image quality, we explain the second method called segmentation which separates the image into a set of regions. The separated regions such as bone, tissue, organs are then delivered to machine learning algorithms to extract representative information. We expect that this paper gives readers basic knowledges of the Computer Aided Diagnosis and intuition about computer vision methods applied in this area.

Improving Chest X-ray Image Classification via Integration of Self-Supervised Learning and Machine Learning Algorithms

  • Tri-Thuc Vo;Thanh-Nghi Do
    • Journal of information and communication convergence engineering
    • /
    • 제22권2호
    • /
    • pp.165-171
    • /
    • 2024
  • In this study, we present a novel approach for enhancing chest X-ray image classification (normal, Covid-19, edema, mass nodules, and pneumothorax) by combining contrastive learning and machine learning algorithms. A vast amount of unlabeled data was leveraged to learn representations so that data efficiency is improved as a means of addressing the limited availability of labeled data in X-ray images. Our approach involves training classification algorithms using the extracted features from a linear fine-tuned Momentum Contrast (MoCo) model. The MoCo architecture with a Resnet34, Resnet50, or Resnet101 backbone is trained to learn features from unlabeled data. Instead of only fine-tuning the linear classifier layer on the MoCopretrained model, we propose training nonlinear classifiers as substitutes for softmax in deep networks. The empirical results show that while the linear fine-tuned ImageNet-pretrained models achieved the highest accuracy of only 82.9% and the linear fine-tuned MoCo-pretrained models an increased highest accuracy of 84.8%, our proposed method offered a significant improvement and achieved the highest accuracy of 87.9%.

Fast Face Gender Recognition by Using Local Ternary Pattern and Extreme Learning Machine

  • Yang, Jucheng;Jiao, Yanbin;Xiong, Naixue;Park, DongSun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제7권7호
    • /
    • pp.1705-1720
    • /
    • 2013
  • Human face gender recognition requires fast image processing with high accuracy. Existing face gender recognition methods used traditional local features and machine learning methods have shortcomings of low accuracy or slow speed. In this paper, a new framework for face gender recognition to reach fast face gender recognition is proposed, which is based on Local Ternary Pattern (LTP) and Extreme Learning Machine (ELM). LTP is a generalization of Local Binary Pattern (LBP) that is in the presence of monotonic illumination variations on a face image, and has high discriminative power for texture classification. It is also more discriminate and less sensitive to noise in uniform regions. On the other hand, ELM is a new learning algorithm for generalizing single hidden layer feed forward networks without tuning parameters. The main advantages of ELM are the less stringent optimization constraints, faster operations, easy implementation, and usually improved generalization performance. The experimental results on public databases show that, in comparisons with existing algorithms, the proposed method has higher precision and better generalization performance at extremely fast learning speed.

데이터 증강을 통한 기계학습 능력 개선 방법 연구 (Study on the Improvement of Machine Learning Ability through Data Augmentation)

  • 김태우;신광성
    • 한국정보통신학회:학술대회논문집
    • /
    • 한국정보통신학회 2021년도 춘계학술대회
    • /
    • pp.346-347
    • /
    • 2021
  • 기계학습을 위한 패턴인식을 위해서는 학습데이터의 양이 많을수록 그 성능이 향상된다. 하지만 일상에서 검출해내야하는 패턴의 종류 및 정보가 항상 많은 양의 학습데이터를 확보할 수는 없다. 따라서 일반적인 기계학습을 위해 적은데이터셋을 의미있게 부풀릴 필요가 있다. 본 연구에서는 기계학습을 수행할 수 있도록 데이터를 증강시키는 기법에 관해 연구한다. 적은데이터셋을 이용하여 기계학습을 수행하는 대표적인 방법이 전이학습(transfer learning) 기법이다. 전이학습은 범용데이터셋으로 기본적인 학습을 수행한 후 목표데이터셋을 최종 단계에 대입함으로써 결과를 얻어내는 방법이다. 본 연구에서는 ImageNet과 같은 범용데이터셋으로 학습시킨 학습모델을 증강된 데이터를 이용하여 특징추출셋으로 사용하여 원하는 패턴에 대한 검출을 수행한다.

  • PDF

머신러닝 기법을 활용한 대용량 시계열 데이터 이상 시점탐지 방법론 : 발전기 부품신호 사례 중심 (Anomaly Detection of Big Time Series Data Using Machine Learning)

  • 권세혁
    • 산업경영시스템학회지
    • /
    • 제43권2호
    • /
    • pp.33-38
    • /
    • 2020
  • Anomaly detection of Machine Learning such as PCA anomaly detection and CNN image classification has been focused on cross-sectional data. In this paper, two approaches has been suggested to apply ML techniques for identifying the failure time of big time series data. PCA anomaly detection to identify time rows as normal or abnormal was suggested by converting subjects identification problem to time domain. CNN image classification was suggested to identify the failure time by re-structuring of time series data, which computed the correlation matrix of one minute data and converted to tiff image format. Also, LASSO, one of feature selection methods, was applied to select the most affecting variables which could identify the failure status. For the empirical study, time series data was collected in seconds from a power generator of 214 components for 25 minutes including 20 minutes before the failure time. The failure time was predicted and detected 9 minutes 17 seconds before the failure time by PCA anomaly detection, but was not detected by the combination of LASSO and PCA because the target variable was binary variable which was assigned on the base of the failure time. CNN image classification with the train data of 10 normal status image and 5 failure status images detected just one minute before.

이미지 분석을 위한 퓨샷 학습의 최신 연구동향 (Recent advances in few-shot learning for image domain: a survey)

  • 석호식
    • 전기전자학회논문지
    • /
    • 제27권4호
    • /
    • pp.537-547
    • /
    • 2023
  • 퓨삿학습(few-shot learning)은 사전에 확보한 관련 지식과 소규모의 학습데이터를 이용하여 학습데이터의 부족으로 인한 어려움을 해결할 수 있는 가능성을 제시해주어 최근 많은 주목을 받고 있다. 본 논문에서는 퓨삿학습의 개념과 주요 접근방법을 빠르게 파악할 수 있도록 데이터 증강, 임베딩과 측도학습, 메타학습의 세 관점에서 최신연구동향을 설명한다. 또한 퓨샷학습을 적용하려는 연구자들에게 도움을 제공할 수 있도록 주요 벤치마크 데이터셋에 대하여 간략하게 소개하였다. 퓨삿학습은 이미지 분석과 자연어 처리 등 다양한 분야에서 활용되고 있으나, 본 논문은 이미지 처리를 위한 퓨삿학습의 접근법에 집중하였다.

딥 러닝 기반의 이미지 압축 알고리즘에 관한 연구 (Study on Image Compression Algorithm with Deep Learning)

  • 이용환
    • 반도체디스플레이기술학회지
    • /
    • 제21권4호
    • /
    • pp.156-162
    • /
    • 2022
  • Image compression plays an important role in encoding and improving various forms of images in the digital era. Recent researches have focused on the principle of deep learning as one of the most exciting machine learning methods to show that it is good scheme to analyze, classify and compress images. Various neural networks are able to adapt for image compressions, such as deep neural networks, artificial neural networks, recurrent neural networks and convolution neural networks. In this review paper, we discussed how to apply the rule of deep learning to obtain better image compression with high accuracy, low loss-ness and high visibility of the image. For those results in performance, deep learning methods are required on justified manner with distinct analysis.

BEGINNER'S GUIDE TO NEURAL NETWORKS FOR THE MNIST DATASET USING MATLAB

  • Kim, Bitna;Park, Young Ho
    • Korean Journal of Mathematics
    • /
    • 제26권2호
    • /
    • pp.337-348
    • /
    • 2018
  • MNIST dataset is a database containing images of handwritten digits, with each image labeled by an integer from 0 to 9. It is used to benchmark the performance of machine learning algorithms. Neural networks for MNIST are regarded as the starting point of the studying machine learning algorithms. However it is not easy to start the actual programming. In this expository article, we will give a step-by-step instruction to build neural networks for MNIST dataset using MATLAB.

Infrared and Visible Image Fusion Based on NSCT and Deep Learning

  • Feng, Xin
    • Journal of Information Processing Systems
    • /
    • 제14권6호
    • /
    • pp.1405-1419
    • /
    • 2018
  • An image fusion method is proposed on the basis of depth model segmentation to overcome the shortcomings of noise interference and artifacts caused by infrared and visible image fusion. Firstly, the deep Boltzmann machine is used to perform the priori learning of infrared and visible target and background contour, and the depth segmentation model of the contour is constructed. The Split Bregman iterative algorithm is employed to gain the optimal energy segmentation of infrared and visible image contours. Then, the nonsubsampled contourlet transform (NSCT) transform is taken to decompose the source image, and the corresponding rules are used to integrate the coefficients in the light of the segmented background contour. Finally, the NSCT inverse transform is used to reconstruct the fused image. The simulation results of MATLAB indicates that the proposed algorithm can obtain the fusion result of both target and background contours effectively, with a high contrast and noise suppression in subjective evaluation as well as great merits in objective quantitative indicators.

Machine learning application for predicting the strawberry harvesting time

  • Yang, Mi-Hye;Nam, Won-Ho;Kim, Taegon;Lee, Kwanho;Kim, Younghwa
    • 농업과학연구
    • /
    • 제46권2호
    • /
    • pp.381-393
    • /
    • 2019
  • A smart farm is a system that combines information and communication technology (ICT), internet of things (IoT), and agricultural technology that enable a farm to operate with minimal labor and to automatically control of a greenhouse environment. Machine learning based on recently data-driven techniques has emerged with big data technologies and high-performance computing to create opportunities to quantify data intensive processes in agricultural operational environments. This paper presents research on the application of machine learning technology to diagnose the growth status of crops and predicting the harvest time of strawberries in a greenhouse according to image processing techniques. To classify the growth stages of the strawberries, we used object inference and detection with machine learning model based on deep learning neural networks and TensorFlow. The classification accuracy was compared based on the training data volume and training epoch. As a result, it was able to classify with an accuracy of over 90% with 200 training images and 8,000 training steps. The detection and classification of the strawberry maturities could be identified with an accuracy of over 90% at the mature and over mature stages of the strawberries. Concurrently, the experimental results are promising, and they show that this approach can be applied to develop a machine learning model for predicting the strawberry harvesting time and can be used to provide key decision support information to both farmers and policy makers about optimal harvest times and harvest planning.