• Title/Summary/Keyword: GAN(Generative Adversarial Network

Search Result 178, Processing Time 0.03 seconds

Assessment and Analysis of Fidelity and Diversity for GAN-based Medical Image Generative Model (GAN 기반 의료영상 생성 모델에 대한 품질 및 다양성 평가 및 분석)

  • Jang, Yoojin;Yoo, Jaejun;Hong, Helen
    • Journal of the Korea Computer Graphics Society
    • /
    • v.28 no.2
    • /
    • pp.11-19
    • /
    • 2022
  • Recently, various researches on medical image generation have been suggested, and it becomes crucial to accurately evaluate the quality and diversity of the generated medical images. For this purpose, the expert's visual turing test, feature distribution visualization, and quantitative evaluation through IS and FID are evaluated. However, there are few methods for quantitatively evaluating medical images in terms of fidelity and diversity. In this paper, images are generated by learning a chest CT dataset of non-small cell lung cancer patients through DCGAN and PGGAN generative models, and the performance of the two generative models are evaluated in terms of fidelity and diversity. The performance is quantitatively evaluated through IS and FID, which are one-dimensional score-based evaluation methods, and Precision and Recall, Improved Precision and Recall, which are two-dimensional score-based evaluation methods, and the characteristics and limitations of each evaluation method are also analyzed in medical imaging.

Waste Classification by Fine-Tuning Pre-trained CNN and GAN

  • Alsabei, Amani;Alsayed, Ashwaq;Alzahrani, Manar;Al-Shareef, Sarah
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.8
    • /
    • pp.65-70
    • /
    • 2021
  • Waste accumulation is becoming a significant challenge in most urban areas and if it continues unchecked, is poised to have severe repercussions on our environment and health. The massive industrialisation in our cities has been followed by a commensurate waste creation that has become a bottleneck for even waste management systems. While recycling is a viable solution for waste management, it can be daunting to classify waste material for recycling accurately. In this study, transfer learning models were proposed to automatically classify wastes based on six materials (cardboard, glass, metal, paper, plastic, and trash). The tested pre-trained models were ResNet50, VGG16, InceptionV3, and Xception. Data augmentation was done using a Generative Adversarial Network (GAN) with various image generation percentages. It was found that models based on Xception and VGG16 were more robust. In contrast, models based on ResNet50 and InceptionV3 were sensitive to the added machine-generated images as the accuracy degrades significantly compared to training with no artificial data.

Predicting Blood Glucose Data and Ensuring Data Integrity Based on Artificial Intelligence (인공지능 기반 혈당 데이터 예측 및 데이터 무결성 보장 연구)

  • Lee, Tae Kang
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.201-203
    • /
    • 2022
  • Over the past five years, the number of patients treated for diabetes has increased by 27.7% to 3.22 million, and since blood sugar is still checked through finger blood collection, continuous blood glucose measurement and blood sugar peak confirmation are difficult and painful. To solve this problem, based on blood sugar data measured for 14 days, three months of blood sugar prediction data are provided to diabetics using artificial intelligence technology.

  • PDF

Vehicle Detection at Night Based on Style Transfer Image Enhancement

  • Jianing Shen;Rong Li
    • Journal of Information Processing Systems
    • /
    • v.19 no.5
    • /
    • pp.663-672
    • /
    • 2023
  • Most vehicle detection methods have poor vehicle feature extraction performance at night, and their robustness is reduced; hence, this study proposes a night vehicle detection method based on style transfer image enhancement. First, a style transfer model is constructed using cycle generative adversarial networks (cycleGANs). The daytime data in the BDD100K dataset were converted into nighttime data to form a style dataset. The dataset was then divided using its labels. Finally, based on a YOLOv5s network, a nighttime vehicle image is detected for the reliable recognition of vehicle information in a complex environment. The experimental results of the proposed method based on the BDD100K dataset show that the transferred night vehicle images are clear and meet the requirements. The precision, recall, mAP@.5, and mAP@.5:.95 reached 0.696, 0.292, 0.761, and 0.454, respectively.

Generating GAN-based Virtual data to Prevent the Spread of Highly Pathogenic Avian Influenza(HPAI) (고위험성 조류인플루엔자(HPAI) 확산 방지를 위한 GAN 기반 가상 데이터 생성)

  • Choi, Dae-Woo;Han, Ye-Ji;Song, Yu-Han;Kang, Tae-Hun;Lee, Won-Been
    • The Journal of Bigdata
    • /
    • v.5 no.2
    • /
    • pp.69-76
    • /
    • 2020
  • This study was conducted with the support of the Information and Communication Technology Promotion Center, funded by the government (Ministry of Science and ICT) in 2019. Highly pathogenic avian influenza (HPAI) is an acute infectious disease of birds caused by highly pathogenic avian influenza virus infection, causing serious damage to poultry such as chickens and ducks. High pathogenic avian influenza (HPAI) is caused by focusing on winter rather than year-round, and sometimes does not occur at all during a certain period of time. Due to these characteristics of HPAI, there is a problem that does not accumulate enough actual data. In this paper study, GAN network was utilized to generate actual similar data containing missing values and the process is introduced. The results of this study can be used to measure risk by generating realistic simulation data for certain times when HPAI did not occur.

A Study on Image Generation from Sentence Embedding Applying Self-Attention (Self-Attention을 적용한 문장 임베딩으로부터 이미지 생성 연구)

  • Yu, Kyungho;No, Juhyeon;Hong, Taekeun;Kim, Hyeong-Ju;Kim, Pankoo
    • Smart Media Journal
    • /
    • v.10 no.1
    • /
    • pp.63-69
    • /
    • 2021
  • When a person sees a sentence and understands the sentence, the person understands the sentence by reminiscent of the main word in the sentence as an image. Text-to-image is what allows computers to do this associative process. The previous deep learning-based text-to-image model extracts text features using Convolutional Neural Network (CNN)-Long Short Term Memory (LSTM) and bi-directional LSTM, and generates an image by inputting it to the GAN. The previous text-to-image model uses basic embedding in text feature extraction, and it takes a long time to train because images are generated using several modules. Therefore, in this research, we propose a method of extracting features by using the attention mechanism, which has improved performance in the natural language processing field, for sentence embedding, and generating an image by inputting the extracted features into the GAN. As a result of the experiment, the inception score was higher than that of the model used in the previous study, and when judged with the naked eye, an image that expresses the features well in the input sentence was created. In addition, even when a long sentence is input, an image that expresses the sentence well was created.

GENERATION OF FUTURE MAGNETOGRAMS FROM PREVIOUS SDO/HMI DATA USING DEEP LEARNING

  • Jeon, Seonggyeong;Moon, Yong-Jae;Park, Eunsu;Shin, Kyungin;Kim, Taeyoung
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.44 no.1
    • /
    • pp.82.3-82.3
    • /
    • 2019
  • In this study, we generate future full disk magnetograms in 12, 24, 36 and 48 hours advance from SDO/HMI images using deep learning. To perform this generation, we apply the convolutional generative adversarial network (cGAN) algorithm to a series of SDO/HMI magnetograms. We use SDO/HMI data from 2011 to 2016 for training four models. The models make AI-generated images for 2017 HMI data and compare them with the actual HMI magnetograms for evaluation. The AI-generated images by each model are very similar to the actual images. The average correlation coefficient between the two images for about 600 data sets are about 0.85 for four models. We are examining hundreds of active regions for more detail comparison. In the future we will use pix2pix HD and video2video translation networks for image prediction.

  • PDF

Generation of global coronal field extrapolation from frontside and AI-generated farside magnetograms

  • Jeong, Hyunjin;Moon, Yong-Jae;Park, Eunsu;Lee, Harim;Kim, Taeyoung
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.44 no.1
    • /
    • pp.52.2-52.2
    • /
    • 2019
  • Global map of solar surface magnetic field, such as the synoptic map or daily synchronic frame, does not tell us real-time information about the far side of the Sun. A deep-learning technique based on Conditional Generative Adversarial Network (cGAN) is used to generate farside magnetograms from EUVI $304{\AA}$ of STEREO spacecrafts by training SDO spacecraft's data pairs of HMI and AIA $304{\AA}$. Farside(or backside) data of daily synchronic frames are replaced by the Ai-generated magnetograms. The new type of data is used to calculate the Potential Field Source Surface (PFSS) model. We compare the results of the global field with observations as well as those of the conventional method. We will discuss advantage and disadvantage of the new method and future works.

  • PDF

A Study on the Video Quality Improvement of National Intangible Cultural Heritage Documentary Film (국가무형문화재 기록영상 화질 개선에 관한 연구)

  • Kwon, Do-Hyung;Yu, Jeong-Min
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2020.07a
    • /
    • pp.439-441
    • /
    • 2020
  • 본 논문에서는 국가무형문화재 기록영상의 화질 개선에 관한 연구를 진행한다. 기록영상의 화질 개선을 위해 SRGAN 기반의 초해상화 복원영상 생성 프레임워크의 적용을 제안한다. Image aumentation과 median filter를 적용한 데이터셋과 적대적 신경망인 Generative Adversarial Network (GAN)을 기반으로 딥러닝 네트워크를 구축하여 입력된 Low-Resolution 이미지를 통해 High-Resolution의 복원 영상을 생성한다. 이 연구를 통해 국가무형문화재 기록영상 뿐만 아니라 문화재 전반의 사진 및 영상 기록 자료의 품질 개선 가능성을 제시하고, 영상 기록 자료의 아카이브 구축을 통해 지속적인 활용의 기초연구가 되는 것을 목표로 한다.

  • PDF

Research on Digital Construction Site Management Using Drone and Vision Processing Technology (드론 및 비전 프로세싱 기술을 활용한 디지털 건설현장 관리에 대한 연구)

  • Seo, Min Jo;Park, Kyung Kyu;Lee, Seung Been;Kim, Si Uk;Choi, Won Jun;Kim, Chee Kyeung
    • Proceedings of the Korean Institute of Building Construction Conference
    • /
    • 2023.11a
    • /
    • pp.239-240
    • /
    • 2023
  • Construction site management involves overseeing tasks from the construction phase to the maintenance stage, and digitalization of construction sites is necessary for digital construction site management. In this study, we aim to conduct research on object recognition at construction sites using drones. Images of construction sites captured by drones are reconstructed into BIM (Building Information Modeling) models, and objects are recognized after partially rendering the models using artificial intelligence. For the photorealistic rendering of the BIM models, both traditional filtering techniques and the generative adversarial network (GAN) model were used, while the YOLO (You Only Look Once) model was employed for object recognition. This study is expected to provide insights into the research direction of digital construction site management and help assess the potential and future value of introducing artificial intelligence in the construction industry.

  • PDF