• Title/Summary/Keyword: GANs

Search Result 60, Processing Time 0.031 seconds

Real-time transmission of 3G point cloud data based on cGANs (cGANs 기반 3D 포인트 클라우드 데이터의 실시간 전송 기법)

  • Shin, Kwang-Seong;Shin, Seong-Yoon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.23 no.11
    • /
    • pp.1482-1484
    • /
    • 2019
  • We present a method for transmitting 3D object information in real time in a telepresence system. Three-dimensional object information consists of a large amount of point cloud data, which requires high performance computing power and ultra-wideband network transmission environment to process and transmit such a large amount of data in real time. In this paper, multiple users can transmit object motion and facial expression information in real time even in small network bands by using GANs (Generative Adversarial Networks), a non-supervised learning machine learning algorithm, for real-time transmission of 3D point cloud data. In particular, we propose the creation of an object similar to the original using only the feature information of 3D objects using conditional GANs.

A Study on Auction-Inspired Multi-GAN Training (경매 메커니즘을 이용한 다중 적대적 생성 신경망 학습에 관한 연구)

  • Joo Yong Shim;Jean Seong Bjorn Choe;Jong-Kook Kim
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.05a
    • /
    • pp.527-529
    • /
    • 2023
  • Generative Adversarial Networks (GANs) models have developed rapidly due to the emergence of various variation models and their wide applications. Despite many recent developments in GANs, mode collapse, and instability are still unresolved issues. To address these problems, we focused on the fact that a single GANs model itself cannot realize local failure during the training phase without external standards. This paper introduces a novel training process involving multiple GANs, inspired by auction mechanisms. During the training, auxiliary performance metrics for each GANs are determined by the others through the process of various auction methods.

A deep learning framework for wind pressure super-resolution reconstruction

  • Xiao Chen;Xinhui Dong;Pengfei Lin;Fei Ding;Bubryur Kim;Jie Song;Yiqing Xiao;Gang Hu
    • Wind and Structures
    • /
    • v.36 no.6
    • /
    • pp.405-421
    • /
    • 2023
  • Strong wind is the main factors of wind-damage of high-rise buildings, which often creates largely economical losses and casualties. Wind pressure plays a critical role in wind effects on buildings. To obtain the high-resolution wind pressure field, it often requires massive pressure taps. In this study, two traditional methods, including bilinear and bicubic interpolation, and two deep learning techniques including Residual Networks (ResNet) and Generative Adversarial Networks (GANs), are employed to reconstruct wind pressure filed from limited pressure taps on the surface of an ideal building from TPU database. It was found that the GANs model exhibits the best performance in reconstructing the wind pressure field. Meanwhile, it was confirmed that k-means clustering based retained pressure taps as model input can significantly improve the reconstruction ability of GANs model. Finally, the generalization ability of k-means clustering based GANs model in reconstructing wind pressure field is verified by an actual engineering structure. Importantly, the k-means clustering based GANs model can achieve satisfactory reconstruction in wind pressure field under the inputs processing by k-means clustering, even the 20% of pressure taps. Therefore, it is expected to save a huge number of pressure taps under the field reconstruction and achieve timely and accurately reconstruction of wind pressure field under k-means clustering based GANs model.

Research Trends of Generative Adversarial Networks and Image Generation and Translation (GAN 적대적 생성 신경망과 이미지 생성 및 변환 기술 동향)

  • Jo, Y.J.;Bae, K.M.;Park, J.Y.
    • Electronics and Telecommunications Trends
    • /
    • v.35 no.4
    • /
    • pp.91-102
    • /
    • 2020
  • Recently, generative adversarial networks (GANs) is a field of research that has rapidly emerged wherein many studies conducted shows overwhelming results. Initially, this was at the level of imitating the training dataset. However, the GAN is currently useful in many fields, such as transformation of data categories, restoration of erased parts of images, copying facial expressions of humans, and creation of artworks depicting a dead painter's style. Although many outstanding research achievements have been attracting attention recently, GANs have encountered many challenges. First, they require a large memory facility for research. Second, there are still technical limitations in processing high-resolution images over 4K. Third, many GAN learning methods have a problem of instability in the training stage. However, recent research results show images that are difficult to distinguish whether they are real or fake, even with the naked eye, and the resolution of 4K and above is being developed. With the increase in image quality and resolution, many applications in the field of design and image and video editing are now available, including those that draw a photorealistic image as a simple sketch or easily modify unnecessary parts of an image or a video. In this paper, we discuss how GANs started, including the base architecture and latest technologies of GANs used in high-resolution, high-quality image creation, image and video editing, style translation, content transfer, and technology.

Study on hole-filling technique of motion capture images using GANs (Generative Adversarial Networks) (GANs(Generative Adversarial Networks)를 활용한 모션캡처 이미지의 hole-filling 기법 연구)

  • Shin, Kwang-Seong;Shin, Seong-Yoon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2019.05a
    • /
    • pp.160-161
    • /
    • 2019
  • As a method for modeling a three-dimensional object, there are a method using a 3D scanner, a method using a motion capture system, and a method using a Kinect system. Through this method, a portion that is not captured due to occlusion occurs in the process of creating a three-dimensional object. In order to implement a perfect three-dimensional object, it is necessary to arbitrarily fill the obscured part. There is a technique to fill the unexposed part by various image processing methods. In this study, we propose a method using GANs, which is the latest trend of unsupervised machine learning, as a method for more natural hole-filling.

  • PDF

Improving Fidelity of Synthesized Voices Generated by Using GANs (GAN으로 합성한 음성의 충실도 향상)

  • Back, Moon-Ki;Yoon, Seung-Won;Lee, Sang-Baek;Lee, Kyu-Chul
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.10 no.1
    • /
    • pp.9-18
    • /
    • 2021
  • Although Generative Adversarial Networks (GANs) have gained great popularity in computer vision and related fields, generating audio signals independently has yet to be presented. Unlike images, an audio signal is a sampled signal consisting of discrete samples, so it is not easy to learn the signals using CNN architectures, which is widely used in image generation tasks. In order to overcome this difficulty, GAN researchers proposed a strategy of applying time-frequency representations of audio to existing image-generating GANs. Following this strategy, we propose an improved method for increasing the fidelity of synthesized audio signals generated by using GANs. Our method is demonstrated on a public speech dataset, and evaluated by Fréchet Inception Distance (FID). When employing our method, the FID showed 10.504, but 11.973 as for the existing state of the art method (lower FID indicates better fidelity).

Text-to-Face Generation Using Multi-Scale Gradients Conditional Generative Adversarial Networks (다중 스케일 그라디언트 조건부 적대적 생성 신경망을 활용한 문장 기반 영상 생성 기법)

  • Bui, Nguyen P.;Le, Duc-Tai;Choo, Hyunseung
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2021.11a
    • /
    • pp.764-767
    • /
    • 2021
  • While Generative Adversarial Networks (GANs) have seen huge success in image synthesis tasks, synthesizing high-quality images from text descriptions is a challenging problem in computer vision. This paper proposes a method named Text-to-Face Generation Using Multi-Scale Gradients for Conditional Generative Adversarial Networks (T2F-MSGGANs) that combines GANs and a natural language processing model to create human faces has features found in the input text. The proposed method addresses two problems of GANs: model collapse and training instability by investigating how gradients at multiple scales can be used to generate high-resolution images. We show that T2F-MSGGANs converge stably and generate good-quality images.

Analysis of Soot Particle Morphology Using Rayleigh Debye Gans Scattering Theory (RDG 산란 이론을 이용한 그을음 탄소 입자의 형상 분석)

  • Seo, Hyoungseock;Kim, Kibum
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.17 no.9
    • /
    • pp.641-646
    • /
    • 2016
  • Soot particles generated by fossil fuel combustion normally have fractal morphology with aggregates consisting of small spherical particles. Thus, Rayleigh or Mie elastic light scattering theory is not feasible for determining the fractal properties of soot aggregates. This paper describes a detailed process for applying Rayleigh-Debye Gans (RDG) scattering theory to effectively extract the morphological properties of any nano-scale particles. The fractal geometry of soot aggregates produced from an isooctane diffusion flame was observed using ex situ transmission electron microscopy (TEM) after thermophoretic sampling. RDG scattering theory was then used to analyze their fractal morphology, and various properties were calculated, such as the diameter of individual soot particles, number density, and volume fraction. The results show indiscernible changes during the soot growth process, but a distinct decreasing trend was observed in the soot oxidation process. The fractal dimension of the soot aggregates was determined to be around 1.82, which is in good agreement with that produced for other types of fuel. Thus, it can be concluded that the value of the fractal dimension is independent of the fuel type.

Spine Computed Tomography to Magnetic Resonance Image Synthesis Using Generative Adversarial Networks : A Preliminary Study

  • Lee, Jung Hwan;Han, In Ho;Kim, Dong Hwan;Yu, Seunghan;Lee, In Sook;Song, You Seon;Joo, Seongsu;Jin, Cheng-Bin;Kim, Hakil
    • Journal of Korean Neurosurgical Society
    • /
    • v.63 no.3
    • /
    • pp.386-396
    • /
    • 2020
  • Objective : To generate synthetic spine magnetic resonance (MR) images from spine computed tomography (CT) using generative adversarial networks (GANs), as well as to determine the similarities between synthesized and real MR images. Methods : GANs were trained to transform spine CT image slices into spine magnetic resonance T2 weighted (MRT2) axial image slices by combining adversarial loss and voxel-wise loss. Experiments were performed using 280 pairs of lumbar spine CT scans and MRT2 images. The MRT2 images were then synthesized from 15 other spine CT scans. To evaluate whether the synthetic MR images were realistic, two radiologists, two spine surgeons, and two residents blindly classified the real and synthetic MRT2 images. Two experienced radiologists then evaluated the similarities between subdivisions of the real and synthetic MRT2 images. Quantitative analysis of the synthetic MRT2 images was performed using the mean absolute error (MAE) and peak signal-to-noise ratio (PSNR). Results : The mean overall similarity of the synthetic MRT2 images evaluated by radiologists was 80.2%. In the blind classification of the real MRT2 images, the failure rate ranged from 0% to 40%. The MAE value of each image ranged from 13.75 to 34.24 pixels (mean, 21.19 pixels), and the PSNR of each image ranged from 61.96 to 68.16 dB (mean, 64.92 dB). Conclusion : This was the first study to apply GANs to synthesize spine MR images from CT images. Despite the small dataset of 280 pairs, the synthetic MR images were relatively well implemented. Synthesis of medical images using GANs is a new paradigm of artificial intelligence application in medical imaging. We expect that synthesis of MR images from spine CT images using GANs will improve the diagnostic usefulness of CT. To better inform the clinical applications of this technique, further studies are needed involving a large dataset, a variety of pathologies, and other MR sequence of the lumbar spine.

Combining Conditional Generative Adversarial Network and Regression-based Calibration for Cloud Removal of Optical Imagery (광학 영상의 구름 제거를 위한 조건부 생성적 적대 신경망과 회귀 기반 보정의 결합)

  • Kwak, Geun-Ho;Park, Soyeon;Park, No-Wook
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.6_1
    • /
    • pp.1357-1369
    • /
    • 2022
  • Cloud removal is an essential image processing step for any task requiring time-series optical images, such as vegetation monitoring and change detection. This paper presents a two-stage cloud removal method that combines conditional generative adversarial networks (cGANs) with regression-based calibration to construct a cloud-free time-series optical image set. In the first stage, the cGANs generate initial prediction results using quantitative relationships between optical and synthetic aperture radar images. In the second stage, the relationships between the predicted results and the actual values in non-cloud areas are first quantified via random forest-based regression modeling and then used to calibrate the cGAN-based prediction results. The potential of the proposed method was evaluated from a cloud removal experiment using Sentinel-2 and COSMO-SkyMed images in the rice field cultivation area of Gimje. The cGAN model could effectively predict the reflectance values in the cloud-contaminated rice fields where severe changes in physical surface conditions happened. Moreover, the regression-based calibration in the second stage could improve the prediction accuracy, compared with a regression-based cloud removal method using a supplementary image that is temporally distant from the target image. These experimental results indicate that the proposed method can be effectively applied to restore cloud-contaminated areas when cloud-free optical images are unavailable for environmental monitoring.