• Title/Summary/Keyword: structural similarity index

Search Result 87, Processing Time 0.025 seconds

Image Denoising Based on Adaptive Fractional Order Anisotropic Diffusion

  • Yu, Jimin;Tan, Lijian;Zhou, Shangbo;Wang, Liping;Wang, Chaomei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.1
    • /
    • pp.436-450
    • /
    • 2017
  • Recently, the method based on fractional order partial differential equation has been used in image processing. Usually, the optional order of fractional differentiation is determined by a lot of experiments. In this paper, a denoising model is proposed based on adaptive fractional order anisotropic diffusion. In the proposed model, the complexity of the local image texture is reflected by the local variance, and the order of the fractional differentiation is determined adaptively. In the process of the adaptive fractional order model, the discrete Fourier transform is applied to compute the fractional order difference as well as the dynamic evolution process. Experimental results show that the peak signal-to-noise ratio (PSNR) and structural similarity index measurement (SSIM) of the proposed image denoising algorithm is better than that of other some algorithms. The proposed algorithm not only can keep the detailed image information and edge information, but also obtain a good visual effect.

An Efficient Frame-Level Rate Control Algorithm for High Efficiency Video Coding

  • Lin, Yubei;Zhang, Xingming;Xiao, Jianen;Su, Shengkai
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.4
    • /
    • pp.1877-1891
    • /
    • 2016
  • In video coding, the goal of rate control (RC) is not only to avoid the undesirable fluctuation in bit allocation, but also to provide a good visual perception. In this paper, a novel frame-level rate control algorithm for High Efficiency Video Coding (HEVC) is proposed. Firstly a model that reveals the relationship between bit per pixel (bpp), the bitrate of the intra frame and the bitrate of the subsequent inter frames in a group of pictures (GOP) is established, based on which the target bitrate of the first intra frame is well estimated. Then a novel frame-level bit allocation algorithm is developed, which provides a robust bit balancing scheme between the intra frame and the inter frames in a GOP to achieve the visual quality smoothness throughout the whole sequence. Our experimental results show that when compared to the RC scheme for HEVC encoder HM-16.0, the proposed algorithm can produce reconstructed frames with more consistent objective video quality. In addition, the objective visual quality of the reconstructed frames can be improved with less bitrate.

Generation of contrast enhanced computed tomography image using deep learning network

  • Woo, Sang-Keun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.3
    • /
    • pp.41-47
    • /
    • 2019
  • In this paper, we propose a application of conditional generative adversarial network (cGAN) for generation of contrast enhanced computed tomography (CT) image. Two types of CT data which were the enhanced and non-enhanced were used and applied by the histogram equalization for adjusting image intensities. In order to validate the generation of contrast enhanced CT data, the structural similarity index measurement (SSIM) was performed. Prepared generated contrast CT data were analyzed the statistical analysis using paired sample t-test. In order to apply the optimized algorithm for the lymph node cancer, they were calculated by short to long axis ratio (S/L) method. In the case of the model trained with CT data and their histogram equalized SSIM were $0.905{\pm}0.048$ and $0.908{\pm}0.047$. The tumor S/L of generated contrast enhanced CT data were validated similar to the ground truth when they were compared to scanned contrast enhanced CT data. It is expected that advantages of Generated contrast enhanced CT data based on deep learning are a cost-effective and less radiation exposure as well as further anatomical information with non-enhanced CT data.

Deep Learning Based Gray Image Generation from 3D LiDAR Reflection Intensity (딥러닝 기반 3차원 라이다의 반사율 세기 신호를 이용한 흑백 영상 생성 기법)

  • Kim, Hyun-Koo;Yoo, Kook-Yeol;Park, Ju H.;Jung, Ho-Youl
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.14 no.1
    • /
    • pp.1-9
    • /
    • 2019
  • In this paper, we propose a method of generating a 2D gray image from LiDAR 3D reflection intensity. The proposed method uses the Fully Convolutional Network (FCN) to generate the gray image from 2D reflection intensity which is projected from LiDAR 3D intensity. Both encoder and decoder of FCN are configured with several convolution blocks in the symmetric fashion. Each convolution block consists of a convolution layer with $3{\times}3$ filter, batch normalization layer and activation function. The performance of the proposed method architecture is empirically evaluated by varying depths of convolution blocks. The well-known KITTI data set for various scenarios is used for training and performance evaluation. The simulation results show that the proposed method produces the improvements of 8.56 dB in peak signal-to-noise ratio and 0.33 in structural similarity index measure compared with conventional interpolation methods such as inverse distance weighted and nearest neighbor. The proposed method can be possibly used as an assistance tool in the night-time driving system for autonomous vehicles.

Photorealistic Ray-traced Visualization Approach for the Interactive Biomimetic Design of Insect Compound Eyes

  • Nguyen, Tung Lam;Trung, Hieu Tran Doan;Lee, Wooseok;Lee, Hocheol
    • Current Optics and Photonics
    • /
    • v.5 no.6
    • /
    • pp.699-710
    • /
    • 2021
  • In this study, we propose a biomimetic optical structure design methodology for investigating micro-optical mechanisms associated with the compound eyes of insects. With these compound eyes, insects can respond fast while maintaining a wide field of view. Also, considerable research attention has been focused on the insect compound eyes to utilize these benefits. However, their nano micro-structures are complex and challenging to demonstrate in real applications. An effectively integrated design methodology is required considering the manufacturing difficulty. We show that photorealistic ray-traced visualization is an effective method for designing the biomimetic of a micro-compound eye of an insect. We analyze the image formation mechanism and create a three-dimensional computer-aided design model. Then, a ray-trace visualization is applied to observe the optical image formation. Finally, the segmented images are stitched together to generate an image with a wide-angle; the image is assessed for quality. The high structural similarity index (SSIM) value (approximately 0.84 to 0.89) of the stitched image proves that the proposed MATLAB-based image stitching algorithm performs effectively and comparably to the commercial software. The results may be employed for the understanding, researching, and design of advanced optical systems based on biological eyes and for other industrial applications.

Spatial Frequency Coverage and Image Reconstruction for Photonic Integrated Interferometric Imaging System

  • Zhang, Wang;Ma, Hongliu;Huang, Kang
    • Current Optics and Photonics
    • /
    • v.5 no.6
    • /
    • pp.606-616
    • /
    • 2021
  • A photonic integrated interferometric imaging system possesses the characteristics of small-scale, low weight, low power consumption, and better image quality. It has potential application for replacing conventional large space telescopes. In this paper, the principle of photonic integrated interferometric imaging is investigated. A novel lenslet array arrangement and lenslet pairing approach are proposed, which are helpful in improving spatial frequency coverage. For the novel lenslet array arrangement, two short interference arms were evenly distributed between two adjacent long interference arms. Each lenslet in the array would be paired twice through the novel lenslet pairing approach. Moreover, the image reconstruction model for optical interferometric imaging based on compressed sensing was established. Image simulation results show that the peak signal to noise ratio (PSNR) of the reconstructed image based on compressive sensing is about 10 dB higher than that of the direct restored image. Meanwhile, the normalized mean square error (NMSE) of the direct restored image is approximately 0.38 higher than that of the reconstructed image. Structural similarity index measure (SSIM) of the reconstructed image based on compressed sensing is about 0.33 higher than that of the direct restored image. The increased spatial frequency coverage and image reconstruction approach jointly contribute to better image quality of the photonic integrated interferometric imaging system.

Floop: An efficient video coding flow for unmanned aerial vehicles

  • Yu Su;Qianqian Cheng;Shuijie Wang;Jian Zhou;Yuhe Qiu
    • ETRI Journal
    • /
    • v.45 no.4
    • /
    • pp.615-626
    • /
    • 2023
  • Under limited transmission conditions, many factors affect the efficiency of video transmission. During the flight of an unmanned aerial vehicle (UAV), frequent network switching often occurs, and the channel transmission condition changes rapidly, resulting in low-video transmission efficiency. This paper presents an efficient video coding flow for UAVs working in the 5G nonstandalone network and proposes two bit controllers, including time and spatial bit controllers, in the flow. When the environment fluctuates significantly, the time bit controller adjusts the depth of the recursive codec to reduce the error propagation caused by excessive network inference. The spatial bit controller combines the spatial bit mask with the channel quality multiplier to adjust the bit allocation in space to allocate resources better and improve the efficiency of information carrying. In the spatial bit controller, a flexible mini graph is proposed to compute the channel quality multiplier. In this study, two bit controllers with end-to-end codec were combined, thereby constructing an efficient video coding flow. Many experiments have been performed in various environments. Concerning the multi-scale structural similarity index and peak signal-to-noise ratio, the performance of the coding flow is close to that of H.265 in the low bits per pixel area. With an increase in bits per pixel, the saturation bottleneck of the coding flow is at the same level as that of H.264.

Simultaneous Motion Recognition Framework using Data Augmentation based on Muscle Activation Model (근육 활성화 모델 기반의 데이터 증강을 활용한 동시 동작 인식 프레임워크)

  • Sejin Kim;Wan Kyun Chung
    • The Journal of Korea Robotics Society
    • /
    • v.19 no.2
    • /
    • pp.203-212
    • /
    • 2024
  • Simultaneous motion is essential in the activities of daily living (ADL). For motion intention recognition, surface electromyogram (sEMG) and corresponding motion label is necessary. However, this process is time-consuming and it may increase the burden of the user. Therefore, we propose a simultaneous motion recognition framework using data augmentation based on muscle activation model. The model consists of multiple point sources to be optimized while the number of point sources and their initial parameters are automatically determined. From the experimental results, it is shown that the framework has generated the data which are similar to the real one. This aspect is quantified with the following two metrics: structural similarity index measure (SSIM) and mean squared error (MSE). Furthermore, with k-nearest neighbor (k-NN) or support vector machine (SVM), the classification accuracy is also enhanced with the proposed framework. From these results, it can be concluded that the generalization property of the training data is enhanced and the classification accuracy is increased accordingly. We expect that this framework reduces the burden of the user from the excessive and time-consuming data acquisition.

A Divide-Conquer U-Net Based High-Quality Ultrasound Image Reconstruction Using Paired Dataset (짝지어진 데이터셋을 이용한 분할-정복 U-net 기반 고화질 초음파 영상 복원)

  • Minha Yoo;Chi Young Ahn
    • Journal of Biomedical Engineering Research
    • /
    • v.45 no.3
    • /
    • pp.118-127
    • /
    • 2024
  • Commonly deep learning methods for enhancing the quality of medical images use unpaired dataset due to the impracticality of acquiring paired dataset through commercial imaging system. In this paper, we propose a supervised learning method to enhance the quality of ultrasound images. The U-net model is designed by incorporating a divide-and-conquer approach that divides and processes an image into four parts to overcome data shortage and shorten the learning time. The proposed model is trained using paired dataset consisting of 828 pairs of low-quality and high-quality images with a resolution of 512x512 pixels obtained by varying the number of channels for the same subject. Out of a total of 828 pairs of images, 684 pairs are used as the training dataset, while the remaining 144 pairs served as the test dataset. In the test results, the average Mean Squared Error (MSE) was reduced from 87.6884 in the low-quality images to 45.5108 in the restored images. Additionally, the average Peak Signal-to-Noise Ratio (PSNR) was improved from 28.7550 to 31.8063, and the average Structural Similarity Index (SSIM) was increased from 0.4755 to 0.8511, demonstrating significant enhancements in image quality.

Human Visual System-aware Dimming Method Combining Pixel Compensation and Histogram Specification for TFT-LCDs

  • Jin, Jeong-Chan;Kim, Young-Jin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.12
    • /
    • pp.5998-6016
    • /
    • 2017
  • In thin-film transistor liquid-crystal displays (TFT-LCDs), which are most commonly used in mobile devices, the backlight accounts for about 70% of the power consumption. Therefore, most low-power-related studies focus on realizing power savings through backlight dimming. Image compensation is performed to mitigate the visual distortion caused by the backlight dimming. Therefore, popular techniques include pixel compensation for brightness recovery and contrast enhancement, such as histogram equalization. However, existing pixel compensation techniques often have limitations with respect to blur owing to the pixel saturation phenomenon, or because contrast enhancement cannot adequately satisfy the human visual system (HVS). To overcome these, in this study, we propose a novel dimming technique to achieve both power saving and HVS-awareness by combining the pixel compensation and histogram specifications, which convert the original cumulative density function (CDF) by designing and using the desired CDF of an image. Because the process of obtaining the desired CDF is customized to consider image characteristics, histogram specification is found to achieve better HVS-awareness than histogram equalization. For the experiments, we employ the LIVE image database, and we use the structural similarity (SSIM) index to measure the degree of visual satisfaction. The experimental results show that the proposed technique achieves up to 15.9% increase in the SSIM index compared with existing dimming techniques that use pixel compensation and histogram equalization in the case of the same low-power ratio. Further, the results indicate that it achieves improved HVS-awareness and increased power saving concurrently compared with previous techniques.