• Title/Summary/Keyword: sub-pixel

Search Result 273, Processing Time 0.019 seconds

Image Dehazing Enhancement Algorithm Based on Mean Guided Filtering

  • Weimin Zhou
    • Journal of Information Processing Systems
    • /
    • v.19 no.4
    • /
    • pp.417-426
    • /
    • 2023
  • To improve the effect of image restoration and solve the image detail loss, an image dehazing enhancement algorithm based on mean guided filtering is proposed. The superpixel calculation method is used to pre-segment the original foggy image to obtain different sub-regions. The Ncut algorithm is used to segment the original image, and it outputs the segmented image until there is no more region merging in the image. By means of the mean-guided filtering method, the minimum value is selected as the value of the current pixel point in the local small block of the dark image, and the dark primary color image is obtained, and its transmittance is calculated to obtain the image edge detection result. According to the prior law of dark channel, a classic image dehazing enhancement model is established, and the model is combined with a median filter with low computational complexity to denoise the image in real time and maintain the jump of the mutation area to achieve image dehazing enhancement. The experimental results show that the image dehazing and enhancement effect of the proposed algorithm has obvious advantages, can retain a large amount of image detail information, and the values of information entropy, peak signal-to-noise ratio, and structural similarity are high. The research innovatively combines a variety of methods to achieve image dehazing and improve the quality effect. Through segmentation, filtering, denoising and other operations, the image quality is effectively improved, which provides an important reference for the improvement of image processing technology.

Super-Resolution Reconstruction of Humidity Fields based on Wasserstein Generative Adversarial Network with Gradient Penalty

  • Tao Li;Liang Wang;Lina Wang;Rui Han
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.5
    • /
    • pp.1141-1162
    • /
    • 2024
  • Humidity is an important parameter in meteorology and is closely related to weather, human health, and the environment. Due to the limitations of the number of observation stations and other factors, humidity data are often not as good as expected, so high-resolution humidity fields are of great interest and have been the object of desire in the research field and industry. This study presents a novel super-resolution algorithm for humidity fields based on the Wasserstein generative adversarial network(WGAN) framework, with the objective of enhancing the resolution of low-resolution humidity field information. WGAN is a more stable generative adversarial networks(GANs) with Wasserstein metric, and to make the training more stable and simple, the gradient cropping is replaced with gradient penalty, and the network feature representation is improved by sub-pixel convolution, residual block combined with convolutional block attention module(CBAM) and other techniques. We evaluate the proposed algorithm using ERA5 relative humidity data with an hourly resolution of 0.25°×0.25°. Experimental results demonstrate that our approach outperforms not only conventional interpolation techniques, but also the super-resolution generative adversarial network(SRGAN) algorithm.

A Frequency Domain DV-to-MPEG-2 Transcoding (DV에서 MPEG-2로의 주파수 영역 변환 부호화)

  • Kim, Do-Nyeon;Yun, Beom-Sik;Choe, Yun-Sik
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.38 no.2
    • /
    • pp.138-148
    • /
    • 2001
  • Digital Video (DV) coding standards for digital video cassette recorder are based mainly on DCT and variable length coding. DV has low hardware complexity but high compressed bit rate of about 26 Mb/s. Thus, it is necessary to encode video with low complex video coding at the studios and then transcode compressed video into MPEG-2 for video-on-demand system. Because these coding methods exploit DCT, transcoding in the DCT domain can reduce computational complexity by excluding duplicated procedures. In transcoding DV into MPEC-2 intra coding, multiplying matrix by transformed data is used for 4:1:1-to-4:2:2 chroma format conversion and the conversion from 2-4-8 to 8-8 DCT mode, and therefore enables parallel processing. Variance of sub block for MPEG-2 rate control is computed completely in the DCT domain. These are verified through experiments. We estimate motion hierarchically using DCT coefficients for transcoding into MPEG-2 inter coding. First, we estimate motion of a macro block (MB) only with 4 DC values of 4 sub blocks and then estimate motion with 16-point MB using IDCT of 2$\times$2 low frequencies in each sub block, and finish estimation at a sub pixel as the fifth step. ME with overlapped search range shows better PSNR performance than ME without overlapping.

  • PDF

A Design of AES-based WiBro Security Processor (AES 기반 와이브로 보안 프로세서 설계)

  • Kim, Jong-Hwan;Shin, Kyung-Wook
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.44 no.7 s.361
    • /
    • pp.71-80
    • /
    • 2007
  • This paper describes an efficient hardware design of WiBro security processor (WBSec) supporting for the security sub-layer of WiBro wireless internet system. The WBSec processor, which is based on AES (Advanced Encryption Standard) block cipher algorithm, performs data oncryption/decryption, authentication/integrity, and key encryption/decryption for packet data protection of wireless network. It carries out the modes of ECB, CTR, CBC, CCM and key wrap/unwrap with two AES cores working in parallel. In order to achieve an area-efficient implementation, two design techniques are considered; First, round transformation block within AES core is designed using a shared structure for encryption/decryption. Secondly, SubByte/InvSubByte blocks that require the largest hardware in AES core are implemented using field transformation technique. It results that the gate count of WBSec is reduced by about 25% compared with conventional LUT (Look-Up Table)-based design. The WBSec processor designed in Verilog-HDL has about 22,350 gates, and the estimated throughput is about 16-Mbps at key wrap mode and maximum 213-Mbps at CCM mode, thus it can be used for hardware design of WiBro security system.

The Development of Stereotest using Gabor Images in 3D Environment: An Explorative Study (3D 환경에서 가보 영상을 이용한 입체 시력 검사도구의 개발: 탐색적 연구)

  • Kham, Keetaek
    • Journal of Broadcast Engineering
    • /
    • v.20 no.6
    • /
    • pp.901-911
    • /
    • 2015
  • Many studies tried to develop a 3D display based stereo test as a substitute for the conventional stereo test. Although many 3D monitor based stereo tests have various advantages over the conventional stereo tests with printed stimuli, they have a crucial limitation in manipulating disparity. The least value for disparity manipulation is one pixel, which is too big to screen the normal stereo acuity in pc environment with a short viewing distance. In this explorative study, a Gabor image was employed as a test stimulus, because its position can be manipulated by changing its phase information, which allows sub-pixel manipulation for disparity. Instead of employing the methods of the conventional stereo tests where measurement was made only once for each of a wide range of disparity values, 10 replications were administrated for each of 6 disparity levels. The results from the test using Gabor images were compared with those using random dot stimulus because the latter stimulus was exactly the same as that of the conventional stereo test. The correlation coefficient between two tests was found to be moderate. After one month later, the whole test was repeated in the same settings. The correlation coefficient between test and retest results from Gabor images was found to be as high as that from a random dot stimulus, implying high test-retest reliability. These results suggest that a Gabor stimulus could be used as a test stimulus for the valid and reliable stereo test, even in the limited condition, such as 3D environment with a short viewing distance and a condition for evaluating the stereo acuity very precisely.

Image Contrast Enhancement Technique for Local Dimming Backlight of Small-sized Mobile Display (소형 모바일 디스플레이의 Local Dimming 백라이트를 위한 영상 컨트라스트 향상 기법)

  • Chung, Jin-Young;Yun, Ki-Bang;Kim, Ki-Doo
    • 전자공학회논문지 IE
    • /
    • v.46 no.4
    • /
    • pp.57-65
    • /
    • 2009
  • This paper presents the image contrast enhancement technique suitable for local dimming backlight of small-sized mobile display while achieving the reduction of the power consumption. In addition to the large-sized TFT-LCD, small-sized one has adopted LED for backlight. Since, conventionally, LED was mounted on the side edge of a display panel, global dimming method has been widely used. However, recently, new advanced method of local dimming by placing the LED to the backside of the display panel and it raised the necessity of sub-blocked processing after partitioning the target image. When the sub-blocked image has low brightness, the supply current of a backlight LED is reduced, which gives both enhancement of contrast ratio and power consumption reduction. In this paper, we propose simple and improved image enhancement algorithm suitable for the small-sized mobile display. After partitioning the input image by equal sized blocks and analyzing the pixel information in each block, we realize the primary contrast enhancement by independently processing the sub-blocks using the information such as histogram, mean, and standard deviation values of luminance(Y) component. And then resulting information is transferred to each backlight control unit for local dimming to realize the secondary contrast enhancement as well as reduction of power consumption.

Development of a Classification Method for Forest Vegetation on the Stand Level, Using KOMPSAT-3A Imagery and Land Coverage Map (KOMPSAT-3A 위성영상과 토지피복도를 활용한 산림식생의 임상 분류법 개발)

  • Song, Ji-Yong;Jeong, Jong-Chul;Lee, Peter Sang-Hoon
    • Korean Journal of Environment and Ecology
    • /
    • v.32 no.6
    • /
    • pp.686-697
    • /
    • 2018
  • Due to the advance in remote sensing technology, it has become easier to more frequently obtain high resolution imagery to detect delicate changes in an extensive area, particularly including forest which is not readily sub-classified. Time-series analysis on high resolution images requires to collect extensive amount of ground truth data. In this study, the potential of land coverage mapas ground truth data was tested in classifying high-resolution imagery. The study site was Wonju-si at Gangwon-do, South Korea, having a mix of urban and natural areas. KOMPSAT-3A imagery taken on March 2015 and land coverage map published in 2017 were used as source data. Two pixel-based classification algorithms, Support Vector Machine (SVM) and Random Forest (RF), were selected for the analysis. Forest only classification was compared with that of the whole study area except wetland. Confusion matrixes from the classification presented that overall accuracies for both the targets were higher in RF algorithm than in SVM. While the overall accuracy in the forest only analysis by RF algorithm was higher by 18.3% than SVM, in the case of the whole region analysis, the difference was relatively smaller by 5.5%. For the SVM algorithm, adding the Majority analysis process indicated a marginal improvement of about 1% than the normal SVM analysis. It was found that the RF algorithm was more effective to identify the broad-leaved forest within the forest, but for the other classes the SVM algorithm was more effective. As the two pixel-based classification algorithms were tested here, it is expected that future classification will improve the overall accuracy and the reliability by introducing a time-series analysis and an object-based algorithm. It is considered that this approach will contribute to improving a large-scale land planning by providing an effective land classification method on higher spatial and temporal scales.

Hough Transform Based Projecton Method for Target Tracking in Image Suquences (투사 및 허프 변환 방식에 의한 연속 영상상의 비행체 궤적 추적)

  • 최재호;곽훈성
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.19 no.11
    • /
    • pp.2094-2105
    • /
    • 1994
  • This paper contains a Hough transform based projection method derived from Radon transform for tracking dim unresolved(sub-pixel) moving targets that move along straight line parths across a time sequential image data. In contrast to several recently presented Hough transform methods using a compressed image referred to as the track map our proposed technique utilizing a set of projections taken along arbitrary orientations effectively increases the changes of target detection, and creates a robust track estimation environment by incorporating all the available knowledge obtained from the projections. Moreover, in order to quantitatively assess the estimation capability of the projection-based Hough transform algorithm, the analytical bounds on the Hough space parameter errors introduced by image space noise contamination are derived. The simulation yielded promising results of estimating the track parameters even under low signal to noise rations when our technique was tested against the time sequential sets of real infrared image data referred to as the HiCamps.

  • PDF

A 2-Dimensional Barcode Detection Algorithm based on Block Contrast and Projection (블록 명암대비와 프로젝션에 기반한 2차원 바코드 검출 알고리즘)

  • Choi, Young-Kyu
    • The KIPS Transactions:PartB
    • /
    • v.15B no.4
    • /
    • pp.259-268
    • /
    • 2008
  • In an effort to increase the data capacity of one-dimensional symbology, 2D barcodes have been proposed a decade ago. In this paper, we present an effective 2D barcode detection algorithm from gray-level images, especially for the handheld 2D barcode recognition system. To locate the symbol inside the image, a criteria based on the block contrast is adopted, and a gray-scale projection with sub-pixel operation is utilized to segment the symbol precisely from the region of interest(ROI). Finally, the segmented ROI is normalized using the inverse perspective transformation for the following decoding processes. We also introduce the post-processing steps for decoding the QR-code. The proposed method ensures high performances under various lighting/printing conditions and strong perspective deformations. Experiments shows that our method is very robust and efficient in detecting the code area for the various types of 2D barcodes in real time.

Image Enhancement Algorithm using Dynamic Range Optimization (다이나믹 레인지 최적화를 통한 영상 화질 개선 알고리즘)

  • Song, Ki Sun;Kim, Min Sub;Kang, Moon Gi
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.53 no.6
    • /
    • pp.101-109
    • /
    • 2016
  • The images captured by digital still cameras or mobile phones are not always satisfactory because the devices have limited dynamic ranges compared with that of the real world. To cope with the problems, tone mapping function based methods and retinex theory based methods are studied. However, these methods generate a halo artifact or limited enhancement of global and local contrasts. The proposed method estimates illumination information used for image enhancement by optimizing a dynamic range of input image. The estimated illumination information has smoothness characteristic where the luminance is flat and does not have where the luminance changes to prevent the halo artifact. Additionally, the estimated illumination information and surrounding pixel values are considered when the tone mapping function is applied to overcome the limitations of the conventional tone mapping function approach. Experimental results show that the proposed algorithm outperforms the conventional methods on objective and subjective criteria.