• Title/Summary/Keyword: image decoder

Search Result 206, Processing Time 0.025 seconds

Demosaicing based Image Compression with Channel-wise Decoder

  • Indra Imanuel;Suk-Ho Lee
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.15 no.4
    • /
    • pp.74-83
    • /
    • 2023
  • In this paper, we propose an image compression scheme which uses a demosaicking network and a channel-wise decoder in the decoding network. For the demosaicing network, we use as the input a colored mosaiced pattern rather than the well-known Bayer pattern. The use of a colored mosaiced pattern results in the mosaiced image containing a greater amount of information pertaining to the original image. Therefore, it contributes to result in a better color reconstruction. The channel-wise decoder is composed of multiple decoders where each decoder is responsible for each channel in the color image, i.e., the R, G, and B channels. The encoder and decoder are both implemented by wavelet based auto-encoders for better performance. Experimental results verify that the separated channel-wise decoders and the colored mosaic pattern produce a better reconstructed color image than a single decoder. When combining the colored CFA with the multi-decoder, the PSNR metric exhibits an increase of over 2dB for three-times compression and approximately 0.6dB for twelve-times compression compared to the Bayer CFA with a single decoder. Therefore, the compression rate is also increased with the proposed method than with the method using a single decoder on the Bayer patterned mosaic image.

Deep Reference-based Dynamic Scene Deblurring

  • Cunzhe Liu;Zhen Hua;Jinjiang Li
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.3
    • /
    • pp.653-669
    • /
    • 2024
  • Dynamic scene deblurring is a complex computer vision problem owing to its difficulty to model mathematically. In this paper, we present a novel approach for image deblurring with the help of the sharp reference image, which utilizes the reference image for high-quality and high-frequency detail results. To better utilize the clear reference image, we develop an encoder-decoder network and two novel modules are designed to guide the network for better image restoration. The proposed Reference Extraction and Aggregation Module can effectively establish the correspondence between blurry image and reference image and explore the most relevant features for better blur removal and the proposed Spatial Feature Fusion Module enables the encoder to perceive blur information at different spatial scales. In the final, the multi-scale feature maps from the encoder and cascaded Reference Extraction and Aggregation Modules are integrated into the decoder for a global fusion and representation. Extensive quantitative and qualitative experimental results from the different benchmarks show the effectiveness of our proposed method.

Design and Implementation of JPEG Image Display Board Using FFGA (FPGA를 이용한 JPEG Image Display Board 설계 및 구현)

  • Kwon Byong-Heon;Seo Burm-Suk
    • Journal of Digital Contents Society
    • /
    • v.6 no.3
    • /
    • pp.169-174
    • /
    • 2005
  • In this paper we propose efficient design and implementation of JPEG image display board that can display JPEG image on TV. we used NAND Flash Memory to save the compressed JPEG bit stream and video encoder to display the decoded JPEG mage on TV. Also we convert YCbCr to RGB to super impose character on JPEG image. The designed B/D is implemented using FPGA.

  • PDF

SHD Digital Cinema Distribution over a Fast Long-Distance Network

  • Takahiro Yamaguchi;Daisuke Shirai;Mitsuru Nomura;Kazuhiro Shirakawa;Tatsuya Fujii;Tetsuro Fujii;Kim, io-Oguchi
    • Journal of Broadcast Engineering
    • /
    • v.9 no.2
    • /
    • pp.119-130
    • /
    • 2004
  • We have developed a prototype super-high-definition (SHD) digital cinema distribution system that can store, transmit, and display eight-million-pixel motion pictures that have the image quality of a 35-mm film movie. The system contains a movie server, a real-time decoder, and an SHB projector. Using a Gigabit Ethernet link and TCP/IP, the server transmits JPEG2000 compressed motion picture data streams to the decoder at transmission speeds as high as 300 Mbps. The received data streams are decompressed by the decoder, and then projected onto a screen via the projector. By using an enlarged TCP window, multiple TCP streams, and a shaping function to control the data transmission quantity, we achieved real-time streaming of SHD movie data at about 300 Mbps between Chicago and Los Angeles, a distance of more than 3000 km. We also improved the decoder performance to show movies with Image qualities of 450 Mbps or higher. Since UDP is more suitable than TCP for fast long-distance streaming, we have developed an SHD digital cinema UDP relay system, in which UDP is used for transmission over a fast long-distance network. By using four pairs of server-side-proxy and decoder-side-proxy, 450-Mbps movie data streams could be transmitted.

A 3-stage Pipelined Architecture for Multi-View Images Decoder3 (단계 파이프라인 구조를 갖는 Multi-View 영상 디코더)

  • Bae, Chang-Ho;Yang, Yeong-Yil
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.39 no.4
    • /
    • pp.104-111
    • /
    • 2002
  • In this paper, we proposed the architecture of the decoder which implements the multi-view images decoding algorithm. The study of the hardware structure of the multi-view image processing has not been accomplished. The proposed multi-view images decoder operates in a three stage pipelined manner and extracts the depth of the pixels of the decoded image every clock. The multi-view images decoder consists of three modules, Node selector which transfers the value of the nodes repeatedly and Depth Extractor which extracts the depth of each pixel from the four values of the nodes and Affine transformer which generates the projecting position on the image plane from the values of the pixels and the specified viewpoint. The proposed architecture is designed and simulated by the Max+plus II design tool and the operating frequency is 30MHz. The image can be constructed in a real time by the decoder with the proposed architecture.

The Object Distance Measurement using Stereo Image (스테레오 영상을 이용한 물체 거리 측정)

  • Jang, Dong-Young;Oh, Eun-Tack
    • Proceedings of the IEEK Conference
    • /
    • 2007.07a
    • /
    • pp.351-352
    • /
    • 2007
  • 본 논문에서는 인간의 시각 특성 중 특정 물체의 거리를 지각하는 메커니즘을 실시간 처리를 위해 하드웨어로 구현하였다. 본 논문에서 구현된 시스템은 크게 스테레오 영상 입 출력부와 영상을 처리하기 위한 IP 부로 구성된다. 입 출력부는 ALTERA 사의 Excalibur을 기반으로 하여 Image Decoder, UART, SDRAM, SRAM, TFT-LCD등으로 구성된 Image Board로 스테레오 영상을 받아 처리할 수 있게 설계하였다. IP부분은 Image Decoder 내부 레지스터를 설정하기 위한 I2C 버스 IP, 두 개의 Image Decoder를 통해 들어오는 스테레오 영상 입력 IP, 에러 보정을 위한 미디언 필터링 IP부, 에지 검출 IP, 거리를 검출하기 위한 스테레오 정합을 IP와 결과 영상을 보여주기 위한 TFT-LCD IP를 구현하였다.

  • PDF

Implementation of Encoder/Decoder to Support SNN Model in an IoT Integrated Development Environment based on Neuromorphic Architecture (뉴로모픽 구조 기반 IoT 통합 개발환경에서 SNN 모델을 지원하기 위한 인코더/디코더 구현)

  • Kim, Hoinam;Yun, Young-Sun
    • Journal of Software Assessment and Valuation
    • /
    • v.17 no.2
    • /
    • pp.47-57
    • /
    • 2021
  • Neuromorphic technology is proposed to complement the shortcomings of existing artificial intelligence technology by mimicking the human brain structure and computational process with hardware. NA-IDE has also been proposed for developing neuromorphic hardware-based IoT applications. To implement an SNN model in NA-IDE, commonly used input data must be transformed for use in the SNN model. In this paper, we implemented a neural coding method encoder component that converts image data into a spike train signal and uses it as an SNN input. The decoder component is implemented to convert the output back to image data when the SNN model generates a spike train signal. If the decoder component uses the same parameters as the encoding process, it can generate static data similar to the original data. It can be used in fields such as image-to-image and speech-to-speech to transform and regenerate input data using the proposed encoder and decoder.

Design of FRACTAL Image Compression Decoder (FRACTAL 영상 압축 Decoder 설계)

  • 김용배;박형근;임순자;김용환
    • Proceedings of the IEEK Conference
    • /
    • 1999.06a
    • /
    • pp.623-626
    • /
    • 1999
  • As the information society is advanced, the needs for mass information storage and retrieval grows. Digital image information is stored in retrieval systems, broadcasting in television transmission and exchanged over several kinds of telecommunication media. A major problem is that digital images are represented with large amount of data. The useful feature of image compression is that transmitting rapidly a lot of data in less time. Therefore we proposed a parallel Fractal trans-formation unit in Fractal Image compression system.

  • PDF

Transform Coding of Arbitrarily-Shaped Image Segments Using Recovery of Truncated Coefficients (삭제된 변환계수의 복원을 이용한 임의형태 영상영역 변환부호화)

  • 김희정;김지홍
    • Proceedings of the IEEK Conference
    • /
    • 2003.07e
    • /
    • pp.2351-2354
    • /
    • 2003
  • A new transform coder for arbitrarily shaped image segments is proposed. In the encoder, a block-based DCT is applied to the resulting image block after shifting pixels within the image segment to block border and padding the mean value of the pixels to empty region. For reducing the transmission bit rate, the transform coefficients located in padded region are truncated and only the remaining transform coefficients are transmitted to the decoder. In the decoder, the transform coefficients truncated in the encoder are recovered using received transform coefficients and a block-based inverse DCT is performed.

  • PDF

MEDU-Net+: a novel improved U-Net based on multi-scale encoder-decoder for medical image segmentation

  • Zhenzhen Yang;Xue Sun;Yongpeng, Yang;Xinyi Wu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.7
    • /
    • pp.1706-1725
    • /
    • 2024
  • The unique U-shaped structure of U-Net network makes it achieve good performance in image segmentation. This network is a lightweight network with a small number of parameters for small image segmentation datasets. However, when the medical image to be segmented contains a lot of detailed information, the segmentation results cannot fully meet the actual requirements. In order to achieve higher accuracy of medical image segmentation, a novel improved U-Net network architecture called multi-scale encoder-decoder U-Net+ (MEDU-Net+) is proposed in this paper. We design the GoogLeNet for achieving more information at the encoder of the proposed MEDU-Net+, and present the multi-scale feature extraction for fusing semantic information of different scales in the encoder and decoder. Meanwhile, we also introduce the layer-by-layer skip connection to connect the information of each layer, so that there is no need to encode the last layer and return the information. The proposed MEDU-Net+ divides the unknown depth network into each part of deconvolution layer to replace the direct connection of the encoder and decoder in U-Net. In addition, a new combined loss function is proposed to extract more edge information by combining the advantages of the generalized dice and the focal loss functions. Finally, we validate our proposed MEDU-Net+ MEDU-Net+ and other classic medical image segmentation networks on three medical image datasets. The experimental results show that our proposed MEDU-Net+ has prominent superior performance compared with other medical image segmentation networks.