• Title/Summary/Keyword: broadcast algorithm

Search Result 1,140, Processing Time 0.018 seconds

A Group Search-based Distributed Dynamic Channel Allocation Algorithm in Uplink Cellular System (상향링크 셀룰러 시스템에서 그룹 탐색 기반의 분산동적채널할당 방법)

  • Yoo, Doh-Kyoung;Kim, Dong-Hoi
    • Journal of Broadcast Engineering
    • /
    • v.15 no.3
    • /
    • pp.407-413
    • /
    • 2010
  • In DCA (Dynamic Channel Allocation) scheme of uplink cellular system appling a frequency reuse factor of one, when a new call requests a channel, the new call will be blocked if its SINR (Carrier to Noise and Interference Ratio) is less than the required SINR or there is no available channel. The additional channel allocation for the blocked new call can be performed with channel borrowing in the adjacent cells. The channel borrowing causes the CCI (Co-Channel Interference), thus the SINR of the existing calls is deteriorated and the channel reallocation for the existing calls is required. As a result, the channel borrowing leads to a complex calculation so that it is a NP-hard problem. Therefore, to overcome the problem, we propose a novel Group Search-based DCA scheme which decreases the number of the blocked new calls and then reduces the number of the channel reallocation by the channel borrowing for the blocked new calls. The proposed scheme searches the all channels in a group of the adjacent cells and home cell at the same time in order to minimizes the number of the blocked new calls. The simulation results show that proposed Group Search-based DCA scheme provides better new call blocking probability and system throughput than the existing Single Search-based DCA scheme which searches only the channels in home cell.

2D-to-3D Stereoscopic conversion: Depth estimation in monoscopic soccer videos (단일 시점 축구 비디오의 3차원 영상 변환을 위한 깊이지도 생성 방법)

  • Ko, Jae-Seung;Kim, Young-Woo;Jung, Young-Ju;Kim, Chang-Ick
    • Journal of Broadcast Engineering
    • /
    • v.13 no.4
    • /
    • pp.427-439
    • /
    • 2008
  • This paper proposes a novel method to convert monoscopic soccer videos to stereoscopic videos. Through the soccer video analysis process, we detect shot boundaries and classify soccer frames into long shot or non-long shot. In the long shot case, the depth mapis generated relying on the size of the extracted ground region. For the non-long shot case, the shot is further partitioned into three types by considering the number of ground blocks and skin blocks which is obtained by a simple skin-color detection method. Then three different depth assignment methods are applied to each non-long shot types: 1) Depth estimation by object region extraction, 2) Foreground estimation by using the skin block and depth value computation by Gaussian function, and 3)the depth map generation for shots not containing the skin blocks. This depth assignment is followed by stereoscopic image generation. Subjective evaluation comparing generated depth maps and corresponding stereoscopic images indicate that the proposed algorithm can yield the sense of depth from a single view images.

Optimized DSP Implementation of Audio Decoders for Digital Multimedia Broadcasting (디지털 방송용 오디오 디코더의 DSP 최적화 구현)

  • Park, Nam-In;Cho, Choong-Sang;Kim, Hong-Kook
    • Journal of Broadcast Engineering
    • /
    • v.13 no.4
    • /
    • pp.452-462
    • /
    • 2008
  • In this paper, we address issues associated with the real-time implementation of the MPEG-1/2 Layer-II (or MUSICAM) and MPEG-4 ER-BSAC decoders for Digital Multimedia Broadcasting (DMB) on TMS320C64x+ that is a fixed-point DSP processor with a clock speed of 330 MHz. To achieve the real-time requirement, they should be optimized in different steps as follows. First of all, a C-code level optimization is performed by sharing the memory, adjusting data types, and unrolling loops. Next, an algorithm level optimization is carried out such as the reconfiguration of bitstream reading, the modification of synthesis filtering, and the rearrangement of the window coefficients for synthesis filtering. In addition, the C-code of a synthesis filtering module of the MPEG-1/2 Layer-II decoder is rewritten by using the linear assembly programming technique. This is because the synthesis filtering module requires the most processing time among all processing modules of the decoder. In order to show how the real-time implementation works, we obtain the percentage of the processing time for decoding and calculate a RMS value between the decoded audio signals by the reference MPEG decoder and its DSP version implemented in this paper. As a result, it is shown that the percentages of the processing time for the MPEG-1/2 Layer-II and MPEG-4 ER-BSAC decoders occupy less than 3% and 11% of the DSP clock cycles, respectively, and the RMS values of the MPEG-1/2 Layer-II and MPEG-4 ER-BSAC decoders implemented in this paper all satisfy the criterion of -77.01 dB which is defined by the MPEG standards.

An Efficient Parallelization Implementation of PU-level ME for Fast HEVC Encoding (고속 HEVC 부호화를 위한 효율적인 PU레벨 움직임예측 병렬화 구현)

  • Park, Soobin;Choi, Kiho;Park, Sang-Hyo;Jang, Euee Seon
    • Journal of Broadcast Engineering
    • /
    • v.18 no.2
    • /
    • pp.178-184
    • /
    • 2013
  • In this paper, we propose an efficient parallelization technique of PU-level motion estimation (ME) in the next generation video coding standard, high efficiency video coding (HEVC) to reduce the time complexity of video encoding. It is difficult to encode video in real-time because ME has significant complexity (i.e., 80 percent at the encoder). In order to solve this problem, various techniques have been studied, and among them is the parallelization, which is carefully concerned in algorithm-level ME design. In this regard, merge estimation method using merge estimation region (MER) that enables ME to be designed in parallel has been proposed; but, parallel ME based on MER has still unconsidered problems to be implemented ideally in HEVC test model (HM). Therefore, we propose two strategies to implement stable parallel ME using MER in HM. Through experimental results, the excellence of our proposed methods is shown; the encoding time using the proposed method is reduced by 25.64 percent on average of that of HM which uses sequential ME.

Fast Decision Method of Geometric Partitioning Mode and Block Partitioning Mode using Hough Transform in VVC (허프 변환을 이용한 VVC의 기하학 분할 모드 및 블록 분할 고속 결정 방법)

  • Lee, Minhun;Park, Juntaek;Bang, Gun;Lim, Woong;Sim, Donggyu;Oh, Seoung-Jun
    • Journal of Broadcast Engineering
    • /
    • v.25 no.5
    • /
    • pp.698-708
    • /
    • 2020
  • VVC (Versatile Video Coding), which has been developing as a next generation video coding standard. Compared to HEVC (High Efficiency Video Coding), VVC is improved by about 34% in RA (Random Access) configuration and about 30% in LDB (Low-Delay B) configuration by adopting various techniques such as recursive block partitioning structure and GPM (Geometric Partitioning Mode). But the encoding complexity is increased by about 10x and 7x, respectively. In this paper, we propose a fast decision method of GPM mode and block partitioning using directionality of block to reduce encoding complexity of VVC. The proposed method is to apply the Hough transform to the current block to identify the directionality of the block, thereby determining the GPM mode and the specific block partitioning method to be skipped in the rate-distortion cost search process. As a result, compared to VTM8.0, the proposed method reduces about 31.01% and 29.84% encoding complexity for RA and LDB configuration with 2.48% and 2.69% BD-rate loss, respectively.

A Design for Extension Codec based on Legacy Codec (레거시 코덱 기반 확장 코덱 설계)

  • Young, Su Heo;Bang, Gun;Park, Gwang Hoon
    • Journal of Broadcast Engineering
    • /
    • v.20 no.4
    • /
    • pp.509-520
    • /
    • 2015
  • A design for the merge mode of three dimensional High Efficiency Video Coding (3D-HEVC) is proposed in this paper. The proposed design can reduce the implementation complexity by removing the duplicated modules of the HEVC. For the extension codec, the implementation complexity is as crucial as coding efficiency, meaning if possible, extension codec needs to be easily implemented through by reusing the design of the legacy codec as-is. However, the existing merging process of 3D-HEVC had been built-in integrated in the inside of the HEVC merging process. Thus the duplicated merging process of HEVC had to be fully re-implemented in the 3D-HEVC. Consequently the implementation complexity of the extension codec was very high. The proposed 3D-HEVC merge mode is divided into following two stages; the process to reuse the HEVC modules without any modification; and the reprocessing process for newly added and modified merging modules in 3D-HEVC. By applying the proposed method, the re-implemented HEVC modules, which accounted for 51.4% of 3D-HEVC merge mode confirmed through the operational analysis of algorithm, can be eliminated, while maintaining the same coding efficiency and computational complexity.

Vehicle Headlight and Taillight Recognition in Nighttime using Low-Exposure Camera and Wavelet-based Random Forest (저노출 카메라와 웨이블릿 기반 랜덤 포레스트를 이용한 야간 자동차 전조등 및 후미등 인식)

  • Heo, Duyoung;Kim, Sang Jun;Kwak, Choong Sub;Nam, Jae-Yeal;Ko, Byoung Chul
    • Journal of Broadcast Engineering
    • /
    • v.22 no.3
    • /
    • pp.282-294
    • /
    • 2017
  • In this paper, we propose a novel intelligent headlight control (IHC) system which is durable to various road lights and camera movement caused by vehicle driving. For detecting candidate light blobs, the region of interest (ROI) is decided as front ROI (FROI) and back ROI (BROI) by considering the camera geometry based on perspective range estimation model. Then, light blobs such as headlights, taillights of vehicles, reflection light as well as the surrounding road lighting are segmented using two different adaptive thresholding. From the number of segmented blobs, taillights are first detected using the redness checking and random forest classifier based on Haar-like feature. For the headlight and taillight classification, we use the random forest instead of popular support vector machine or convolutional neural networks for supporting fast learning and testing in real-life applications. Pairing is performed by using the predefined geometric rules, such as vertical coordinate similarity and association check between blobs. The proposed algorithm was successfully applied to various driving sequences in night-time, and the results show that the performance of the proposed algorithms is better than that of recent related works.

File System Support for Multimedia Streaming in Internet Home Appliances (인터넷 홈서버를 위한 스트리밍 전용 파일 시스템)

  • 박진연;송승호;진종현;원유집;박승민;김정기
    • Journal of Broadcast Engineering
    • /
    • v.6 no.3
    • /
    • pp.246-259
    • /
    • 2001
  • Due to recent rapid deployment of Internet streaming service and digital broadcasting service, the issue of how to efficiently support streaming workload in so called "Internet Home Appliance" receives prime interests from industry as well as academia. The underlying dilemma is that it may not be feasible to put cutting edge CPU, boards, disks and other peripherals into that type of device. The primary reason is its cost. Usually, Internet Home Appliances has its dedicated usage, e.g. Internet Radio, and thus it does not require high-end CPU nor high-end Va subsystem. The same reasoning applies to I/O subsystem. In Internet Home Appliances dedicated to handle compressed moving picture, it is not equipped with high end SCSI disk with fast rotational speed. Thus, it is mandatory to devise elaborate software algorithm to exploit the available hardware resources and maximize the efficiency of the system. This paper presents our experiences in the design and implementation of a new multimedia file system which can efficiently deliver the required disk bandwidth for a periodic I/O workload. We have implemented the file system on the Linux operating system, and examined itsperformance under streaming I/O workload. The results of the study show that the proposed file system exhibits superior performance than the Linux Ext2 file system under streaming I/O workload. The result of this work not only contribute to advance the state f art file system technology for multimedia streaming but also put forth the software which is readily available and can be deployed. deployed.

  • PDF

Performance Evaluation of Channel Estimation Scheme for ATSC 3.0 MIMO under Fixed Reception Environment (고정 수신 환경에서 ATSC 3.0 MIMO의 채널 추정 방법에 따른 성능 평가)

  • Kim, Hyeongseok;Yeom, Myeonggil;Kim, Jeongchang;Park, Sung-Ik;Jung, Hoiyoon;Hur, Namho
    • Journal of Broadcast Engineering
    • /
    • v.24 no.5
    • /
    • pp.879-891
    • /
    • 2019
  • This paper provides performance evaluations of various channel estimation schemes for Advanced Television Systems Committee (ATSC) 3.0 multiple-input multiple-output (MIMO) system under a fixed reception environment. ATSC 3.0 MIMO system can obtain high spectral efficiency and improved reception performance compared to conventional terrestrial broadcasting systems. The ATSC 3.0 MIMO defines Walsh-Hadamard and null pilot encoding algorithms and the amplitude and phase of MIMO pilots are different from those of single-input single-output pilots. At the receiver, linear and discrete Fourier transform (DFT)-based interpolations can be used for the channel estimation. This paper provides the various combinations of the interpolation schemes for channel estimation in time and frequency dimensions, and then analyzes the performance of the various combinations through the computer simulation. The results of computer simulation show that the combination of the linear interpolation in the time dimension and then DFT-based interpolation in the frequency dimension can obtain the best performance among the considered combinations.

Deep Learning-based SISR (Single Image Super Resolution) Method using RDB (Residual Dense Block) and Wavelet Prediction Network (RDB 및 웨이블릿 예측 네트워크 기반 단일 영상을 위한 심층 학습기반 초해상도 기법)

  • NGUYEN, HUU DUNG;Kim, Eung-Tae
    • Journal of Broadcast Engineering
    • /
    • v.24 no.5
    • /
    • pp.703-712
    • /
    • 2019
  • Single image Super-Resolution (SISR) aims to generate a visually pleasing high-resolution image from its degraded low-resolution measurement. In recent years, deep learning - based super - resolution methods have been actively researched and have shown more reliable and high performance. A typical method is WaveletSRNet, which restores high-resolution images through wavelet coefficient learning based on feature maps of images. However, there are two disadvantages in WaveletSRNet. One is a big processing time due to the complexity of the algorithm. The other is not to utilize feature maps efficiently when extracting input image's features. To improve this problems, we propose an efficient single image super resolution method, named RDB-WaveletSRNet. The proposed method uses the residual dense block to effectively extract low-resolution feature maps to improve single image super-resolution performance. We also adjust appropriated growth rates to solve complex computational problems. In addition, wavelet packet decomposition is used to obtain the wavelet coefficients according to the possibility of large scale ratio. In the experimental result on various images, we have proven that the proposed method has faster processing time and better image quality than the conventional methods. Experimental results have shown that the proposed method has better image quality by increasing 0.1813dB of PSNR and 1.17 times faster than the conventional method.