• Title/Summary/Keyword: Complex scene

Search Result 134, Processing Time 0.024 seconds

Abrupt Scene Change Detection Algorithm Using Macroblock Type and DC Coefficient in Compressed Domain (압축 도메인 상에서 메크로 블록 타입과 DC 계수를 사용한 급격한 장면 변화 검출 알고리즘)

  • 이흥렬;이웅희;이웅호;정동석
    • Proceedings of the IEEK Conference
    • /
    • 2003.07d
    • /
    • pp.1527-1530
    • /
    • 2003
  • Video is an important and challenge media and requires sophisticated indexing schemes for efficient retrieval from visual databases. Scene change detection is the first step for automatic indexing of video data. Recently, several scene change detection algorithms in the pixel and compressed domains have been reported in the literature. However, using pixel methods are computationally complex and are not very robust in detecting scene change detection. In this paper, we propose robust abrupt scene change detection using macroblock type and DC coefficient. Experimental results show that the proposed algorithm is robust for detection of most abrupt scene changes in the compressed domain.

  • PDF

Relation between Game Motivation and Preference to Cutscenes

  • Ruan, Xiao-Yin;Cho, Dong-Min
    • Cartoon and Animation Studies
    • /
    • s.36
    • /
    • pp.573-592
    • /
    • 2014
  • Following rapid development of software and hardware technologies and increasing enhancement in arithmetic capability, there are more and more content that can be accommodated and processed in video games, which is also increasingly complex and fine. Cutscene as a main narrative method have been developed, which have become necessary to express some key plots and important scenarios in games. Good cutscene can strengthen engagement of players with virtual world in games and make players share affection and sorrow with roles in games; while badly-designed cut-scene or overused cut-scene will impair immersion of players and affect players' gaming experience; for this reason, developers should not continue cut-scene design just from opinions of designers nor make players passive receivers, instead, they should reduce as much as possible interruption by cut-scene to players' immersion and grant players with better immersion. After all, only designs depending on demands and preferences of players by having some knowledge of impacts of cut-scene on players' immersion can be accepted by players.

Dynamic Scene Segmentation Algorithm Using a Cross Mask and Edge Information (Cross Mask와 에지 정보를 사용한 동영상 분할)

  • 강정숙;박래홍;이상욱
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.26 no.8
    • /
    • pp.1247-1256
    • /
    • 1989
  • In this paper, we propose the dynamic scene segmentation algorithm using a cross mask and edge information. This method, a combination of the conventioanl feature-based and pixel-based approaches, uses edges as features and determines moving pixels, with a cross mask centered on each edge pixel, by computing similarity measure between two consecutive image frames. With simple calcualtion the proposed method works well for image consisting of complex background or several moving objects. Also this method works satisfactorily in case of rotaitional motion.

  • PDF

A Segmentation Method for a Moving Object on A Static Complex Background Scene. (복잡한 배경에서 움직이는 물체의 영역분할에 관한 연구)

  • Park, Sang-Min;Kwon, Hui-Ung;Kim, Dong-Sung;Jeong, Kyu-Sik
    • The Transactions of the Korean Institute of Electrical Engineers A
    • /
    • v.48 no.3
    • /
    • pp.321-329
    • /
    • 1999
  • Moving Object segmentation extracts an interested moving object on a consecutive image frames, and has been used for factory automation, autonomous navigation, video surveillance, and VOP(Video Object Plane) detection in a MPEG-4 method. This paper proposes new segmentation method using difference images are calculated with three consecutive input image frames, and used to calculate both coarse object area(AI) and it's movement area(OI). An AI is extracted by removing background using background area projection(BAP). Missing parts in the AI is recovered with help of the OI. Boundary information of the OI confines missing parts of the object and gives inital curves for active contour optimization. The optimized contours in addition to the AI make the boundaries of the moving object. Experimental results of a fast moving object on a complex background scene are included.

  • PDF

Arabic Words Extraction and Character Recognition from Picturesque Image Macros with Enhanced VGG-16 based Model Functionality Using Neural Networks

  • Ayed Ahmad Hamdan Al-Radaideh;Mohd Shafry bin Mohd Rahim;Wad Ghaban;Majdi Bsoul;Shahid Kamal;Naveed Abbas
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.7
    • /
    • pp.1807-1822
    • /
    • 2023
  • Innovation and rapid increased functionality in user friendly smartphones has encouraged shutterbugs to have picturesque image macros while in work environment or during travel. Formal signboards are placed with marketing objectives and are enriched with text for attracting people. Extracting and recognition of the text from natural images is an emerging research issue and needs consideration. When compared to conventional optical character recognition (OCR), the complex background, implicit noise, lighting, and orientation of these scenic text photos make this problem more difficult. Arabic language text scene extraction and recognition adds a number of complications and difficulties. The method described in this paper uses a two-phase methodology to extract Arabic text and word boundaries awareness from scenic images with varying text orientations. The first stage uses a convolution autoencoder, and the second uses Arabic Character Segmentation (ACS), which is followed by traditional two-layer neural networks for recognition. This study presents the way that how can an Arabic training and synthetic dataset be created for exemplify the superimposed text in different scene images. For this purpose a dataset of size 10K of cropped images has been created in the detection phase wherein Arabic text was found and 127k Arabic character dataset for the recognition phase. The phase-1 labels were generated from an Arabic corpus of quotes and sentences, which consists of 15kquotes and sentences. This study ensures that Arabic Word Awareness Region Detection (AWARD) approach with high flexibility in identifying complex Arabic text scene images, such as texts that are arbitrarily oriented, curved, or deformed, is used to detect these texts. Our research after experimentations shows that the system has a 91.8% word segmentation accuracy and a 94.2% character recognition accuracy. We believe in the future that the researchers will excel in the field of image processing while treating text images to improve or reduce noise by processing scene images in any language by enhancing the functionality of VGG-16 based model using Neural Networks.

Text Region Extraction of Natural Scene Images using Gray-level Information and Split/Merge Method (명도 정보와 분할/합병 방법을 이용한 자연 영상에서의 텍스트 영역 추출)

  • Kim Ji-Soo;Kim Soo-Hyung;Choi Yeong-Woo
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.6
    • /
    • pp.502-511
    • /
    • 2005
  • In this paper, we propose a hybrid analysis method(HAM) based on gray-intensity information from natural scene images. The HAM is composed of GIA(Gray-intensity Information Analysis) and SMA(Split/Merge Analysis). Our experimental results show that the proposed approach is superior to conventional methods both in simple and complex images.

Real Scene Text Image Super-Resolution Based on Multi-Scale and Attention Fusion

  • Xinhua Lu;Haihai Wei;Li Ma;Qingji Xue;Yonghui Fu
    • Journal of Information Processing Systems
    • /
    • v.19 no.4
    • /
    • pp.427-438
    • /
    • 2023
  • Plenty of works have indicated that single image super-resolution (SISR) models relying on synthetic datasets are difficult to be applied to real scene text image super-resolution (STISR) for its more complex degradation. The up-to-date dataset for realistic STISR is called TextZoom, while the current methods trained on this dataset have not considered the effect of multi-scale features of text images. In this paper, a multi-scale and attention fusion model for realistic STISR is proposed. The multi-scale learning mechanism is introduced to acquire sophisticated feature representations of text images; The spatial and channel attentions are introduced to capture the local information and inter-channel interaction information of text images; At last, this paper designs a multi-scale residual attention module by skillfully fusing multi-scale learning and attention mechanisms. The experiments on TextZoom demonstrate that the model proposed increases scene text recognition's (ASTER) average recognition accuracy by 1.2% compared to text super-resolution network.

Adaptive Face Mask Detection System based on Scene Complexity Analysis

  • Kang, Jaeyong;Gwak, Jeonghwan
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.5
    • /
    • pp.1-8
    • /
    • 2021
  • Coronavirus disease 2019 (COVID-19) has affected the world seriously. Every person is required for wearing a mask properly in a public area to prevent spreading the virus. However, many people are not wearing a mask properly. In this paper, we propose an efficient mask detection system. In our proposed system, we first detect the faces of input images using YOLOv5 and classify them as the one of three scene complexity classes (Simple, Moderate, and Complex) based on the number of detected faces. After that, the image is fed into the Faster-RCNN with the one of three ResNet (ResNet-18, 50, and 101) as backbone network depending on the scene complexity for detecting the face area and identifying whether the person is wearing the mask properly or not. We evaluated our proposed system using public mask detection datasets. The results show that our proposed system outperforms other models.

Extraction of Text Alignment by Tensor Voting and its Application to Text Detection (텐서보팅을 이용한 텍스트 배열정보의 획득과 이를 이용한 텍스트 검출)

  • Lee, Guee-Sang;Dinh, Toan Nguyen;Park, Jong-Hyun
    • Journal of KIISE:Software and Applications
    • /
    • v.36 no.11
    • /
    • pp.912-919
    • /
    • 2009
  • A novel algorithm using 2D tensor voting and edge-based approach is proposed for text detection in natural scene images. The tensor voting is used based on the fact that characters in a text line are usually close together on a smooth curve and therefore the tokens corresponding to centers of these characters have high curve saliency values. First, a suitable edge-based method is used to find all possible text regions. Since the false positive rate of text detection result generated from the edge-based method is high, 2D tensor voting is applied to remove false positives and find only text regions. The experimental results show that our method successfully detects text regions in many complex natural scene images.

이동로봇주행을 위한 영상처리 기술

  • 허경식;김동수
    • The Magazine of the IEIE
    • /
    • v.23 no.12
    • /
    • pp.115-125
    • /
    • 1996
  • This paper presents a new algorithm for the self-localization of a mobile robot using one degree perspective Invariant(Cross Ratio). Most of conventional model-based self-localization methods have some problems that data structure building, map updating and matching processes are very complex. Use of a simple cross ratio can be effective to the above problems. The algorithm is based on two basic assumptions that the ground plane is flat and two locally parallel sloe-lines are available. Also it is assumed that an environmental map is available for matching between the scene and the model. To extract an accurate steering angle for a mobile robot, we take advantage of geometric features such as vanishing points. Feature points for cross ratio are extracted robustly using a vanishing point and intersection points between two locally parallel side-lines and vertical lines. Also the local position estimation problem has been treated when feature points exist less than 4points in the viewed scene. The robustness and feasibility of our algorithms have been demonstrated through real world experiments In Indoor environments using an indoor mobile robot, KASIRI-II(KAist Simple Roving Intelligence).

  • PDF