• Title/Summary/Keyword: computer image analysis

Search Result 1,460, Processing Time 0.03 seconds

Robust Action Recognition Using Multiple View Image Sequences (다중 시점 영상 시퀀스를 이용한 강인한 행동 인식)

  • Ahmad, Mohiuddin;Lee, Seong-Whan
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2006.10b
    • /
    • pp.509-514
    • /
    • 2006
  • Human action recognition is an active research area in computer vision. In this paper, we present a robust method for human action recognition by using combined information of human body shape and motion information with multiple views image sequence. The principal component analysis is used to extract the shape feature of human body and multiple block motion of the human body is used to extract the motion features of human. This combined information with multiple view sequences enhances the recognition of human action. We represent each action using a set of hidden Markov model and we model each action by multiple views. This characterizes the human action recognition from arbitrary view information. Several daily actions of elderly persons are modeled and tested by using this approach and they are correctly classified, which indicate the robustness of our method.

  • PDF

Scalable Big Data Pipeline for Video Stream Analytics Over Commodity Hardware

  • Ayub, Umer;Ahsan, Syed M.;Qureshi, Shavez M.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.4
    • /
    • pp.1146-1165
    • /
    • 2022
  • A huge amount of data in the form of videos and images is being produced owning to advancements in sensor technology. Use of low performance commodity hardware coupled with resource heavy image processing and analyzing approaches to infer and extract actionable insights from this data poses a bottleneck for timely decision making. Current approach of GPU assisted and cloud-based architecture video analysis techniques give significant performance gain, but its usage is constrained by financial considerations and extremely complex architecture level details. In this paper we propose a data pipeline system that uses open-source tools such as Apache Spark, Kafka and OpenCV running over commodity hardware for video stream processing and image processing in a distributed environment. Experimental results show that our proposed approach eliminates the need of GPU based hardware and cloud computing infrastructure to achieve efficient video steam processing for face detection with increased throughput, scalability and better performance.

Linear Regression-based 1D Invariant Image for Shadow Detection and Removal in Single Natural Image (단일 자연 영상에서 그림자 검출 및 제거를 위한 선형 회귀 기반의 1D 불변 영상)

  • Park, Ki-Hong
    • Journal of Digital Contents Society
    • /
    • v.19 no.9
    • /
    • pp.1787-1793
    • /
    • 2018
  • Shadow is a common phenomenon observed in natural scenes, but it has a negative influence on image analysis such as object recognition, feature detection and scene analysis. Therefore, the process of detecting and removing shadows included in digital images must be considered as a pre-processing process of image analysis. In this paper, the existing methods for acquiring 1D invariant images, one of the feature elements for detecting and removing shadows contained in a single natural image, are described, and a method for obtaining 1D invariant images based on linear regression has been proposed. The proposed method calculates the log of the band-ratio between each channel of the RGB color image, and obtains the grayscale image line by linear regression. The final 1D invariant images were obtained by projecting the log image of the band-ratio onto the estimated grayscale image line. Experimental results show that the proposed method has lower computational complexity than the existing projection method using entropy minimization, and shadow detection and removal based on 1D invariant images are performed effectively.

Efficient Content-Based Image Retrieval Method using Shape and Color feature (형태와 칼러성분을 이용한 효율적인 내용 기반의 이미지 검색 방법)

  • Youm, Sung-Ju;Kim, Woo-Saeng
    • The Transactions of the Korea Information Processing Society
    • /
    • v.3 no.4
    • /
    • pp.733-744
    • /
    • 1996
  • Content-based image retrieval(CBIR) is an image data retrieval methodology using characteristic values of image data those are generated by system automatically without any caption or text information. In this paper, we propose a content-based image data retrieval method using shape and color features of image data as characteristic values. For this, we present some image processing techniques used for feature extraction and indexing techniques based on trie and R tree for fast image data retrieval. In our approach, image query result is more reliable because both shape and color features are considered. Also, we how an image database which implemented according to our approaches and sample retrieval results which are selected by our system from 200 sample images, and an analysis about the result by considering the effect of characteristic values of shape and color.

  • PDF

Virtual Contamination Lane Image and Video Generation Method for the Performance Evaluation of the Lane Departure Warning System (차선 이탈 경고 시스템의 성능 검증을 위한 가상의 오염 차선 이미지 및 비디오 생성 방법)

  • Kwak, Jae-Ho;Kim, Whoi-Yul
    • Transactions of the Korean Society of Automotive Engineers
    • /
    • v.24 no.6
    • /
    • pp.627-634
    • /
    • 2016
  • In this paper, an augmented video generation method to evaluate the performance of lane departure warning system is proposed. In our system, the input is a video which have road scene with general clean lane, and the content of output video is the same but the lane is synthesized with contamination image. In order to synthesize the contamination lane image, two approaches were used. One is example-based image synthesis, and the other is background-based image synthesis. Example-based image synthesis is generated in the assumption of the situation that contamination is applied to the lane, and background-based image synthesis is for the situation that the lane is erased due to aging. In this paper, a new contamination pattern generation method using Gaussian function is also proposed in order to produce contamination with various shape and size. The contamination lane video can be generated by shifting synthesized image as lane movement amount obtained empirically. Our experiment showed that the similarity between the generated contamination lane image and real lane image is over 90 %. Futhermore, we can verify the reliability of the video generated from the proposed method through the analysis of the change of lane recognition rate. In other words, the recognition rate based on the video generated from the proposed method is very similar to that of the real contamination lane video.

Computer Vision Based Measurement, Error Analysis and Calibration (컴퓨터 시각(視覺)에 의거한 측정기술(測定技術) 및 측정오차(測定誤差)의 분석(分析)과 보정(補正))

  • Hwang, H.;Lee, C.H.
    • Journal of Biosystems Engineering
    • /
    • v.17 no.1
    • /
    • pp.65-78
    • /
    • 1992
  • When using a computer vision system for a measurement, the geometrically distorted input image usually restricts the site and size of the measuring window. A geometrically distorted image caused by the image sensing and processing hardware degrades the accuracy of the visual measurement and prohibits the arbitrary selection of the measuring scope. Therefore, an image calibration is inevitable to improve the measuring accuracy. A calibration process is usually done via four steps such as measurement, modeling, parameter estimation, and compensation. In this paper, the efficient error calibration technique of a geometrically distorted input image was developed using a neural network. After calibrating a unit pixel, the distorted image was compensated by training CMLAN(Cerebellar Model Linear Associator Network) without modeling the behavior of any system element. The input/output training pairs for the network was obtained by processing the image of the devised sampled pattern. The generalization property of the network successfully compensates the distortion errors of the untrained arbitrary pixel points on the image space. The error convergence of the trained network with respect to the network control parameters were also presented. The compensated image through the network was then post processed using a simple DDA(Digital Differential Analyzer) to avoid the pixel disconnectivity. The compensation effect was verified using known sized geometric primitives. A way to extract directly a real scaled geometric quantity of the object from the 8-directional chain coding was also devised and coded. Since the developed calibration algorithm does not require any knowledge of modeling system elements and estimating parameters, it can be applied simply to any image processing system. Furthermore, it efficiently enhances the measurement accuracy and allows the arbitrary sizing and locating of the measuring window. The applied and developed algorithms were coded as a menu driven way using MS-C language Ver. 6.0, PC VISION PLUS library functions, and VGA graphic functions.

  • PDF

An Analysis on the Properties of Features against Various Distortions in Deep Neural Networks

  • Kang, Jung Heum;Jeong, Hye Won;Choi, Chang Kyun;Ali, Muhammad Salman;Bae, Sung-Ho;Kim, Hui Yong
    • Journal of Broadcast Engineering
    • /
    • v.26 no.7
    • /
    • pp.868-876
    • /
    • 2021
  • Deploying deep neural network model training performs remarkable performance in the fields of Object detection and Instance segmentation. To train these models, features are first extracted from the input image using a backbone network. The extracted features can be reused by various tasks. Research has been actively conducted to serve various tasks by using these learned features. In this process, standardization discussions about encoding, decoding, and transmission methods are proceeding actively. In this scenario, it is necessary to analyze the response characteristics of features against various distortions that may occur in the data transmission or data compression process. In this paper, experiment was conducted to inject various distortions into the feature in the object recognition task. And analyze the mAP (mean Average Precision) metric between the predicted value output from the neural network and the target value as the intensity of various distortions was increased. Experiments have shown that features are more robust to distortion than images. And this points out that using the feature as transmission means can prevent the loss of information against the various distortions during data transmission and compression process.

A Study on RCS and Scattering Point Analysis Based on Measured Data for Maritime Ship (실측자료 기반 함정 RCS 측정 및 산란점 분석 연구)

  • Jung, Hoi-In;Park, Sang-Hong;Choi, Jae-Ho;Kim, Kyung-Tae
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.23 no.2
    • /
    • pp.97-105
    • /
    • 2020
  • In order to set up radar cross section(RCS) reduction factors for a target, the scattering point position of the target should be identified through inverse synthetic aperture radar(ISAR) image analysis. For this purpose, ISAR image focusing is important. Maritime ship is non-linear maneuvering in the sea, however, which blur the ISAR image. To solve this problem, translational and rotational motion compensation are essential to form focused ISAR image. In this paper, hourglass and ISAR image analysis are performed on the collected data in the sea instead of using the prediction software tool, which takes much time and cost to make computer-aided design(CAD) model of the ship.

Facial Expression Analysis System based on Image Feature Extraction (이미지 특징점 추출 기반 얼굴 표정 분석 시스템)

  • Jeon, Jin-Hwan;Song, Jeo;Lee, Sang-Moon
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2016.07a
    • /
    • pp.293-294
    • /
    • 2016
  • 스마트폰, 블랙박스, CCTV 등을 통해 다양하고 방대한 영상 데이터가 발생하고 있다. 그중에서 사람의 얼굴 영상을 통해 개인을 인식 및 식별하고 감정 상태를 분석하려는 다양한 연구가 진행되고 있다. 본 논문에서는 디지털영상처리 분야에서 널리 사용되고 있는 SIFT알고리즘을 이용하여, 얼굴영상에 대한 특징점을 추출하고 이를 기반으로 성별, 나이 및 기초적인 감정 상태를 분류할 수 있는 시스템을 제안한다.

  • PDF