• Title/Summary/Keyword: Image Signal Recognition

Search Result 185, Processing Time 0.027 seconds

Automatic Measurement Method of Traffic Signs Using Image Recognition and Photogrammetry Technology (영상인식과 사진측량 기술을 이용한 교통표지 자동측정 방법)

  • Chang, Sang Kyu;Kim, Jin Soo
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.21 no.3
    • /
    • pp.19-25
    • /
    • 2013
  • Recently, more accurate database information of facilities is being required, with the increase in importance of urban road facility management. Therefore, this study proposed how to automatically detect particular traffic signs necessary for efficient construction of road facility DB. For this study, central locations of facilities were searched, after recognition and automatic detection of particular traffic signs through an image. Then, coordinate values of traffic signs calculated in the study were compared with real coordinate values, in order to evaluate the accuracy of traffic sign locations which were finally detected. Computer vision technology was used in recognizing and detecting traffic signs through OPEN CV-based coding, and photogrammetry was used in calculating accurate locations of detected traffic signs. For the experiment, circular road signal(No Parking) and triangular road signal(Crosswalk) were chosen out of various kinds of road signals. The research result showed that the circular road signal had a nearly 50cm error value, and the triangular road signal had a nearly 60cm error value, when comparing the calculated coordinates with the real coordinates. Though this result is not satisfactory, it is considered that there would be no problem to find locations of traffic signs.

Gendered innovation for algorithm through case studies (음성·영상 신호 처리 알고리즘 사례를 통해 본 젠더혁신의 필요성)

  • Lee, JiYeoun;Lee, Heisook
    • Journal of Digital Convergence
    • /
    • v.16 no.12
    • /
    • pp.459-466
    • /
    • 2018
  • Gendered innovations is a term used by policy makers and academics to refer the process of creating better research and development (R&D) for both men and women. In this paper, we analyze the literatures in image and speech signal processing that can be used in ICT, examine the importance of gendered innovations through case study. Therefore the latest domestic and foreign literature related to image and speech signal processing based on gender research is searched and a total of 9 papers are selected. In terms of gender analysis, research subjects, research environment, and research design are examined separately. Especially, through the case analysis of algorithms of the elderly voice signal processing, machine learning, machine translation technology, and facial gender recognition technology, we found that there is gender bias in existing algorithms, and which leads to gender analysis is required. We also propose a gendered innovations method integrating sex and gender analysis in algorithm development. Gendered innovations in ICT can contribute to the creation of new markets by developing products and services that reflect the needs of both men and women.

Development of Data Fusion Human Identification System Based on Finger-Vein Pattern-Matching Method and photoplethysmography Identification

  • Ko, Kuk Won;Lee, Jiyeon;Moon, Hongsuk;Lee, Sangjoon
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.7 no.2
    • /
    • pp.149-154
    • /
    • 2015
  • Biometric techniques for authentication using body parts such as a fingerprint, face, iris, voice, finger-vein and also photoplethysmography have become increasingly important in the personal security field, including door access control, finance security, electronic passport, and mobile device. Finger-vein images are now used to human identification, however, difficulties in recognizing finger-vein images are caused by capturing under various conditions, such as different temperatures and illumination, and noise in the acquisition camera. The human photoplethysmography is also important signal for human identification. In this paper To increase the recognition rate, we develop camera based identification method by combining finger vein image and photoplethysmography signal. We use a compact CMOS camera with a penetrating infrared LED light source to acquire images of finger vein and photoplethysmography signal. In addition, we suggest a simple pattern matching method to reduce the calculation time for embedded environments. The experimental results show that our simple system has good results in terms of speed and accuracy for personal identification compared to the result of only finger vein images.

Deep learning of sweep signal for damage detection on the surface of concrete

  • Gao Shanga;Jun Chen
    • Computers and Concrete
    • /
    • v.32 no.5
    • /
    • pp.475-486
    • /
    • 2023
  • Nondestructive evaluation (NDE) is an important task of civil engineering structure monitoring and inspection, but minor damage such as small cracks in local structure is difficult to observe. If cracks continued expansion may cause partial or even overall damage to the structure. Therefore, monitoring and detecting the structure in the early stage of crack propagation is important. The crack detection technology based on machine vision has been widely studied, but there are still some problems such as bad recognition effect for small cracks. In this paper, we proposed a deep learning method based on sweep signals to evaluate concrete surface crack with a width less than 1 mm. Two convolutional neural networks (CNNs) are used to analyze the one-dimensional (1D) frequency sweep signal and the two-dimensional (2D) time-frequency image, respectively, and the probability value of average damage (ADPV) is proposed to evaluate the minor damage of structural. Finally, we use the standard deviation of energy ratio change (ERVSD) and infrared thermography (IRT) to compare with ADPV to verify the effectiveness of the method proposed in this paper. The experiment results show that the method proposed in this paper can effectively predict whether the concrete surface is damaged and the severity of damage.

A Study on the Measurement of Respiratory Rate Using Image Alignment and Statistical Pattern Classification (영상 정합 및 통계학적 패턴 분류를 이용한 호흡률 측정에 관한 연구)

  • Moon, Sujin;Lee, Eui Chul
    • Asia-pacific Journal of Multimedia Services Convergent with Art, Humanities, and Sociology
    • /
    • v.8 no.10
    • /
    • pp.63-70
    • /
    • 2018
  • Biomedical signal measurement technology using images has been developed, and researches on respiration signal measurement technology for maintaining life have been continuously carried out. The existing technology measured respiratory signals through a thermal imaging camera that measures heat emitted from a person's body. In addition, research was conducted to measure respiration rate by analyzing human chest movement in real time. However, the image processing using the infrared thermal image may be difficult to detect the respiratory organ due to the external environmental factors (temperature change, noise, etc.), and thus the accuracy of the measurement of the respiration rate is low.In this study, the images were acquired using visible light and infrared thermal camera to enhance the area of the respiratory tract. Then, based on the two images, features of the respiratory tract region are extracted through processes such as face recognition and image matching. The pattern of the respiratory signal is classified through the k-nearest neighbor classifier, which is one of the statistical classification methods. The respiration rate was calculated according to the characteristics of the classified patterns and the possibility of breathing rate measurement was verified by analyzing the measured respiration rate with the actual respiration rate.

An Implementation of Automatic Transmission System of Traffic Event Information (교통이벤트 정보의 자동 전송시스템 구현)

  • Jeong, Yeong-Rae;Jang, Jae-Hoon;Kang, Seog Geun
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.13 no.5
    • /
    • pp.987-994
    • /
    • 2018
  • In this paper, an automatic transmission system of traffic information is presented. Here, a traffic event is defined as an obstacle to an emergency vehicle such as an ambulance or a fire truck. When a traffic event is detected from a video recorded by a black box installed in a vehicle, the implemented system automatically transmits a proof image and corresponding information to the control center through an e-mail. For this purpose, we realize an algorithm of identifying the numbers and a character from the license plate, and an algorithm for determining the occurrence of a traffic event. To report the event, a function for automatic transmission of the text and image files through e-mail and file transfer protocol (FTP) is also appended. Therefore, if the traffic event is extended and applied to the presented system, it will be possible to establish a convenient reporting system for the violation of various traffic regulations. In addition, it will contribute to significantly reduce the number of traffic violations against the regulations.

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.

Study of Apparent Diffusion Coefficient Changes According to Spinal Disease in MR Diffusion-weighted Image

  • Heo, Yeong-Cheol;Cho, Jae-Hwan
    • Journal of Magnetics
    • /
    • v.22 no.1
    • /
    • pp.146-149
    • /
    • 2017
  • In this study, we compared the standardized value of each signal intensity, the apparent diffusion coefficient (ADC) that digitizes the diffusion of water molecules, and the signal to noise ratio (SNR) using b value 0 400, 1400 ($s/mm^2$). From March 2013 to December 2013, patients with suspicion of simple compound fracture and metastatic spine cancer were included in the MR readout. We used a 1.5 Tesla Achieva MRI system and a Syn-Spine Coil. Sequence is a DWI SE-EPI sagittal (diffusion weighted imaging spin echo-echo planar imaging sagittal) image with b-factor ($s/mm^2$) 0, 400, 1400 were used. Data analysis showed ROI (Region of Interest) in diseased area with high SI (signal intensity) in diffusion-weighted image b value 0 ($s/mm^2$) Using the MRIcro program, each SI was calculated with images of b-value 0, 400, and 1400 ($s/mm^2$), ADC map was obtained using Metlab Software with each image of b-value, The ADC is obtained by applying the ROI to the same position. The standardized values ($SI_{400}/SI_0$, $SI_{400}/SI_0$) of simple compression fractures were $0.47{\pm}0.04$ and $0.23{\pm}0.03$ and the standardized values ($SI_{400}/SI_0$, $SI_{400}/SI_0$) of the metastatic spine were $0.57{\pm}0.07$ and $0.32{\pm}0.08$ And the standardized values of the two diseases were statistically significant (p < 0.05). The ADC ($mm^2/s$) for b value 400 ($s/mm^2$) and 1400 ($s/mm^2$) of the simple compression fracture disease site were $1.70{\pm}0.16$ and $0.93{\pm}0.28$ and $1.24{\pm}0.21$ and $0.80{\pm}0.15$ for the metastatic spine. The ADC ($mm^2/s$) for b value 400($s/mm^2$) was statistically significant (p < 0.05) but the ADC ($mm^2/s$) for b value 1400 (p > 0.05). In conclusion, multi - b value recognition of signal changes in diffusion - weighted imaging is very important for the diagnosis of various spinal diseases.

A Pilot Study on Outpainting-powered Pet Pose Estimation (아웃페인팅 기반 반려동물 자세 추정에 관한 예비 연구)

  • Gyubin Lee;Youngchan Lee;Wonsang You
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.24 no.1
    • /
    • pp.69-75
    • /
    • 2023
  • In recent years, there has been a growing interest in deep learning-based animal pose estimation, especially in the areas of animal behavior analysis and healthcare. However, existing animal pose estimation techniques do not perform well when body parts are occluded or not present. In particular, the occlusion of dog tail or ear might lead to a significant degradation of performance in pet behavior and emotion recognition. In this paper, to solve this intractable problem, we propose a simple yet novel framework for pet pose estimation where pet pose is predicted on an outpainted image where some body parts hidden outside the input image are reconstructed by the image inpainting network preceding the pose estimation network, and we performed a preliminary study to test the feasibility of the proposed approach. We assessed CE-GAN and BAT-Fill for image outpainting, and evaluated SimpleBaseline for pet pose estimation. Our experimental results show that pet pose estimation on outpainted images generated using BAT-Fill outperforms the existing methods of pose estimation on outpainting-less input image.

Machine Learning based Traffic Light Detection and Recognition Algorithm using Shape Information (기계학습 기반의 신호등 검출과 형태적 정보를 이용한 인식 알고리즘)

  • Kim, Jung-Hwan;Kim, Sun-Kyu;Lee, Tae-Min;Lim, Yong-Jin;Lim, Joonhong
    • Journal of IKEEE
    • /
    • v.22 no.1
    • /
    • pp.46-52
    • /
    • 2018
  • The problem of traffic light detection and recognition has recently become one of the most important topics in various researches on autonomous driving. Most algorithms are based on colors to detect and recognize traffic light signals. These methods have disadvantage in that the recognition rate is lowered due to the change of the color of the traffic light, the influence of the angle, distance, and surrounding illumination environment of the image. In this paper, we propose machine learning based detection and recognition algorithm using shape information to solve these problems. Unlike the existing algorithms, the proposed algorithm detects and recognizes the traffic signals based on the morphological characteristics of the traffic lights, which is advantageous in that it is robust against the influence from the surrounding environments. Experimental results show that the recognition rate of the signal is higher than those of other color-based algorithms.