• Title/Summary/Keyword: video recognition

Search Result 696, Processing Time 0.026 seconds

A model to secure storage space for CCTV video files using YOLO v3

  • Seong-Ik, Kim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.1
    • /
    • pp.65-70
    • /
    • 2023
  • In this paper, we propose a CCTV storage space securing model using YOLO v3. CCTV is installed and operated in various parts of society for disasters, disasters and safety such as crime prevention, fire prevention, and monitoring, and the number of CCTV is increasing and the quality of the video quality is improving. Due to this, as the number and size of image files increase, it is difficult to cope with the existing storage space. In order to solve this problem, we propose a model that detects specific objects in CCTV images using YOLO v3 library and deletes unnecessary frames by saving only the corresponding frames, thereby securing storage space by reducing the size of the image file, and thereby Periodic images can be stored and managed. After applying the proposed model, it was confirmed that the average image file size was reduced by 94.9%, and it was confirmed that the storage period was increased by about 20 times compared to before the application of the proposed model.

The Design and Experiment of AI Device Communication System Equipped with 5G (5G를 탑재한 AI 디바이스 통신 시스템의 설계 및 실험)

  • Han Seongil;Lee Daesik;Han Jihwan;Moon Hhyunjin;Lim Changmin;Lee Sangku
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.19 no.2
    • /
    • pp.69-78
    • /
    • 2023
  • In this paper, IO+5G dedicated hardware is developed and an AI device communication system equipped with a 5G is designed and tested. The AI device communication system equipped with a 5G receives the collected real-time images and the information collected from the IoT sensor in real time is to analyze the information and generates the risk detection events in the AI processing board. The event generated in the AI processing board creates a 5G channel in the dedicated hardware equipped with IO+5G. The created 5G channel delivers event video to the control video server. The 5G based dongle network enables faster data collection and more precise data measurement compared to wireless LAN and 5G routers. As a result of the experiment in this paper, the average test result of the 5G dongle network is about 51% faster than the Wi-Fi average test result in downlink and about 40% faster in uplink. In addition, when comparing the test result with terms of the 5G rounter to be set to 80% upload and 20% download, the average test result is that the 5G dongle network is about 11.27% faster when downloading and about 17.93% faster when uploading. when comparing the test result with terms of the the router to be set to 60% upload and 40% download, the 5G dongle network is about 11.19% faster when downlinking and about 13.61% faster when uplinking. Therefore, in this paper it describes that the developed 5G dongle network can improve the results by collecting data and analyzing it faster than wireless LAN and 5G routers.

A multidisciplinary analysis of the main actor's conflict emotions in Animation film's Turning Point (장편 애니메이션 극적전환점에서 주인공의 갈등 정서에 대한 다학제적 분석)

  • Lee, Tae Rin;Kim, Jong Dae;Liu, Guoxu;Ingabire, Jesse;Kim, Jae Ho
    • Korea Science and Art Forum
    • /
    • v.34
    • /
    • pp.275-290
    • /
    • 2018
  • The study began with the recognition that the animations movie need objective and reasonable methods to classify conflicts in visual to analyze conflicts centering on narratives. Study the emotions of the hero in conflict. The purpose of the study is to analyze conflict intensity and emotion. The results and contents of the study are as follows. First, we found a Turning Point and suggested a conflict classification model (Conflict 6B Model). Second, Based on the conflict classification model, the conflict based shot DB was extracted. Third, I found strength and emotion in inner and super personal conflicts. Fourth, Experiments and tests of strength and emotion were conducted in internal and super personal conflicts. The results of this study are metadata extracted from the emotional research on conflict. It is expected to be applied to video indexing of conflicts.

Automatic Text Extraction from News Video using Morphology and Text Shape (형태학과 문자의 모양을 이용한 뉴스 비디오에서의 자동 문자 추출)

  • Jang, In-Young;Ko, Byoung-Chul;Kim, Kil-Cheon;Byun, Hye-Ran
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.4
    • /
    • pp.479-488
    • /
    • 2002
  • In recent years the amount of digital video used has risen dramatically to keep pace with the increasing use of the Internet and consequently an automated method is needed for indexing digital video databases. Textual information, both superimposed and embedded scene texts, appearing in a digital video can be a crucial clue for helping the video indexing. In this paper, a new method is presented to extract both superimposed and embedded scene texts in a freeze-frame of news video. The algorithm is summarized in the following three steps. For the first step, a color image is converted into a gray-level image and applies contrast stretching to enhance the contrast of the input image. Then, a modified local adaptive thresholding is applied to the contrast-stretched image. The second step is divided into three processes: eliminating text-like components by applying erosion, dilation, and (OpenClose+CloseOpen)/2 morphological operations, maintaining text components using (OpenClose+CloseOpen)/2 operation with a new Geo-correction method, and subtracting two result images for eliminating false-positive components further. In the third filtering step, the characteristics of each component such as the ratio of the number of pixels in each candidate component to the number of its boundary pixels and the ratio of the minor to the major axis of each bounding box are used. Acceptable results have been obtained using the proposed method on 300 news images with a recognition rate of 93.6%. Also, my method indicates a good performance on all the various kinds of images by adjusting the size of the structuring element.

Deep Learning Braille Block Recognition Method for Embedded Devices (임베디드 기기를 위한 딥러닝 점자블록 인식 방법)

  • Hee-jin Kim;Jae-hyuk Yoon;Soon-kak Kwon
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.28 no.4
    • /
    • pp.1-9
    • /
    • 2023
  • In this paper, we propose a method to recognize the braille blocks for embedded devices in real time through deep learning. First, a deep learning model for braille block recognition is trained on a high-performance computer, and the learning model is applied to a lightweight tool to apply to an embedded device. To recognize the walking information of the braille block, an algorithm is used to determine the path using the distance from the braille block in the image. After detecting braille blocks, bollards, and crosswalks through the YOLOv8 model in the video captured by the embedded device, the walking information is recognized through the braille block path discrimination algorithm. We apply the model lightweight tool to YOLOv8 to detect braille blocks in real time. The precision of YOLOv8 model weights is lowered from the existing 32 bits to 8 bits, and the model is optimized by applying the TensorRT optimization engine. As the result of comparing the lightweight model through the proposed method with the existing model, the path recognition accuracy is 99.05%, which is almost the same as the existing model, but the recognition speed is reduced by 59% compared to the existing model, processing about 15 frames per second.

A Study on the Performance of Music Retrieval Based on the Emotion Recognition (감정 인식을 통한 음악 검색 성능 분석)

  • Seo, Jin Soo
    • The Journal of the Acoustical Society of Korea
    • /
    • v.34 no.3
    • /
    • pp.247-255
    • /
    • 2015
  • This paper presents a study on the performance of the music search based on the automatically recognized music-emotion labels. As in the other media data, such as speech, image, and video, a song can evoke certain emotions to the listeners. When people look for songs to listen, the emotions, evoked by songs, could be important points to consider. However; very little study has been done on the performance of the music-emotion labels to the music search. In this paper, we utilize the three axes of human music perception (valence, activity, tension) and the five basic emotion labels (happiness, sadness, tenderness, anger, fear) in measuring music similarity for music search. Experiments were conducted on both genre and singer datasets. The search accuracy of the proposed emotion-based music search was up to 75 % of that of the conventional feature-based music search. By combining the proposed emotion-based method with the feature-based method, we achieved up to 14 % improvement of search accuracy.

Development of a Point Tracking System for Measuring Structural Deformations Using Commercial Video Cameras

  • Kim, Hong-Il;Kim, Ho-Young;Park, Hyun-Jin;Han, Jae-Hung;Kim, Jun-Bum;Kim, Do-Hyung;Han, Jeong-Ho
    • International Journal of Aeronautical and Space Sciences
    • /
    • v.10 no.2
    • /
    • pp.86-94
    • /
    • 2009
  • This paper deals with the creation of a new, low-cost point/position tracking system that can measure deformations in engineering structures with simple commercially widespread cameras. Though point tracking systems do exist today, such as Stereo Pattern Recognition (SPR) and Projection Moir$\acute{e}$ Interferometry (PMI) systems, they are far too costly to use to analyze small, simple structures because complex optical components such as large flashes, high-resolution cameras and data acquisition systems with several computers are required. We developed a point tracking system using commercial cameras. This system used IR LEDs and commercial IR CCD cameras to minimize the interference posed by other extraneous light sources. The main algorithm used for this system is an optical point tracking algorithm, which is composed of the point extraction algorithm and the point matching algorithm for 3-D motion estimation. a series of verification tests were performed. Then, the developed point tracking system was applied to measure deformations of an acrylic plate under a mechanical load. The measured deformations of the acrylic plate matched well with the numerical analysis results. The results indicate that the developed point tracking system is reliable enough to measure continuous deformed shapes of various engineering structures.

Virtual Dialog System Based on Multimedia Signal Processing for Smart Home Environments (멀티미디어 신호처리에 기초한 스마트홈 가상대화 시스템)

  • Kim, Sung-Ill;Oh, Se-Jin
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.15 no.2
    • /
    • pp.173-178
    • /
    • 2005
  • This paper focuses on the use of the virtual dialog system whose aim is to build more convenient living environments. In order to realize this, the main emphasis of the paper lies on the description of the multimedia signal processing on the basis of the technologies such as speech recognition, speech synthesis, video, or sensor signal processing. For essential modules of the dialog system, we incorporated the real-time speech recognizer based on HM-Net(Hidden Markov Network) as well as speech synthesis into the overall system. In addition, we adopted the real-time motion detector based on the changes of brightness in pixels, as well as the touch sensor that was used to start system. In experimental evaluation, the results showed that the proposed system was relatively easy to use for controlling electric appliances while sitting in a sofa, even though the performance of the system was not better than the simulation results owing to the noisy environments.

A Feature Point Recognition Ratio Improvement Method for Immersive Contents Using Deep Learning (딥 러닝을 이용한 실감형 콘텐츠 특징점 인식률 향상 방법)

  • Park, Byeongchan;Jang, Seyoung;Yoo, Injae;Lee, Jaechung;Kim, Seok-Yoon;Kim, Youngmo
    • Journal of IKEEE
    • /
    • v.24 no.2
    • /
    • pp.419-425
    • /
    • 2020
  • The market size of immersive 360-degree video contents, which are noted as one of the main technology of the fourth industry, increases every year. However, since most of the images are distributed through illegal distribution networks such as Torrent after the DRM gets lifted, the damage caused by illegal copying is also increasing. Although filtering technology is used as a technology to respond to these issues in 2D videos, most of those filtering technology has issues in that it has to overcome the technical limitation such as huge feature-point data volume and the related processing capacity due to ultra high resolution such as 4K UHD or higher in order to apply the existing technology to immersive 360° videos. To solve these problems, this paper proposes a feature-point recognition ratio improvement method for immersive 360-degree videos using deep learning technology.

Adaptive Counting Line Detection for Traffic Analysis in CCTV Videos (CCTV영상 내 교통량 분석을 위한 적응적 계수선 검출 방법)

  • Jung, Hyeonseok;Lim, Seokjae;Lee, Ryong;Park, Minwoo;Lee, Sang-Hwan;Kim, Wonjun
    • Journal of Broadcast Engineering
    • /
    • v.25 no.1
    • /
    • pp.48-57
    • /
    • 2020
  • Recently, with the rapid development of image recognition technology, the demand for object analysis in road CCTV videos is increasing. In this paper, we propose a method that can adaptively find the counting line for traffic analysis in road CCTV videos. First, vehicles on the road are detected, and the corresponding positions of the detected vehicles are modeled as the two-dimensional pointwise Gaussian map. The paths of vehicles are estimated by accumulating pointwise Gaussian maps on successive video frames. Then, we apply clustering and linear regression to the accumulated Gaussian map to find the principal direction of the road, which is highly relevant to the counting line. Experimental results show that the proposed method for detecting the counting line is effective in various situations.