• 제목/요약/키워드: Video Object Detection

검색결과 354건 처리시간 0.022초

YOLOv5 based Anomaly Detection for Subway Safety Management Using Dilated Convolution

  • Nusrat Jahan Tahira;Ju-Ryong Park;Seung-Jin Lim;Jang-Sik Park
    • 한국산업융합학회 논문집
    • /
    • 제26권2_1호
    • /
    • pp.217-223
    • /
    • 2023
  • With the rapid advancement of technologies, need for different research fields where this technology can be used is also increasing. One of the most researched topic in computer vision is object detection, which has widely been implemented in various fields which include healthcare, video surveillance and education. The main goal of object detection is to identify and categorize all the objects in a target environment. Specifically, methods of object detection consist of a variety of significant techniq ues, such as image processing and patterns recognition. Anomaly detection is a part of object detection, anomalies can be found various scenarios for example crowded places such as subway stations. An abnormal event can be assumed as a variation from the conventional scene. Since the abnormal event does not occur frequently, the distribution of normal and abnormal events is thoroughly imbalanced. In terms of public safety, abnormal events should be avoided and therefore immediate action need to be taken. When abnormal events occur in certain places, real time detection is required to prevent and protect the safety of the people. To solve the above problems, we propose a modified YOLOv5 object detection algorithm by implementing dilated convolutional layers which achieved 97% mAP50 compared to other five different models of YOLOv5. In addition to this, we also created a simple mobile application to avail the abnormal event detection on mobile phones.

Deep Learning Object Detection to Clearly Differentiate Between Pedestrians and Motorcycles in Tunnel Environment Using YOLOv3 and Kernelized Correlation Filters

  • Mun, Sungchul;Nguyen, Manh Dung;Kweon, Seokkyu;Bae, Young Hoon
    • 방송공학회논문지
    • /
    • 제24권7호
    • /
    • pp.1266-1275
    • /
    • 2019
  • With increasing criminal rates and number of CCTVs, much attention has been paid to intelligent surveillance system on the horizon. Object detection and tracking algorithms have been developed to reduce false alarms and accurately help security agents immediately response to undesirable changes in video clips such as crimes and accidents. Many studies have proposed a variety of algorithms to improve accuracy of detecting and tracking objects outside tunnels. The proposed methods might not work well in a tunnel because of low illuminance significantly susceptible to tail and warning lights of driving vehicles. The detection performance has rarely been tested against the tunnel environment. This study investigated a feasibility of object detection and tracking in an actual tunnel environment by utilizing YOLOv3 and Kernelized Correlation Filter. We tested 40 actual video clips to differentiate pedestrians and motorcycles to evaluate the performance of our algorithm. The experimental results showed significant difference in detection between pedestrians and motorcycles without false positive rates. Our findings are expected to provide a stepping stone of developing efficient detection algorithms suitable for tunnel environment and encouraging other researchers to glean reliable tracking data for smarter and safer City.

Fuzzy Based Shadow Removal and Integrated Boundary Detection for Video Surveillance

  • Niranjil, Kumar A.;Sureshkumar, C.
    • Journal of Electrical Engineering and Technology
    • /
    • 제9권6호
    • /
    • pp.2126-2133
    • /
    • 2014
  • We present a scalable object tracking framework, which is capable of removing shadows and tracking the people. The framework consists of background subtraction, fuzzy based shadow removal and boundary tracking algorithm. This work proposes a general-purpose method that combines statistical assumptions with the object-level knowledge of moving objects, apparent objects, and shadows acquired in the processing of the previous frames. Pixels belonging to moving objects and shadows are processed differently in order to supply an object-based selective update. Experimental results demonstrate that the proposed method is able to track the object boundaries under significant shadows with noise and background clutter.

AnoVid: 비디오 주석을 위한 심층 신경망 기반의 도구 (AnoVid: A Deep Neural Network-based Tool for Video Annotation)

  • 황지수;김인철
    • 한국멀티미디어학회논문지
    • /
    • 제23권8호
    • /
    • pp.986-1005
    • /
    • 2020
  • In this paper, we propose AnoVid, an automated video annotation tool based on deep neural networks, that automatically generates various meta data for each scene or shot in a long drama video containing rich elements. To this end, a novel meta data schema for drama video is designed. Based on this schema, the AnoVid video annotation tool has a total of six deep neural network models for object detection, place recognition, time zone recognition, person recognition, activity detection, and description generation. Using these models, the AnoVid can generate rich video annotation data. In addition, AnoVid provides not only the ability to automatically generate a JSON-type video annotation data file, but also provides various visualization facilities to check the video content analysis results. Through experiments using a real drama video, "Misaeing", we show the practical effectiveness and performance of the proposed video annotation tool, AnoVid.

Extensible Hierarchical Method of Detecting Interactive Actions for Video Understanding

  • Moon, Jinyoung;Jin, Junho;Kwon, Yongjin;Kang, Kyuchang;Park, Jongyoul;Park, Kyoung
    • ETRI Journal
    • /
    • 제39권4호
    • /
    • pp.502-513
    • /
    • 2017
  • For video understanding, namely analyzing who did what in a video, actions along with objects are primary elements. Most studies on actions have handled recognition problems for a well-trimmed video and focused on enhancing their classification performance. However, action detection, including localization as well as recognition, is required because, in general, actions intersect in time and space. In addition, most studies have not considered extensibility for a newly added action that has been previously trained. Therefore, proposed in this paper is an extensible hierarchical method for detecting generic actions, which combine object movements and spatial relations between two objects, and inherited actions, which are determined by the related objects through an ontology and rule based methodology. The hierarchical design of the method enables it to detect any interactive actions based on the spatial relations between two objects. The method using object information achieves an F-measure of 90.27%. Moreover, this paper describes the extensibility of the method for a new action contained in a video from a video domain that is different from the dataset used.

다면기법 SPFACS 영상객체를 이용한 AAM 알고리즘 적용 미소검출 설계 분석 (Using a Multi-Faced Technique SPFACS Video Object Design Analysis of The AAM Algorithm Applies Smile Detection)

  • 최병관
    • 디지털산업정보학회논문지
    • /
    • 제11권3호
    • /
    • pp.99-112
    • /
    • 2015
  • Digital imaging technology has advanced beyond the limits of the multimedia industry IT convergence, and to develop a complex industry, particularly in the field of object recognition, face smart-phones associated with various Application technology are being actively researched. Recently, face recognition technology is evolving into an intelligent object recognition through image recognition technology, detection technology, the detection object recognition through image recognition processing techniques applied technology is applied to the IP camera through the 3D image object recognition technology Face Recognition been actively studied. In this paper, we first look at the essential human factor, technical factors and trends about the technology of the human object recognition based SPFACS(Smile Progress Facial Action Coding System)study measures the smile detection technology recognizes multi-faceted object recognition. Study Method: 1)Human cognitive skills necessary to analyze the 3D object imaging system was designed. 2)3D object recognition, face detection parameter identification and optimal measurement method using the AAM algorithm inside the proposals and 3)Face recognition objects (Face recognition Technology) to apply the result to the recognition of the person's teeth area detecting expression recognition demonstrated by the effect of extracting the feature points.

동적 비디오 기반 안정화 및 객체 추적 방법 (A Method for Object Tracking Based on Background Stabilization)

  • 정훈조;이동은
    • 디지털산업정보학회논문지
    • /
    • 제14권1호
    • /
    • pp.77-85
    • /
    • 2018
  • This paper proposes a robust digital video stabilization algorithm to extract and track an object, which uses a phase correlation-based motion correction. The proposed video stabilization algorithm consists of background stabilization based on motion estimation and extraction of a moving object. The motion vectors can be estimated by calculating the phase correlation of a series of frames in the eight sub-images, which are located in the corner of the video. The global motion vector can be estimated and the image can be compensated by using the multiple local motions of sub-images. Through the calculations of the phase correlation, the motion of the background can be subtracted from the former frame and the compensated frame, which share the same background. The moving objects in the video can also be extracted. In this paper, calculating the phase correlation to track the robust motion vectors results in the compensation of vibrations, such as movement, rotation, expansion and the downsize of videos from all directions of the sub-images. Experimental results show that the proposed digital image stabilization algorithm can provide continuously stabilized videos and tracking object movements.

Object Recognition Algorithm with Partial Information

  • Yoo, Suk Won
    • International Journal of Advanced Culture Technology
    • /
    • 제7권4호
    • /
    • pp.229-235
    • /
    • 2019
  • Due to the development of video and optical technology today, video equipments are being used in a variety of fields such as identification, security maintenance, and factory automation systems that generate products. In this paper, we investigate an algorithm that effectively recognizes an experimental object in an input image with a partial problem due to the mechanical problem of the input imaging device. The object recognition algorithm proposed in this paper moves and rotates the vertices constituting the outline of the experimental object to the positions of the respective vertices constituting the outline of the DB model. Then, the discordance values between the moved and rotated experimental object and the corresponding DB model are calculated, and the minimum discordance value is selected. This minimum value is the final discordance value between the experimental object and the corresponding DB model, and the DB model with the minimum discordance value is selected as the recognition result for the experimental object. The proposed object recognition method obtains satisfactory recognition results using only partial information of the experimental object.

배경 모델 학습을 통한 객체 분할/검출 및 파티클 필터를 이용한 분할된 객체의 움직임 추적 방법 (Object Segmentation/Detection through learned Background Model and Segmented Object Tracking Method using Particle Filter)

  • 임수창;김도연
    • 한국정보통신학회논문지
    • /
    • 제20권8호
    • /
    • pp.1537-1545
    • /
    • 2016
  • 실시간영상에서 객체의 분할 및 추적은 침입자 감시와 로봇의 물체 추적, 증강현실의 객체 추적등 다양한 분야에서 사용되고 있다. 본 논문에서는 초기 입력 영상의 일부를 학습하여 배경모델로 제작한 후, 배경제거 방법을 이용하여 움직이는 객체의 분할을 통해 객체를 검출하였다. 검출된 객체의 영역을 기반으로 HSV 색상히스토그램과 파티클 필터를 이용하여 객체의 움직임을 추적하는 방법을 제안한다. 제안한 분할 방법은 평균 배경모델을 이용한 방법보다 주변환경 변화의 영향을 적게 받으며, 움직이는 객체의 검출 성능이 더욱 우수하였다. 또한 단일 객체 및 다수의 객체가 존재하는 환경에서 추적 객체가 유사한 색상 객체와 겹치는 경우, 추적 객체의 영역 절반 이상이 가려지는 경우에도 지속적으로 추적하는 결과를 얻을 수 있었다. 2개의 비디오 영상을 사용한 실험결과는 평균 중첩율 85.9%, 추적률 96.3%의 성능을 보여준다.

Vanishing point-based 3D object detection method for improving traffic object recognition accuracy

  • Jeong-In, Park
    • 한국컴퓨터정보학회논문지
    • /
    • 제28권1호
    • /
    • pp.93-101
    • /
    • 2023
  • 이 논문은 영상 카메라를 이용하여 교통 객체를 인식하고자 하는 경우, 영상 내 객체 인식 정확도를 높이기 위해 소실점을 이용하여 객체에 대한 3D 바운딩 박스를 생성하는 방법이다. 최근 인공지능을 이용하여 교통 영상 카메라로 촬영된 차량을 검출하고자 하는 경우 이 3D 바운딩 박스 생성 알고리즘을 적용하고자 한다. 카메라 설치 각도와 카메라가 촬영한 영상의 방향성을 분석하여 종 방향 소실점(VP1)과 횡 방향 소실점(VP2)을 도출하고 이를 기반으로 분석 대상 동영상에서 이동하는 객체를 특정하게 된다. 이 알고리즘을 적용하면 감지된 객체의 위치, 종류, 크기 등 객체 정보 검출이 용이하고, 이를 자동차와 같은 이동류에 적용하는 경우 이를 트래킹하여 각 객체가 이동한 위치와 좌표, 이동속도 및 방향 등을 알 수 있다. 실제 도로에 적용한 결과 트래킹이 10% 향상되었으며 특히 음영지역(큰 차에 가려진 극히 적은 차량 부위)의 인식율과 트래킹이 100% 개선되는 등 교통 데이터 분석 정확성을 향상시킬 수 있었다.