• 제목/요약/키워드: Computer Vision

검색결과 2,208건 처리시간 0.035초

Intelligent User Pattern Recognition based on Vision, Audio and Activity for Abnormal Event Detections of Single Households

  • Jung, Ju-Ho;Ahn, Jun-Ho
    • 한국컴퓨터정보학회논문지
    • /
    • 제24권5호
    • /
    • pp.59-66
    • /
    • 2019
  • According to the KT telecommunication statistics, people stayed inside their houses on an average of 11.9 hours a day. As well as, according to NSC statistics in the united states, people regardless of age are injured for a variety of reasons in their houses. For purposes of this research, we have investigated an abnormal event detection algorithm to classify infrequently occurring behaviors as accidents, health emergencies, etc. in their daily lives. We propose a fusion method that combines three classification algorithms with vision pattern, audio pattern, and activity pattern to detect unusual user events. The vision pattern algorithm identifies people and objects based on video data collected through home CCTV. The audio and activity pattern algorithms classify user audio and activity behaviors using the data collected from built-in sensors on their smartphones in their houses. We evaluated the proposed individual pattern algorithm and fusion method based on multiple scenarios.

Detection of Traditional Costumes: A Computer Vision Approach

  • Marwa Chacha Andrea;Mi Jin Noh;Choong Kwon Lee
    • 스마트미디어저널
    • /
    • 제12권11호
    • /
    • pp.125-133
    • /
    • 2023
  • Traditional attire has assumed a pivotal role within the contemporary fashion industry. The objective of this study is to construct a computer vision model tailored to the recognition of traditional costumes originating from five distinct countries, namely India, Korea, Japan, Tanzania, and Vietnam. Leveraging a dataset comprising 1,608 images, we proceeded to train the cutting-edge computer vision model YOLOv8. The model yielded an impressive overall mean average precision (MAP) of 96%. Notably, the Indian sari exhibited a remarkable MAP of 99%, the Tanzanian kitenge 98%, the Japanese kimono 92%, the Korean hanbok 89%, and the Vietnamese ao dai 83%. Furthermore, the model demonstrated a commendable overall box precision score of 94.7% and a recall rate of 84.3%. Within the realm of the fashion industry, this model possesses considerable utility for trend projection and the facilitation of personalized recommendation systems.

Essential Computer Vision Methods for Maximal Visual Quality of Experience on Augmented Reality

  • Heo, Suwoong;Song, Hyewon;Kim, Jinwoo;Nguyen, Anh-Duc;Lee, Sanghoon
    • Journal of International Society for Simulation Surgery
    • /
    • 제3권2호
    • /
    • pp.39-45
    • /
    • 2016
  • The augmented reality is the environment which consists of real-world view and information drawn by computer. Since the image which user can see through augmented reality device is a synthetic image composed by real-view and virtual image, it is important to make the virtual image generated by computer well harmonized with real-view image. In this paper, we present reviews of several works about computer vision and graphics methods which give user realistic augmented reality experience. To generate visually harmonized synthetic image which consists of a real and a virtual image, 3D geometry and environmental information such as lighting or material surface reflectivity should be known by the computer. There are lots of computer vision methods which aim to estimate those. We introduce some of the approaches related to acquiring geometric information, lighting environment and material surface properties using monocular or multi-view images. We expect that this paper gives reader's intuition of the computer vision methods for providing a realistic augmented reality experience.

로보트와 Vision System Interface (Robot vision interface)

  • 김선일;여인택;박찬웅
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1987년도 한국자동제어학술회의논문집; 한국과학기술대학, 충남; 16-17 Oct. 1987
    • /
    • pp.101-104
    • /
    • 1987
  • This paper shows the robot-vision system which consists of robot, vision system, single board computer and IBM-PC. IBM-PC based system has a great flexibility in expansion for a vision system interfacing. Easy human interfacing and great calculation ability are the benefits of this system. It was carried to interface between each component. The calibration between two coordinate systems is studied. The robot language for robot-vision system was written in "C" language. User also can write job program in "C" language in which the robot and vision related functions reside in the library.side in the library.

  • PDF

Construction of Confusion Lines for Color Vision Deficiency and Verification by Ishihara Chart

  • Cho, Keuyhong;Lee, Jusun;Song, Sanghoon;Han, Dongil
    • IEIE Transactions on Smart Processing and Computing
    • /
    • 제4권4호
    • /
    • pp.272-280
    • /
    • 2015
  • This paper proposes color databases that can be used for various purposes for people with a color vision deficiency (CVD). The purpose of this paper is to group colors within the sRGB gamut into the CIE $L^*a^*b^*$ color space using the Brettel algorithm to simulate the representative colors of each group into colors visible to people with a CVD, and to establish a confusion line database by comparing colors that might cause confusion for people with different types of color vision deficiency. The validity of the established confusion lines were verified by using an Ishihara chart. The different colors that confuse those with a CVD in an Ishihara chart are located in the same confusion line database for both protanopia and deutanopia. Instead of the 3D RGB color space, we have grouped confusion colors to the CIE $L^*a^*b^*$ space coordinates in a more distinctive and intuitive manner, and can establish a database of colors that can be perceived by people with a CVD more accurately. Editor - Highlight - Do these changes reflect the intended meaning? If not, please rephrase as intended.

A Vision-Based Method to Find Fingertips in a Closed Hand

  • Chaudhary, Ankit;Vatwani, Kapil;Agrawal, Tushar;Raheja, J.L.
    • Journal of Information Processing Systems
    • /
    • 제8권3호
    • /
    • pp.399-408
    • /
    • 2012
  • Hand gesture recognition is an important area of research in the field of Human Computer Interaction (HCI). The geometric attributes of the hand play an important role in hand shape reconstruction and gesture recognition. That said, fingertips are one of the important attributes for the detection of hand gestures and can provide valuable information from hand images. Many methods are available in scientific literature for fingertips detection with an open hand but very poor results are available for fingertips detection when the hand is closed. This paper presents a new method for the detection of fingertips in a closed hand using the corner detection method and an advanced edge detection algorithm. It is important to note that the skin color segmentation methodology did not work for fingertips detection in a closed hand. Thus the proposed method applied Gabor filter techniques for the detection of edges and then applied the corner detection algorithm for the detection of fingertips through the edges. To check the accuracy of the method, this method was tested on a vast number of images taken with a webcam. The method resulted in a higher accuracy rate of detections from the images. The method was further implemented on video for testing its validity on real time image capturing. These closed hand fingertips detection would help in controlling an electro-mechanical robotic hand via hand gesture in a natural way.

비전 센서를 사용하는 실시간 물류 파악 시스템 구현 (Implementation of Real-time Logistics Identification System using Vision Sensors)

  • 김동휘;박민혁;박성재;박정규
    • 한국정보통신학회:학술대회논문집
    • /
    • 한국정보통신학회 2022년도 추계학술대회
    • /
    • pp.172-174
    • /
    • 2022
  • 국내에서 물류를 처리하는 회사에서는 다양한 형태의 제품을 입출고 처리하고 있다. 다양한 형태의 제품을 처리하기 위해서 수작업으로 분류 업부를 수행하고 있다. 본 논문에서는 적은 인력으로 고효율을 내기 위해 비전 센서를 사용하는 실시간 QR코드 탐지 방법을 제안한다. 제한하는 시스템은 비전 센서를 사용하여 물류의 QR코드 인식을 실시간으로 처리가 가능하다. 제안하는 시스템은 물류의 단인 QR 코드 인식이 아닌 다중 인식을 통해서 다량의 QR 코드를 빠르게 파악할 수 있다. 연구에서는 시스템을 실제 구현하여 검증을 진행하여 비전 센터를 통해서 이미지에서 다중 QR 인식을 확인하였다.

  • PDF

Estimation of Angular Acceleration By a Monocular Vision Sensor

  • Lim, Joonhoo;Kim, Hee Sung;Lee, Je Young;Choi, Kwang Ho;Kang, Sung Jin;Chun, Sebum;Lee, Hyung Keun
    • Journal of Positioning, Navigation, and Timing
    • /
    • 제3권1호
    • /
    • pp.1-10
    • /
    • 2014
  • Recently, monitoring of two-body ground vehicles carrying extremely hazardous materials has been considered as one of the most important national issues. This issue induces large cost in terms of national economy and social benefit. To monitor and counteract accidents promptly, an efficient methodology is required. For accident monitoring, GPS can be utilized in most cases. However, it is widely known that GPS cannot provide sufficient continuity in urban cannons and tunnels. To complement the weakness of GPS, this paper proposes an accident monitoring method based on a monocular vision sensor. The proposed method estimates angular acceleration from a sequence of image frames captured by a monocular vision sensor. The possibility of using angular acceleration is investigated to determine the occurrence of accidents such as jackknifing and rollover. By an experiment based on actual measurements, the feasibility of the proposed method is evaluated.

A completely non-contact recognition system for bridge unit influence line using portable cameras and computer vision

  • Dong, Chuan-Zhi;Bas, Selcuk;Catbas, F. Necati
    • Smart Structures and Systems
    • /
    • 제24권5호
    • /
    • pp.617-630
    • /
    • 2019
  • Currently most of the vision-based structural identification research focus either on structural input (vehicle location) estimation or on structural output (structural displacement and strain responses) estimation. The structural condition assessment at global level just with the vision-based structural output cannot give a normalized response irrespective of the type and/or load configurations of the vehicles. Combining the vision-based structural input and the structural output from non-contact sensors overcomes the disadvantage given above, while reducing cost, time, labor force including cable wiring work. In conventional traffic monitoring, sometimes traffic closure is essential for bridge structures, which may cause other severe problems such as traffic jams and accidents. In this study, a completely non-contact structural identification system is proposed, and the system mainly targets the identification of bridge unit influence line (UIL) under operational traffic. Both the structural input (vehicle location information) and output (displacement responses) are obtained by only using cameras and computer vision techniques. Multiple cameras are synchronized by audio signal pattern recognition. The proposed system is verified with a laboratory experiment on a scaled bridge model under a small moving truck load and a field application on a footbridge on campus under a moving golf cart load. The UILs are successfully identified in both bridge cases. The pedestrian loads are also estimated with the extracted UIL and the predicted weights of pedestrians are observed to be in acceptable ranges.