• Title/Summary/Keyword: Object Recognition Technology

Search Result 471, Processing Time 0.03 seconds

Yolo based Light Source Object Detection for Traffic Image Big Data Processing (교통 영상 빅데이터 처리를 위한 Yolo 기반 광원 객체 탐지)

  • Kang, Ji-Soo;Shim, Se-Eun;Jo, Sun-Moon;Chung, Kyungyong
    • Journal of Convergence for Information Technology
    • /
    • v.10 no.8
    • /
    • pp.40-46
    • /
    • 2020
  • As interest in traffic safety increases, research on autonomous driving, which reduces the incidence of traffic accidents, is increased. Object recognition and detection are essential for autonomous driving. Therefore, research on object recognition and detection through traffic image big data is being actively conducted to determine the road conditions. However, because most existing studies use only daytime data, it is difficult to recognize objects on night roads. Particularly, in the case of a light source object, it is difficult to use the features of the daytime as it is due to light smudging and whitening. Therefore, this study proposes Yolo based light source object detection for traffic image big data processing. The proposed method performs image processing by applying color model transitions to night traffic image. The object group is determined by extracting the characteristics of the object through image processing. It is possible to increase the recognition rate of light source object detection on a night road through a deep learning model using candidate group data.

3-D Underwater Object Recognition Using PZT-Epoxy 3-3 Type Composite Ultrasonic Transducers (PZT-에폭시 3-3형 복합압전체 초음파 트랜스듀서를 사용한 3차원 수중 물체인식)

  • Cho, Hyun-Chul;Heo, Jin;SaGong, Geon
    • Journal of Sensor Science and Technology
    • /
    • v.10 no.6
    • /
    • pp.286-294
    • /
    • 2001
  • In this study, 3-D underwater object recognition using the self-made 3-3 type composite ultrasonic transducer and modified SOFM(Self Organizing Feature Map) neural network are investigated. Properties of the self-made 3-3 type composite specimens are satisfied considerably with requirements as an underwater ultrasonic transducer's materials. 3-D underwater all object's recognition rates obtained from both the training data and testing data in different objects, such as a rectangular block, regular triangular block, square block and cylinderical block, were 100% and 94.0%, respectively. All object's recognition rates are obtained by utilizing the self-made 3-3 type composite transducer and SOFM neural network. From the object recognition rates, it could be seen that an ultrasonic transducer fabricated with the self-made 3-3 type composite resonator will be able to have application for the underwater object recognition.

  • PDF

Multi-Object Tracking based on Reliability Assessment of Learning in Mobile Environment (모바일 환경 신뢰도 평가 학습에 의한 다중 객체 추적)

  • Han, Woo ri;Kim, Young-Seop;Lee, Yong-Hwan
    • Journal of the Semiconductor & Display Technology
    • /
    • v.14 no.3
    • /
    • pp.73-77
    • /
    • 2015
  • This paper proposes an object tracking system according to reliability assessment of learning in mobile environments. The proposed system is based on markerless tracking, and there are four modules which are recognition, tracking, detecting and learning module. Recognition module detects and identifies an object to be matched on current frame correspond to the database using LSH through SURF, and then this module generates a standard object information that has the best reliability of learning. The standard object information is used for evaluating and learning the object that is successful tracking in tracking module. Detecting module finds out the object based on having the best possible knowledge available among the learned objects information, when the system fails to track. The experimental results show that the proposed system is able to recognize and track the reliable objects with reliability assessment of learning for the use of mobile platform.

Recent Trends of Object and Scene Recognition Technologies for Mobile/Embedded Devices (모바일/임베디드 객체 및 장면 인식 기술 동향)

  • Lee, S.W.;Lee, G.D.;Ko, J.G.;Lee, S.J.;Yoo, W.Y.
    • Electronics and Telecommunications Trends
    • /
    • v.34 no.6
    • /
    • pp.133-144
    • /
    • 2019
  • Although deep learning-based visual image recognition technology has evolved rapidly, most of the commonly used methods focus solely on recognition accuracy. However, the demand for low latency and low power consuming image recognition with an acceptable accuracy is rising for practical applications in edge devices. For example, most Internet of Things (IoT) devices have a low computing power requiring more pragmatic use of these technologies; in addition, drones or smartphones have limited battery capacity again requiring practical applications that take this into consideration. Furthermore, some people do not prefer that central servers process their private images, as is required by high performance serverbased recognition technologies. To address these demands, the object and scene recognition technologies for mobile/embedded devices that enable optimized neural networks to operate in mobile and embedded environments are gaining attention. In this report, we briefly summarize the recent trends and issues of object and scene recognition technologies for mobile and embedded devices.

YOLO based Optical Music Recognition and Virtual Reality Content Creation Method (YOLO 기반의 광학 음악 인식 기술 및 가상현실 콘텐츠 제작 방법)

  • Oh, Kyeongmin;Hong, Yoseop;Baek, Geonyeong;Chun, Chanjun
    • Smart Media Journal
    • /
    • v.10 no.4
    • /
    • pp.80-90
    • /
    • 2021
  • Using optical music recognition technology based on deep learning, we propose to apply the results derived to VR games. To detect the music objects in the music sheet, the deep learning model used YOLO v5, and Hough transform was employed to detect undetected objects, modifying the size of the staff. It analyzes and uses BPM, maximum number of combos, and musical notes in VR games using output result files, and prevents the backlog of notes through Object Pooling technology for resource management. In this paper, VR games can be produced with music elements derived from optical music recognition technology to expand the utilization of optical music recognition along with providing VR contents.

A Vehicle Recognition Method based on Radar and Camera Fusion in an Autonomous Driving Environment

  • Park, Mun-Yong;Lee, Suk-Ki;Shin, Dong-Jin
    • International journal of advanced smart convergence
    • /
    • v.10 no.4
    • /
    • pp.263-272
    • /
    • 2021
  • At a time when securing driving safety is the most important in the development and commercialization of autonomous vehicles, AI and big data-based algorithms are being studied to enhance and optimize the recognition and detection performance of various static and dynamic vehicles. However, there are many research cases to recognize it as the same vehicle by utilizing the unique advantages of radar and cameras, but they do not use deep learning image processing technology or detect only short distances as the same target due to radar performance problems. Radars can recognize vehicles without errors in situations such as night and fog, but it is not accurate even if the type of object is determined through RCS values, so accurate classification of the object through images such as cameras is required. Therefore, we propose a fusion-based vehicle recognition method that configures data sets that can be collected by radar device and camera device, calculates errors in the data sets, and recognizes them as the same target.

Learning Methods for Effective Object Tracking in 3D Storytelling Augmented Reality (3D 스토리텔링 증강현실에서 효과적인 객체 추적을 위한 학습 방법)

  • Choi, Dae han;Han, Woo ri;Lee, Yong-Hwan;Kim, Youngseop
    • Journal of the Semiconductor & Display Technology
    • /
    • v.15 no.3
    • /
    • pp.46-50
    • /
    • 2016
  • Recently, Depending on expectancy effect and ripple effect of augmented reality, the convergence between augmented reality and culture & arts are being actively conducted. This paper proposes a learning method for effective object tracking in 3D storytelling augmented reality in cultural properties. The proposed system is based on marker-less tracking, and there are four modules that are recognition, tracking, detecting and learning module. Recognition module is composed of SURF and LSH, and then this module generates standard object information. Tracking module tracks an object using object tracking based on reliability. This information is stored in Learning module along with learned time information. Detecting module finds out the object based on having the best possible knowledge available among the learned objects information, when the system fails to track. Also, it proposes a method for robustly implementing a 3D storytelling augmented reality in cultural properties in the future.

Implementation of Character and Object Metadata Generation System for Media Archive Construction (미디어 아카이브 구축을 위한 등장인물, 사물 메타데이터 생성 시스템 구현)

  • Cho, Sungman;Lee, Seungju;Lee, Jaehyeon;Park, Gooman
    • Journal of Broadcast Engineering
    • /
    • v.24 no.6
    • /
    • pp.1076-1084
    • /
    • 2019
  • In this paper, we introduced a system that extracts metadata by recognizing characters and objects in media using deep learning technology. In the field of broadcasting, multimedia contents such as video, audio, image, and text have been converted to digital contents for a long time, but the unconverted resources still remain vast. Building media archives requires a lot of manual work, which is time consuming and costly. Therefore, by implementing a deep learning-based metadata generation system, it is possible to save time and cost in constructing media archives. The whole system consists of four elements: training data generation module, object recognition module, character recognition module, and API server. The deep learning network module and the face recognition module are implemented to recognize characters and objects from the media and describe them as metadata. The training data generation module was designed separately to facilitate the construction of data for training neural network, and the functions of face recognition and object recognition were configured as an API server. We trained the two neural-networks using 1500 persons and 80 kinds of object data and confirmed that the accuracy is 98% in the character test data and 42% in the object data.

A Study on The Classification of Target-objects with The Deep-learning Model in The Vision-images (딥러닝 모델을 이용한 비전이미지 내의 대상체 분류에 관한 연구)

  • Cho, Youngjoon;Kim, Jongwon
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.22 no.2
    • /
    • pp.20-25
    • /
    • 2021
  • The target-object classification method was implemented using a deep-learning-based detection model in real-time images. The object detection model was a deep-learning-based detection model that allowed extensive data collection and machine learning processes to classify similar target-objects. The recognition model was implemented by changing the processing structure of the detection model and combining developed the vision-processing module. To classify the target-objects, the identity and similarity were defined and applied to the detection model. The use of the recognition model in industry was also considered by verifying the effectiveness of the recognition model using the real-time images of an actual soccer game. The detection model and the newly constructed recognition model were compared and verified using real-time images. Furthermore, research was conducted to optimize the recognition model in a real-time environment.

Development of IoT System Based on Context Awareness to Assist the Visually Impaired

  • Song, Mi-Hwa
    • International Journal of Advanced Culture Technology
    • /
    • v.9 no.4
    • /
    • pp.320-328
    • /
    • 2021
  • As the number of visually impaired people steadily increases, interest in independent walking is also increasing. However, there are various inconveniences in the independent walking of the visually impaired at present, reducing the quality of life of the visually impaired. The white cane, which is an existing walking aid for the visually impaired, has difficulty in recognizing upper obstacles and obstacles outside the effective distance. In addition, it is inconvenient to cross the street because the sound signal to help the visually impaired cross the crosswalk is lacking or damaged. These factors make it difficult for the visually impaired to walk independently. Therefore, we propose the design of an embedded system that provides traffic light recognition through object recognition technology, voice guidance using TTS, and upper obstacle recognition through ultrasonic sensors so that blind people can realize safe and high-quality independent walking.