• Title/Summary/Keyword: camera model

Search Result 1,516, Processing Time 0.033 seconds

Vehicle Type Classification Model based on Deep Learning for Smart Traffic Control Systems (스마트 교통 단속 시스템을 위한 딥러닝 기반 차종 분류 모델)

  • Kim, Doyeong;Jang, Sungjin;Jang, Jongwook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.469-472
    • /
    • 2022
  • With the recent development of intelligent transportation systems, various technologies applying deep learning technology are being used. To crackdown on illegal vehicles and criminal vehicles driving on the road, a vehicle type classification system capable of accurately determining the type of vehicle is required. This study proposes a vehicle type classification system optimized for mobile traffic control systems using YOLO(You Only Look Once). The system uses a one-stage object detection algorithm YOLOv5 to detect vehicles into six classes: passenger cars, subcompact, compact, and midsize vans, full-size vans, trucks, motorcycles, special vehicles, and construction machinery. About 5,000 pieces of domestic vehicle image data built by the Korea Institute of Science and Technology for the development of artificial intelligence technology were used as learning data. It proposes a lane designation control system that applies a vehicle type classification algorithm capable of recognizing both front and side angles with one camera.

  • PDF

Development of CanSat System for Vehicle Tracking based on Jetson Nano (젯슨 나노 기반의 차량 추적 캔위성 시스템 개발)

  • Lee, Younggun;Lee, Sanghyun;You, Seunghoon;Lee, Sangku
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.556-558
    • /
    • 2022
  • This paper proposes a CanSat system with a vehicle tracking function based on Jetson Nano, a high-performance small computer capable of operating artificial intelligence algorithms. The CanSat system consists of a CanSat and a ground station. The CanSat falls in the atmosphere and transmits the data obtained through the installed sensors to the ground station using wireless communication. The existing CanSat is limited to the mission of simply transmitting the collected information to the ground station, and there is a limit to efficiently performing the mission due to the limited fall time and bandwidth limitation of wireless communication. The Jetson Nano based CanSat proposed in this paper uses a pre-trained neural network model to detect the location of a vehicle in each image taken from the air in real time, and then uses a 2-axis motor to move the camera to track the vehicle.

  • PDF

Utilization of UAV Remote Sensing in Small-scale Field Experiment : Case Study in Evaluation of Plat-based LAI for Sweetcorn Production

  • Hyunjin Jung;Rongling Ye;Yang Yi;Naoyuki Hashimoto;Shuhei Yamamoto;Koki Homma
    • Proceedings of the Korean Society of Crop Science Conference
    • /
    • 2022.10a
    • /
    • pp.75-75
    • /
    • 2022
  • Traditional agriculture mostly focused on activity in the field, but current agriculture faces problems such as reduction of agricultural inputs, labor shortage and so on. Accordingly, traditional agricultural experiments generally considered the simple treatment effects, but current agricultural experiments need to consider the several and complicate treatment effects. To analyze such several and complicate treatment effects, data collection has the first priority. Remote sensing is a quite effective tool to collect information in agriculture, and recent easier availability of UAVs (Unmanned Aerial Vehicles) enhances the effectiveness. LAI (Leaf Area Index) is one of the most important information for evaluating the condition of crop growth. In this study, we utilized UAV with multispectral camera to evaluate plant-based LAI of sweetcorn in a small-scale field experiment and discussed the feasibility of a new experimental design to analyze the several and complicate treatment effects. The plant-based SR measured by UAV showed the highest correlation coefficient with LAI measured by a canopy analyzer in 2018 and 2019. Application of linear mix model showed that plant-based SR data had higher detection power due to its huge number of data although SR was inferior to evaluate LAI than the canopy analyzer. The distribution of plant-based data also statistically revealed the border effect in treatment plots in the traditional experimental design. These results suggest that remote sensing with UAVs has the advantage even in a small-scale experimental plot and has a possibility to provide a new experimental design if combined with various analytical applications such as plant size, shape, and color.

  • PDF

A Method of Extracting Features of Sensor-only Facilities for Autonomous Cooperative Driving

  • Hyung Lee;Chulwoo Park;Handong Lee;Sanyeon Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.12
    • /
    • pp.191-199
    • /
    • 2023
  • In this paper, we propose a method to extract the features of five sensor-only facilities built as infrastructure for autonomous cooperative driving, which are from point cloud data acquired by LiDAR. In the case of image acquisition sensors installed in autonomous vehicles, the acquisition data is inconsistent due to the climatic environment and camera characteristics, so LiDAR sensor was applied to replace them. In addition, high-intensity reflectors were designed and attached to each facility to make it easier to distinguish it from other existing facilities with LiDAR. From the five sensor-only facilities developed and the point cloud data acquired by the data acquisition system, feature points were extracted based on the average reflective intensity of the high-intensity reflective paper attached to the facility, clustered by the DBSCAN method, and changed to two-dimensional coordinates by a projection method. The features of the facility at each distance consist of three-dimensional point coordinates, two-dimensional projected coordinates, and reflection intensity, and will be used as training data for a model for facility recognition to be developed in the future.

A Study on the Elevator System Using Real-time Object Detection Technology YOLOv5 (실시간 객체 검출 기술 YOLOv5를 이용한 스마트 엘리베이터 시스템에 관한 연구)

  • Sun-Been Park;Yu-Jeong Jeong;Da-Eun Lee;Tae-Kook Kim
    • Journal of Internet of Things and Convergence
    • /
    • v.10 no.2
    • /
    • pp.103-108
    • /
    • 2024
  • In this paper, a smart elevator system was studied using real-time object detection technology based on YOLO(You only look once)v5. When an external elevator button is pressed, the YOLOv5 model analyzes the camera video to determine whether there are people waiting, and if it determines that there are no people waiting, the button is automatically canceled. The study introduces an effective method of implementing object detection and communication technology through YOLOv5 and MQTT (Message Queuing Telemetry Transport) used in the Internet of Things. And using this, we implemented a smart elevator system that determines in real time whether there are people waiting. The proposed system can play the role of CCTV (closed-circuit television) while reducing unnecessary power consumption. Therefore, the proposed smart elevator system is expected to contribute to safety and security issues.

Assessment of Fire-Damaged Mortar using Color image Analysis (색도 이미지 분석을 이용한 화재 피해 모르타르의 손상 평가)

  • Park, Kwang-Min;Lee, Byung-Do;Yoo, Sung-Hun;Ham, Nam-Hyuk;Roh, Young-Sook
    • Journal of the Korea institute for structural maintenance and inspection
    • /
    • v.23 no.3
    • /
    • pp.83-91
    • /
    • 2019
  • The purpose of this study is to assess a fire-damaged concrete structure using a digital camera and image processing software. To simulate it, mortar and paste samples of W/C=0.5(general strength) and 0.3(high strength) were put into an electric furnace and simulated from $100^{\circ}C$ to $1000^{\circ}C$. Here, the paste was processed into a powder to measure CIELAB chromaticity, and the samples were taken with a digital camera. The RGB chromaticity was measured by color intensity analyzer software. As a result, the residual compressive strength of W/C=0.5 and 0.3 was 87.2 % and 86.7 % at the heating temperature of $400^{\circ}C$. However there was a sudden decrease in strength at the temperature above $500^{\circ}C$, while the residual compressive strength of W/C=0.5 and 0.3 was 55.2 % and 51.9 % of residual strength. At the temperature $700^{\circ}C$ or higher, W/C=0.5 and W/C=0.3 show 26.3% and 27.8% of residual strength, so that the durability of the structure could not be secured. The results of $L^*a^*b$ color analysis show that $b^*$ increases rapidly after $700^{\circ}C$. It is analyzed that the intensity of yellow becomes strong after $700^{\circ}C$. Further, the RGB analysis found that the histogram kurtosis and frequency of Red and Green increases after $700^{\circ}C$. It is analyzed that number of Red and Green pixels are increased. Therefore, it is deemed possible to estimate the degree of damage by checking the change in yellow($b^*$ or R+G) when analyzing the chromaticity of the fire-damaged concrete structures.

An Analysis on the Sinking Resistance of Purse Seine - 2. In the Case of the Model Purse Seine with Different Netting Material and Sinkers - (旋網의 沈降 抵抗 解析 - 2. 網地材料와 沈子量 다른 模型網의 경우 -)

  • Kim, Suk-Jong
    • Journal of the Korean Society of Fisheries and Ocean Technology
    • /
    • v.40 no.1
    • /
    • pp.29-36
    • /
    • 2004
  • This study deals with an analysis on the sinking resistance for the model purse seine, in the case of different netting material and sinkers. The experiment was carried out using rune simplified model seines of knotless nettings. Dimension of model seines 420cm for corkline and 85cm for seine depth, three groups of models rigged 25, 45 and 60g with the same weighted sinkers in water were used. These were named PP-25, PA-25, PES-25, PP-45, PA-45, PES-45, PP-60, PA-60 and PES-60 seine. The densitie($\rho$) of netting materials were 0.91g/cm$cm^3$, 1.14g/cm$cm^3$ and 1.38g/cm$m^3$. Experiments carried out in the observation channel in a flume tank under still water conditions. Sinking motion was recorded by the one set of TV-camera for VTR, and reading coordinate carried out by the video digitization system. Differential equations were derived from the conservation of momenta of the model purse seines and used to determine the sinking speeds of the depths of leadline and the other portions of the seines. An analysis carried out by simultaneous differential equations for numerical method by sub-routine Runge-Kutta-Gill The results obtained were as follows : 1. Average sinking speed of leadline for the model seines rigged 60g with the same weighted sinkers in water was fastest for 12.2cm/sec of PES seine, followed by 11.4cm/sec of PA and 10.7cm/sec of PP seines. 2. The coefficient of resistance for netting of seine was estimated to be $K_D=0.09(\frac{\rho}{\rho_w})^4$ 3. The coefficient of resistance for netting bundle of seine was estimated to be $C_R=0.91(\frac{\rho}{\rho_w})$ 4. In all seines, the calculated depths of leadline closely agreed with the measured ones, each 25g, 45g, 60g of weighted sinkers were put into formulas meas.=1.04cal., meas.=0.99cal. and meas.=0.98 cal.

Estimation of Rice Heading Date of Paddy Rice from Slanted and Top-view Images Using Deep Learning Classification Model (딥 러닝 분류 모델을 이용한 직하방과 경사각 영상 기반의 벼 출수기 판별)

  • Hyeok-jin Bak;Wan-Gyu Sang;Sungyul Chang;Dongwon Kwon;Woo-jin Im;Ji-hyeon Lee;Nam-jin Chung;Jung-Il Cho
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.25 no.4
    • /
    • pp.337-345
    • /
    • 2023
  • Estimating the rice heading date is one of the most crucial agricultural tasks related to productivity. However, due to abnormal climates around the world, it is becoming increasingly challenging to estimate the rice heading date. Therefore, a more objective classification method for estimating the rice heading date is needed than the existing methods. This study, we aimed to classify the rice heading stage from various images using a CNN classification model. We collected top-view images taken from a drone and a phenotyping tower, as well as slanted-view images captured with a RGB camera. The collected images underwent preprocessing to prepare them as input data for the CNN model. The CNN architectures employed were ResNet50, InceptionV3, and VGG19, which are commonly used in image classification models. The accuracy of the models all showed an accuracy of 0.98 or higher regardless of each architecture and type of image. We also used Grad-CAM to visually check which features of the image the model looked at and classified. Then verified our model accurately measure the rice heading date in paddy fields. The rice heading date was estimated to be approximately one day apart on average in the four paddy fields. This method suggests that the water head can be estimated automatically and quantitatively when estimating the rice heading date from various paddy field monitoring images.

3D Modeling from 2D Stereo Image using 2-Step Hybrid Method (2단계 하이브리드 방법을 이용한 2D 스테레오 영상의 3D 모델링)

  • No, Yun-Hyang;Go, Byeong-Cheol;Byeon, Hye-Ran;Yu, Ji-Sang
    • Journal of KIISE:Software and Applications
    • /
    • v.28 no.7
    • /
    • pp.501-510
    • /
    • 2001
  • Generally, it is essential to estimate exact disparity for the 3D modeling from stereo images. Because existing methods calculate disparities from a whole image, they require too much cimputational time and bring about the mismatching problem. In this article, using the characteristic that the disparity vectors in stereo images are distributed not equally in a whole image but only exist about the background and obhect, we do a wavelet transformation on stereo images and estimate coarse disparity fields from the reduced lowpass field using area-based method at first-step. From these coarse disparity vectors, we generate disparity histogram and then separate object from background area using it. Afterwards, we restore only object area to the original image and estimate dense and accurate disparity by our two-step pixel-based method which does not use pixel brightness but use second gradient. We also extract feature points from the separated object area and estimate depth information by applying disparity vectors and camera parameters. Finally, we generate 3D model using both feature points and their z coordinates. By using our proposed, we can considerably reduce the computation time and estimate the precise disparity through the additional pixel-based method using LOG filter. Furthermore, our proposed foreground/background method can solve the mismatching problem of existing Delaunay triangulation and generate accurate 3D model.

  • PDF

A Study on the Selection and Applicability Analysis of 3D Terrain Modeling Sensor for Intelligent Excavation Robot (지능형 굴삭 로봇의 개발을 위한 로컬영역 3차원 모델링 센서 선정 및 현장 적용성 분석에 관한 연구)

  • Yoo, Hyun-Seok;Kwon, Soon-Wook;Kim, Young-Suk
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.33 no.6
    • /
    • pp.2551-2562
    • /
    • 2013
  • Since 2006, an Intelligent Excavation Robot which automatically performs the earth-work without operator has been developed in Korea. The technologies for automatically recognizing the terrain of work environment and detecting the objects such as obstacles or dump trucks are essential for its work quality and safety. In several countries, terrestrial 3D laser scanner and stereo vision camera have been used to model the local area around workspace of the automated construction equipment. However, these attempts have some problems that require high cost to make the sensor system or long processing time to eliminate the noise from 3D model outcome. The objectives of this study are to analyze the advantages of the existing 3D modeling sensors and to examine the applicability for practical use by using Analytic Hierarchical Process(AHP). In this study, 3D modeling quality and accuracy of modeling sensors were tested at the real earth-work environment.