• Title/Summary/Keyword: Recognition algorithm

Search Result 3,539, Processing Time 0.032 seconds

Artificial neural network model for predicting sex using dental and orthodontic measurements

  • Sandra Anic-Milosevic;Natasa Medancic;Martina Calusic-Sarac;Jelena Dumancic;Hrvoje Brkic
    • The korean journal of orthodontics
    • /
    • v.53 no.3
    • /
    • pp.194-204
    • /
    • 2023
  • Objective: To investigate sex-specific correlations between the dimensions of permanent canines and the anterior Bolton ratio and to construct a statistical model capable of identifying the sex of an unknown subject. Methods: Odontometric data were collected from 121 plaster study models derived from Caucasian orthodontic patients aged 12-17 years at the pretreatment stage by measuring the dimensions of the permanent canines and Bolton's anterior ratio. Sixteen variables were collected for each subject: 12 dimensions of the permanent canines, sex, age, anterior Bolton ratio, and Angle's classification. Data were analyzed using inferential statistics, principal component analysis, and artificial neural network modeling. Results: Sex-specific differences were identified in all odontometric variables, and an artificial neural network model was prepared that used odontometric variables for predicting the sex of the participants with an accuracy of > 80%. This model can be applied for forensic purposes, and its accuracy can be further improved by adding data collected from new subjects or adding new variables for existing subjects. The improvement in the accuracy of the model was demonstrated by an increase in the percentage of accurate predictions from 72.0-78.1% to 77.8-85.7% after the anterior Bolton ratio and age were added. Conclusions: The described artificial neural network model combines forensic dentistry and orthodontics to improve subject recognition by expanding the initial space of odontometric variables and adding orthodontic parameters.

Analysis of the effects of non-face-to-face SW·AI education for Pre-service teachers (예비교사 대상 비대면 SW·AI 교육 효과 분석)

  • Park, SunJu
    • 한국정보교육학회:학술대회논문집
    • /
    • 2021.08a
    • /
    • pp.315-320
    • /
    • 2021
  • In order to prepare for future social changes, SW·AI education is essential. In this paper, after conducting non-face-to-face SW·AI education for pre-service teachers, the effectiveness of SW education before and after education was measured using the measurement tool on the software educational effectiveness. As a result of the analysis, the overall average and the average of the 'computational thinking' and 'SW literacy' domains increased significantly, and the difference between the averages before and after education was statistically significant in decomposition, pattern recognition, abstraction, and algorithm, which are sub domains of 'computational thinking'. Through SW·AI education, students not only recognize the necessity of SW education and the importance of computational thinking, but also understand the process of decomposing information, recognizing and extracting patterns, and expressing problem-solving processes. It can be seen that non-face-to-face SW·AI education has the effect of improving computational thinking and SW literacy beyond recognizing the importance of SW.

  • PDF

A Study on the Use of Haar Cascade Filtering to check Wearing Masks and Fever Abnormality (Haar Cascade 필터링을 통한 마스크 착용 여부와 발열 체크)

  • Kim, Eui-Jeong;Kim, In-Jung
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.474-477
    • /
    • 2021
  • Recently, in order to prevent the proliferation of COVID-19, which began in earnest in 2020, an increasing number of places have been measuring the temperature and required to wear a mask. However, as wearing a mask and checking the temperature are typically measured directly by a person or by a single individual positioned in front of the machine, standards may vary based on the person's manual measurement method, wasting workforce. While standing in front of a device often measures the maximum temperature of the face, the standard of fever is also unclear. Both approaches can create bottleneck situations when checking large numbers of people. Furthermore, it is unable to conduct periodic measurements and tracking because the measuring machines are generally put only at the entrance. Thus, this study suggests a method for preventing the spread of infectious diseases by automatically identifying and displaying unmasked people and those with fever in real-time using a general camera, a thermal imaging camera, and an artificial intelligence algorithm.

  • PDF

Time delay estimation algorithm using Elastic Net (Elastic Net를 이용한 시간 지연 추정 알고리즘)

  • Jun-Seok Lim;Keunwa Lee
    • The Journal of the Acoustical Society of Korea
    • /
    • v.42 no.4
    • /
    • pp.364-369
    • /
    • 2023
  • Time-delay estimation between two receivers is a technique that has been applied in a variety of fields, from underwater acoustics to room acoustics and robotics. There are two types of time delay estimation techniques: one that estimates the amount of time delay from the correlation between receivers, and the other that parametrically models the time delay between receivers and estimates the parameters by system recognition. The latter has the characteristic that only a small fraction of the system's parameters are directly related to the delay. This characteristic can be exploited to improve the accuracy of the estimation by methods such as Lasso regularization. However, in the case of Lasso regularization, the necessary information is lost. In this paper, we propose a method using Elastic Net that adds Ridge regularization to Lasso regularization to compensate for this. Comparing the proposed method with the conventional Generalized Cross Correlation (GCC) method and the method using Lasso regularization, we show that the estimation variance is very small even for white Gaussian signal sources and colored signal sources.

Parameter Analysis for Super-Resolution Network Model Optimization of LiDAR Intensity Image (LiDAR 반사 강도 영상의 초해상화 신경망 모델 최적화를 위한 파라미터 분석)

  • Seungbo Shim
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.22 no.5
    • /
    • pp.137-147
    • /
    • 2023
  • LiDAR is used in autonomous driving and various industrial fields to measure the size and distance of an object. In addition, the sensor also provides intensity images based on the amount of reflected light. This has a positive effect on sensor data processing by providing information on the shape of the object. LiDAR guarantees higher performance as the resolution increases but at an increased cost. These conditions also apply to LiDAR intensity images. Expensive equipment is essential to acquire high-resolution LiDAR intensity images. This study developed artificial intelligence to improve low-resolution LiDAR intensity images into high-resolution ones. Therefore, this study performed parameter analysis for the optimal super-resolution neural network model. The super-resolution algorithm was trained and verified using 2,500 LiDAR intensity images. As a result, the resolution of the intensity images were improved. These results can be applied to the autonomous driving field and help improve driving environment recognition and obstacle detection performance

Study of Deep Learning Based Specific Person Following Mobility Control for Logistics Transportation (물류 이송을 위한 딥러닝 기반 특정 사람 추종 모빌리티 제어 연구)

  • Yeong Jun Yu;SeongHoon Kang;JuHwan Kim;SeongIn No;GiHyeon Lee;Seung Yong Lee;Chul-hee Lee
    • Journal of Drive and Control
    • /
    • v.20 no.4
    • /
    • pp.1-8
    • /
    • 2023
  • In recent years, robots have been utilized in various industries to reduce workload and enhance work efficiency. The following mobility offers users convenience by autonomously tracking specific locations and targets without the need for additional equipment such as forklifts or carts. In this paper, deep learning techniques were employed to recognize individuals and assign each of them a unique identifier to enable the recognition of a specific person even among multiple individuals. To achieve this, the distance and angle between the robot and the targeted individual are transmitted to respective controllers. Furthermore, this study explored the control methodology for mobility that tracks a specific person, utilizing Simultaneous Localization and Mapping (SLAM) and Proportional-Integral-Derivative (PID) control techniques. In the PID control method, a genetic algorithm is employed to extract the optimal gain value, subsequently evaluating PID performance through simulation. The SLAM method involves generating a map by synchronizing data from a 2D LiDAR and a depth camera using Real-Time Appearance-Based Mapping (RTAB-MAP). Experiments are conducted to compare and analyze the performance of the two control methods, visualizing the paths of both the human and the following mobility.

Voice Activity Detection Method Using Psycho-Acoustic Model Based on Speech Energy Maximization in Noisy Environments (잡음 환경에서 심리음향모델 기반 음성 에너지 최대화를 이용한 음성 검출 방법)

  • Choi, Gab-Keun;Kim, Soon-Hyob
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.5
    • /
    • pp.447-453
    • /
    • 2009
  • This paper introduces the method for detect voices and exact end point at low SNR by maximizing voice energy. Conventional VAD (Voice Activity Detection) algorithm estimates noise level so it tends to detect the end point inaccurately. Moreover, because it uses relatively long analysis range for reflecting temporal change of noise, computing load too high for application. In this paper, the SEM-VAD (Speech Energy Maximization-Voice Activity Detection) method which uses psycho-acoustical bark scale filter banks to maximize voice energy within frames is introduced. Stable threshold values are obtained at various noise environments (SNR 15 dB, 10 dB, 5 dB, 0 dB). At the test for voice detection in car noisy environment, PHR (Pause Hit Rate) was 100%accurate at every noise environment, and FAR (False Alarm Rate) shows 0% at SNR15 dB and 10 dB, 5.6% at SNR5 dB and 9.5% at SNR0 dB.

Object Pose Estimation and Motion Planning for Service Automation System (서비스 자동화 시스템을 위한 물체 자세 인식 및 동작 계획)

  • Youngwoo Kwon;Dongyoung Lee;Hosun Kang;Jiwook Choi;Inho Lee
    • The Journal of Korea Robotics Society
    • /
    • v.19 no.2
    • /
    • pp.176-187
    • /
    • 2024
  • Recently, automated solutions using collaborative robots have been emerging in various industries. Their primary functions include Pick & Place, Peg in the Hole, fastening and assembly, welding, and more, which are being utilized and researched in various fields. The application of these robots varies depending on the characteristics of the grippers attached to the end of the collaborative robots. To grasp a variety of objects, a gripper with a high degree of freedom is required. In this paper, we propose a service automation system using a multi-degree-of-freedom gripper, collaborative robots, and vision sensors. Assuming various products are placed at a checkout counter, we use three cameras to recognize the objects, estimate their pose, and create grasping points for grasping. The grasping points are grasped by the multi-degree-of-freedom gripper, and experiments are conducted to recognize barcodes, a key task in service automation. To recognize objects, we used a CNN (Convolutional Neural Network) based algorithm and point cloud to estimate the object's 6D pose. Using the recognized object's 6d pose information, we create grasping points for the multi-degree-of-freedom gripper and perform re-grasping in a direction that facilitates barcode scanning. The experiment was conducted with four selected objects, progressing through identification, 6D pose estimation, and grasping, recording the success and failure of barcode recognition to prove the effectiveness of the proposed system.

A Comparative Study of Deep Learning Techniques for Alzheimer's disease Detection in Medical Radiography

  • Amal Alshahrani;Jenan Mustafa;Manar Almatrafi;Layan Albaqami;Raneem Aljabri;Shahad Almuntashri
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.5
    • /
    • pp.53-63
    • /
    • 2024
  • Alzheimer's disease is a brain disorder that worsens over time and affects millions of people around the world. It leads to a gradual deterioration in memory, thinking ability, and behavioral and social skills until the person loses his ability to adapt to society. Technological progress in medical imaging and the use of artificial intelligence, has provided the possibility of detecting Alzheimer's disease through medical images such as magnetic resonance imaging (MRI). However, Deep learning algorithms, especially convolutional neural networks (CNNs), have shown great success in analyzing medical images for disease diagnosis and classification. Where CNNs can recognize patterns and objects from images, which makes them ideally suited for this study. In this paper, we proposed to compare the performances of Alzheimer's disease detection by using two deep learning methods: You Only Look Once (YOLO), a CNN-enabled object recognition algorithm, and Visual Geometry Group (VGG16) which is a type of deep convolutional neural network primarily used for image classification. We will compare our results using these modern models Instead of using CNN only like the previous research. In addition, the results showed different levels of accuracy for the various versions of YOLO and the VGG16 model. YOLO v5 reached 56.4% accuracy at 50 epochs and 61.5% accuracy at 100 epochs. YOLO v8, which is for classification, reached 84% accuracy overall at 100 epochs. YOLO v9, which is for object detection overall accuracy of 84.6%. The VGG16 model reached 99% accuracy for training after 25 epochs but only 78% accuracy for testing. Hence, the best model overall is YOLO v9, with the highest overall accuracy of 86.1%.

Accelerometer-based Gesture Recognition for Robot Interface (로봇 인터페이스 활용을 위한 가속도 센서 기반 제스처 인식)

  • Jang, Min-Su;Cho, Yong-Suk;Kim, Jae-Hong;Sohn, Joo-Chan
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.1
    • /
    • pp.53-69
    • /
    • 2011
  • Vision and voice-based technologies are commonly utilized for human-robot interaction. But it is widely recognized that the performance of vision and voice-based interaction systems is deteriorated by a large margin in the real-world situations due to environmental and user variances. Human users need to be very cooperative to get reasonable performance, which significantly limits the usability of the vision and voice-based human-robot interaction technologies. As a result, touch screens are still the major medium of human-robot interaction for the real-world applications. To empower the usability of robots for various services, alternative interaction technologies should be developed to complement the problems of vision and voice-based technologies. In this paper, we propose the use of accelerometer-based gesture interface as one of the alternative technologies, because accelerometers are effective in detecting the movements of human body, while their performance is not limited by environmental contexts such as lighting conditions or camera's field-of-view. Moreover, accelerometers are widely available nowadays in many mobile devices. We tackle the problem of classifying acceleration signal patterns of 26 English alphabets, which is one of the essential repertoires for the realization of education services based on robots. Recognizing 26 English handwriting patterns based on accelerometers is a very difficult task to take over because of its large scale of pattern classes and the complexity of each pattern. The most difficult problem that has been undertaken which is similar to our problem was recognizing acceleration signal patterns of 10 handwritten digits. Most previous studies dealt with pattern sets of 8~10 simple and easily distinguishable gestures that are useful for controlling home appliances, computer applications, robots etc. Good features are essential for the success of pattern recognition. To promote the discriminative power upon complex English alphabet patterns, we extracted 'motion trajectories' out of input acceleration signal and used them as the main feature. Investigative experiments showed that classifiers based on trajectory performed 3%~5% better than those with raw features e.g. acceleration signal itself or statistical figures. To minimize the distortion of trajectories, we applied a simple but effective set of smoothing filters and band-pass filters. It is well known that acceleration patterns for the same gesture is very different among different performers. To tackle the problem, online incremental learning is applied for our system to make it adaptive to the users' distinctive motion properties. Our system is based on instance-based learning (IBL) where each training sample is memorized as a reference pattern. Brute-force incremental learning in IBL continuously accumulates reference patterns, which is a problem because it not only slows down the classification but also downgrades the recall performance. Regarding the latter phenomenon, we observed a tendency that as the number of reference patterns grows, some reference patterns contribute more to the false positive classification. Thus, we devised an algorithm for optimizing the reference pattern set based on the positive and negative contribution of each reference pattern. The algorithm is performed periodically to remove reference patterns that have a very low positive contribution or a high negative contribution. Experiments were performed on 6500 gesture patterns collected from 50 adults of 30~50 years old. Each alphabet was performed 5 times per participant using $Nintendo{(R)}$ $Wii^{TM}$ remote. Acceleration signal was sampled in 100hz on 3 axes. Mean recall rate for all the alphabets was 95.48%. Some alphabets recorded very low recall rate and exhibited very high pairwise confusion rate. Major confusion pairs are D(88%) and P(74%), I(81%) and U(75%), N(88%) and W(100%). Though W was recalled perfectly, it contributed much to the false positive classification of N. By comparison with major previous results from VTT (96% for 8 control gestures), CMU (97% for 10 control gestures) and Samsung Electronics(97% for 10 digits and a control gesture), we could find that the performance of our system is superior regarding the number of pattern classes and the complexity of patterns. Using our gesture interaction system, we conducted 2 case studies of robot-based edutainment services. The services were implemented on various robot platforms and mobile devices including $iPhone^{TM}$. The participating children exhibited improved concentration and active reaction on the service with our gesture interface. To prove the effectiveness of our gesture interface, a test was taken by the children after experiencing an English teaching service. The test result showed that those who played with the gesture interface-based robot content marked 10% better score than those with conventional teaching. We conclude that the accelerometer-based gesture interface is a promising technology for flourishing real-world robot-based services and content by complementing the limits of today's conventional interfaces e.g. touch screen, vision and voice.