• Title/Summary/Keyword: 지능형 알고리즘

Search Result 911, Processing Time 0.027 seconds

Proposal for Research Model of High-Function Patrol Robot using Integrated Sensor System (통합 센서 시스템을 이용한 고기능 순찰 로봇의 연구모델 제안)

  • Byeong-Cheon Yoo;Seung-Jung Shin
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.24 no.3
    • /
    • pp.77-85
    • /
    • 2024
  • In this dissertation, a we designed and implemented a patrol robot that integrates a thermal imaging camera, speed dome camera, PTZ camera, radar, lidar sensor, and smartphone. This robot has the ability to monitor and respond efficiently even in complex environments, and is especially designed to demonstrate high performance even at night or in low visibility conditions. An orbital movement system was selected for the robot's mobility, and a smartphone-based control system was developed for real-time data processing and decision-making. The combination of various sensors allows the robot to comprehensively perceive the environment and quickly detect hazards. Thermal imaging cameras are used for night surveillance, speed domes and PTZ cameras are used for wide-area monitoring, and radar and LIDAR are used for obstacle detection and avoidance. The smartphone-based control system provides a user-friendly interface. The proposed robot system can be used in various fields such as security, surveillance, and disaster response. Future research should include improving the robot's autonomous patrol algorithm, developing a multi-robot collaboration system, and long-term testing in a real environment. This study is expected to contribute to the development of the field of intelligent surveillance robots.

A Study on Atmospheric Data Anomaly Detection Algorithm based on Unsupervised Learning Using Adversarial Generative Neural Network (적대적 생성 신경망을 활용한 비지도 학습 기반의 대기 자료 이상 탐지 알고리즘 연구)

  • Yang, Ho-Jun;Lee, Seon-Woo;Lee, Mun-Hyung;Kim, Jong-Gu;Choi, Jung-Mu;Shin, Yu-mi;Lee, Seok-Chae;Kwon, Jang-Woo;Park, Ji-Hoon;Jung, Dong-Hee;Shin, Hye-Jung
    • Journal of Convergence for Information Technology
    • /
    • v.12 no.4
    • /
    • pp.260-269
    • /
    • 2022
  • In this paper, We propose an anomaly detection model using deep neural network to automate the identification of outliers of the national air pollution measurement network data that is previously performed by experts. We generated training data by analyzing missing values and outliers of weather data provided by the Institute of Environmental Research and based on the BeatGAN model of the unsupervised learning method, we propose a new model by changing the kernel structure, adding the convolutional filter layer and the transposed convolutional filter layer to improve anomaly detection performance. In addition, by utilizing the generative features of the proposed model to implement and apply a retraining algorithm that generates new data and uses it for training, it was confirmed that the proposed model had the highest performance compared to the original BeatGAN models and other unsupervised learning model like Iforest and One Class SVM. Through this study, it was possible to suggest a method to improve the anomaly detection performance of proposed model while avoiding overfitting without additional cost in situations where training data are insufficient due to various factors such as sensor abnormalities and inspections in actual industrial sites.

Analytical Evaluation of PPG Blood Glucose Monitoring System - researcher clinical trial (PPG 혈당 모니터링 시스템의 분석적 평가 - 연구자 임상)

  • Cheol-Gu Park;Sang-Ki Choi;Seong-Geun Jo;Kwon-Min Kim
    • Journal of Digital Convergence
    • /
    • v.21 no.3
    • /
    • pp.33-39
    • /
    • 2023
  • This study is a performance evaluation of a blood sugar monitoring system that combines a PPG sensor, which is an evaluation device for blood glucose monitoring, and a DNN algorithm when monitoring capillary blood glucose. The study is a researcher-led clinical trial conducted on participants from September 2023 to November 2023. PPG-BGMS compared predicted blood sugar levels for evaluation using 1-minute heart rate and heart rate variability information and the DNN prediction algorithm with capillary blood glucose levels measured with a blood glucose meter of the standard personal blood sugar management system. Of the 100 participants, 50 had type 2 diabetes (T2DM), and the average age was 67 years (range, 28 to 89 years). It was found that 100% of the predicted blood sugar level of PPG-BGMS was distributed in the A+B area of the Clarke error grid and Parker(Consensus) error grid. The MARD value of PPG-BGMS predicted blood glucose is 5.3 ± 4.0%. Consequentially, the non-blood-based PPG-BGMS was found to be non-inferior to the instantaneous blood sugar level of the clinical standard blood-based personal blood glucose measurement system.

Ontology-based User Customized Search Service Considering User Intention (온톨로지 기반의 사용자 의도를 고려한 맞춤형 검색 서비스)

  • Kim, Sukyoung;Kim, Gunwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.4
    • /
    • pp.129-143
    • /
    • 2012
  • Recently, the rapid progress of a number of standardized web technologies and the proliferation of web users in the world bring an explosive increase of producing and consuming information documents on the web. In addition, most companies have produced, shared, and managed a huge number of information documents that are needed to perform their businesses. They also have discretionally raked, stored and managed a number of web documents published on the web for their business. Along with this increase of information documents that should be managed in the companies, the need of a solution to locate information documents more accurately among a huge number of information sources have increased. In order to satisfy the need of accurate search, the market size of search engine solution market is becoming increasingly expended. The most important functionality among much functionality provided by search engine is to locate accurate information documents from a huge information sources. The major metric to evaluate the accuracy of search engine is relevance that consists of two measures, precision and recall. Precision is thought of as a measure of exactness, that is, what percentage of information considered as true answer are actually such, whereas recall is a measure of completeness, that is, what percentage of true answer are retrieved as such. These two measures can be used differently according to the applied domain. If we need to exhaustively search information such as patent documents and research papers, it is better to increase the recall. On the other hand, when the amount of information is small scale, it is better to increase precision. Most of existing web search engines typically uses a keyword search method that returns web documents including keywords which correspond to search words entered by a user. This method has a virtue of locating all web documents quickly, even though many search words are inputted. However, this method has a fundamental imitation of not considering search intention of a user, thereby retrieving irrelevant results as well as relevant ones. Thus, it takes additional time and effort to set relevant ones out from all results returned by a search engine. That is, keyword search method can increase recall, while it is difficult to locate web documents which a user actually want to find because it does not provide a means of understanding the intention of a user and reflecting it to a progress of searching information. Thus, this research suggests a new method of combining ontology-based search solution with core search functionalities provided by existing search engine solutions. The method enables a search engine to provide optimal search results by inferenceing the search intention of a user. To that end, we build an ontology which contains concepts and relationships among them in a specific domain. The ontology is used to inference synonyms of a set of search keywords inputted by a user, thereby making the search intention of the user reflected into the progress of searching information more actively compared to existing search engines. Based on the proposed method we implement a prototype search system and test the system in the patent domain where we experiment on searching relevant documents associated with a patent. The experiment shows that our system increases the both recall and precision in accuracy and augments the search productivity by using improved user interface that enables a user to interact with our search system effectively. In the future research, we will study a means of validating the better performance of our prototype system by comparing other search engine solution and will extend the applied domain into other domains for searching information such as portal.

New transfer standard for low vacuum region

  • 우삼용;한승웅;김부식;이상균
    • Proceedings of the Korean Vacuum Society Conference
    • /
    • 1999.07a
    • /
    • pp.44-44
    • /
    • 1999
  • 저진공(1 kPa~ 100 kPa)은 대기압 측정, 비행고도, 기체의 온도 측정, 질량의 부력 보정, 레이저의 굴절률 측정등에 사용되는 영역으로 과학적 중요성을 갖고 있다. 또한 대기압 이상의 압력 측정과 고진공 측정의 경계적 역할도 수행하고 있어 압력 표준기의 국제 비교에 필수적으로 권장되는 역역이다. 이 영역에 주로 사용되는 압력 표준기는 수은 압력계(Mercury manometer)와 분동식 압력계(Deadweight piston gauge or Pressure)가 있다. 이들은 이동이 불편하거나 불가능하므로 표준기의 국제 비교에 사용되는 전달 표준기로는 보다 이동이 간편한 탄성 압력계인 CDG(Capacitance diaphragm Gauge)가 있다. 이 게이지는 반도체 산업의 공정 제어용으로도 많이 사용되고 있다. 그러나 게이지와 함께 사용되는 컨트롤러의 부피가 크고 무거우며 영점 이동이 커서 측정때 마다 재조정하여야 하는 단점이 있다. 본 논문에서는 이 같은 단점을 극복하기 위해 수정빔 진동형 진공 센서를 잔달 표준기로 사용하는 것에 대한 연구를 수행하였다. 수정빔 진동형 압력 센서는 수정빔으 공진주파수가 스트레인에 비례하는 것을 이용하여 제작된 센서로 주로 대기압 이상의 고압 측정에 많이 사용되고 있다. 먼저 수정빔의 압력과 주파수간의 관계를 측정하고 또한 내장된 수정 온도센서의 공진 주파수를 측정하여 온도 보상을 위한 자료로 사용하였다. 규격에 나와 있는 수정빔의 기하학적 형상으로부터 거동에 관한 이론 모델식을 구하고 압력교정 자료로부터 얻어진 데이터를 이 식과 비교 분석하여 적합한 특성식과 인자를 구하였으며 게이지의 불확도를 추정하였다.모델은 길이가 유한한 0-차원 실린더 모델로 가정하였고, 이에 대한 기하학적 성질 및 열역학적 성질은 유효계수를 고려하여 산출하였다. 진공용기 이중 벽 내부로 흐르는 질소가스의 유량과 온도의 계산은 진공용기 내벽과 외벽을 각각 독립적인 열전달 요소로 가정하여 구성한 모델을 이용하였다. 전체 해석에서 각 열전달 요소의 비열 값은 온도에 따라 변화하는 비열의 특성을 반영하였으며. 진공용기와 플라즈마 대향 부품의 방사율(emissivity)은 앞서 가정했던 각 온도 상승 곡선에 대해서 각각 0.1, 0.2, 1.3의 경우를 가정하여 계산하였다. 직선적으로 증가하는 온도 상승 곡선중 2$0^{\circ}C$/hr의 온도상승율을 갖는 경우가 다른 베이킹 시나리오 모델에 비해 효과적이라 생각되며 초대 필요 공급열량은 200kW 정도로 산출되었다. 실질적인 수치를 얻기 위해 보다 고차원 모델로의 해석이 필요하리라 생각된다. 끝으로 장기적인 관점에서 KSTAR 장치의 베이킹 계획도 살펴본다.습파라미터와 더불어, 본 연구에서 새롭게 제시된 주기분할층의 파라미터들이 모형의 학습성과를 높이기 위해 함께 고려된다. 한편, 이러한 학습과정에서 추가적으로 고려해야 할 파라미터 갯수가 증가함에 따라서, 본 모델의 학습성과가 local minimum에 빠지는 문제점이 발생될 수 있다. 즉, 웨이블릿분석과 인공신경망모형을 모두 전역적으로 최적화시켜야 하는 문제가 발생한다. 본 연구에서는 이 문제를 해결하기 위해서, 최근 local minimum의 가능성을 최소화하여 전역적인 학습성과를 높여 주는 인공지능기법으로서 유전자알고리즘기법을 본 연구이 통합모델에 반영하였다. 이에 대한 실

  • PDF

A Research on Ball-Balancing Robot (볼 벨런싱 로봇에 관한 연구)

  • Kim, Ji-Tae;Kim, Dae-young;Lee, Won-Joon;Jin, Tae-Seok
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2017.05a
    • /
    • pp.463-466
    • /
    • 2017
  • The purpose of this paper is to develop a module capable of all-directional driving different from conventional wheeled robots, and to solve the problems of the conventional mobile robot with side driving performance degradation, It is possible to overcome the disadvantages such as an increase in the time required for the unnecessary driving. The all - direction spherical wheel drive module for driving a ball - balancing robot is required to develop a power transfer mechanism and a driving algorithm for driving the robot in all directions using three rotor casters. 3DoF (Axis) A driver with built-in forward motion algorithm is embedded in the module and a driving motor module with 3DoF (axis) for driving direction and speed is installed. The movement mechanism depends on the sum of the rotation vectors of the respective driving wheels. It is possible to create various movement directions depending on the rotation and the vector sum of two or three drive wheels. It is possible to move in different directions according to the rotation vector field of each driving wheel. When a more innovative all-round spherical wheel drive module for forward movement is developed, it can be used in the driving part of the mobile robot to improve the performance of the robot more technically, and through the forward-direction robot platform with the drive module Conventional wheeled robots can overcome the disadvantage that the continuous straightening performance is lowered due to resistance to various environments. Therefore, it is necessary to use a full-direction driving function as well as a cleaning robot and a mobile robot applicable in the Americas and Europe It will be an essential technology for guide robots, boarding robots, mobile means, etc., and will contribute to the expansion of the intelligent service robot market and future automobile market.

  • PDF

Recognition of Resident Registration Card using ART2-based RBF Network and face Verification (ART2 기반 RBF 네트워크와 얼굴 인증을 이용한 주민등록증 인식)

  • Kim Kwang-Baek;Kim Young-Ju
    • Journal of Intelligence and Information Systems
    • /
    • v.12 no.1
    • /
    • pp.1-15
    • /
    • 2006
  • In Korea, a resident registration card has various personal information such as a present address, a resident registration number, a face picture and a fingerprint. A plastic-type resident card currently used is easy to forge or alter and tricks of forgery grow to be high-degree as time goes on. So, whether a resident card is forged or not is difficult to judge by only an examination with the naked eye. This paper proposed an automatic recognition method of a resident card which recognizes a resident registration number by using a refined ART2-based RBF network newly proposed and authenticates a face picture by a template image matching method. The proposed method, first, extracts areas including a resident registration number and the date of issue from a resident card image by applying Sobel masking, median filtering and horizontal smearing operations to the image in turn. To improve the extraction of individual codes from extracted areas, the original image is binarized by using a high-frequency passing filter and CDM masking is applied to the binaried image fur making image information of individual codes better. Lastly, individual codes, which are targets of recognition, are extracted by applying 4-directional contour tracking algorithm to extracted areas in the binarized image. And this paper proposed a refined ART2-based RBF network to recognize individual codes, which applies ART2 as the loaming structure of the middle layer and dynamicaly adjusts a teaming rate in the teaming of the middle and the output layers by using a fuzzy control method to improve the performance of teaming. Also, for the precise judgement of forgey of a resident card, the proposed method supports a face authentication by using a face template database and a template image matching method. For performance evaluation of the proposed method, this paper maked metamorphoses of an original image of resident card such as a forgey of face picture, an addition of noise, variations of contrast variations of intensity and image blurring, and applied these images with original images to experiments. The results of experiment showed that the proposed method is excellent in the recognition of individual codes and the face authentication fur the automatic recognition of a resident card.

  • PDF

A Study on Risk Parity Asset Allocation Model with XGBoos (XGBoost를 활용한 리스크패리티 자산배분 모형에 관한 연구)

  • Kim, Younghoon;Choi, HeungSik;Kim, SunWoong
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.1
    • /
    • pp.135-149
    • /
    • 2020
  • Artificial intelligences are changing world. Financial market is also not an exception. Robo-Advisor is actively being developed, making up the weakness of traditional asset allocation methods and replacing the parts that are difficult for the traditional methods. It makes automated investment decisions with artificial intelligence algorithms and is used with various asset allocation models such as mean-variance model, Black-Litterman model and risk parity model. Risk parity model is a typical risk-based asset allocation model which is focused on the volatility of assets. It avoids investment risk structurally. So it has stability in the management of large size fund and it has been widely used in financial field. XGBoost model is a parallel tree-boosting method. It is an optimized gradient boosting model designed to be highly efficient and flexible. It not only makes billions of examples in limited memory environments but is also very fast to learn compared to traditional boosting methods. It is frequently used in various fields of data analysis and has a lot of advantages. So in this study, we propose a new asset allocation model that combines risk parity model and XGBoost machine learning model. This model uses XGBoost to predict the risk of assets and applies the predictive risk to the process of covariance estimation. There are estimated errors between the estimation period and the actual investment period because the optimized asset allocation model estimates the proportion of investments based on historical data. these estimated errors adversely affect the optimized portfolio performance. This study aims to improve the stability and portfolio performance of the model by predicting the volatility of the next investment period and reducing estimated errors of optimized asset allocation model. As a result, it narrows the gap between theory and practice and proposes a more advanced asset allocation model. In this study, we used the Korean stock market price data for a total of 17 years from 2003 to 2019 for the empirical test of the suggested model. The data sets are specifically composed of energy, finance, IT, industrial, material, telecommunication, utility, consumer, health care and staple sectors. We accumulated the value of prediction using moving-window method by 1,000 in-sample and 20 out-of-sample, so we produced a total of 154 rebalancing back-testing results. We analyzed portfolio performance in terms of cumulative rate of return and got a lot of sample data because of long period results. Comparing with traditional risk parity model, this experiment recorded improvements in both cumulative yield and reduction of estimated errors. The total cumulative return is 45.748%, about 5% higher than that of risk parity model and also the estimated errors are reduced in 9 out of 10 industry sectors. The reduction of estimated errors increases stability of the model and makes it easy to apply in practical investment. The results of the experiment showed improvement of portfolio performance by reducing the estimated errors of the optimized asset allocation model. Many financial models and asset allocation models are limited in practical investment because of the most fundamental question of whether the past characteristics of assets will continue into the future in the changing financial market. However, this study not only takes advantage of traditional asset allocation models, but also supplements the limitations of traditional methods and increases stability by predicting the risks of assets with the latest algorithm. There are various studies on parametric estimation methods to reduce the estimated errors in the portfolio optimization. We also suggested a new method to reduce estimated errors in optimized asset allocation model using machine learning. So this study is meaningful in that it proposes an advanced artificial intelligence asset allocation model for the fast-developing financial markets.

A Proposal of a Keyword Extraction System for Detecting Social Issues (사회문제 해결형 기술수요 발굴을 위한 키워드 추출 시스템 제안)

  • Jeong, Dami;Kim, Jaeseok;Kim, Gi-Nam;Heo, Jong-Uk;On, Byung-Won;Kang, Mijung
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.3
    • /
    • pp.1-23
    • /
    • 2013
  • To discover significant social issues such as unemployment, economy crisis, social welfare etc. that are urgent issues to be solved in a modern society, in the existing approach, researchers usually collect opinions from professional experts and scholars through either online or offline surveys. However, such a method does not seem to be effective from time to time. As usual, due to the problem of expense, a large number of survey replies are seldom gathered. In some cases, it is also hard to find out professional persons dealing with specific social issues. Thus, the sample set is often small and may have some bias. Furthermore, regarding a social issue, several experts may make totally different conclusions because each expert has his subjective point of view and different background. In this case, it is considerably hard to figure out what current social issues are and which social issues are really important. To surmount the shortcomings of the current approach, in this paper, we develop a prototype system that semi-automatically detects social issue keywords representing social issues and problems from about 1.3 million news articles issued by about 10 major domestic presses in Korea from June 2009 until July 2012. Our proposed system consists of (1) collecting and extracting texts from the collected news articles, (2) identifying only news articles related to social issues, (3) analyzing the lexical items of Korean sentences, (4) finding a set of topics regarding social keywords over time based on probabilistic topic modeling, (5) matching relevant paragraphs to a given topic, and (6) visualizing social keywords for easy understanding. In particular, we propose a novel matching algorithm relying on generative models. The goal of our proposed matching algorithm is to best match paragraphs to each topic. Technically, using a topic model such as Latent Dirichlet Allocation (LDA), we can obtain a set of topics, each of which has relevant terms and their probability values. In our problem, given a set of text documents (e.g., news articles), LDA shows a set of topic clusters, and then each topic cluster is labeled by human annotators, where each topic label stands for a social keyword. For example, suppose there is a topic (e.g., Topic1 = {(unemployment, 0.4), (layoff, 0.3), (business, 0.3)}) and then a human annotator labels "Unemployment Problem" on Topic1. In this example, it is non-trivial to understand what happened to the unemployment problem in our society. In other words, taking a look at only social keywords, we have no idea of the detailed events occurring in our society. To tackle this matter, we develop the matching algorithm that computes the probability value of a paragraph given a topic, relying on (i) topic terms and (ii) their probability values. For instance, given a set of text documents, we segment each text document to paragraphs. In the meantime, using LDA, we can extract a set of topics from the text documents. Based on our matching process, each paragraph is assigned to a topic, indicating that the paragraph best matches the topic. Finally, each topic has several best matched paragraphs. Furthermore, assuming there are a topic (e.g., Unemployment Problem) and the best matched paragraph (e.g., Up to 300 workers lost their jobs in XXX company at Seoul). In this case, we can grasp the detailed information of the social keyword such as "300 workers", "unemployment", "XXX company", and "Seoul". In addition, our system visualizes social keywords over time. Therefore, through our matching process and keyword visualization, most researchers will be able to detect social issues easily and quickly. Through this prototype system, we have detected various social issues appearing in our society and also showed effectiveness of our proposed methods according to our experimental results. Note that you can also use our proof-of-concept system in http://dslab.snu.ac.kr/demo.html.

VKOSPI Forecasting and Option Trading Application Using SVM (SVM을 이용한 VKOSPI 일 중 변화 예측과 실제 옵션 매매에의 적용)

  • Ra, Yun Seon;Choi, Heung Sik;Kim, Sun Woong
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.4
    • /
    • pp.177-192
    • /
    • 2016
  • Machine learning is a field of artificial intelligence. It refers to an area of computer science related to providing machines the ability to perform their own data analysis, decision making and forecasting. For example, one of the representative machine learning models is artificial neural network, which is a statistical learning algorithm inspired by the neural network structure of biology. In addition, there are other machine learning models such as decision tree model, naive bayes model and SVM(support vector machine) model. Among the machine learning models, we use SVM model in this study because it is mainly used for classification and regression analysis that fits well to our study. The core principle of SVM is to find a reasonable hyperplane that distinguishes different group in the data space. Given information about the data in any two groups, the SVM model judges to which group the new data belongs based on the hyperplane obtained from the given data set. Thus, the more the amount of meaningful data, the better the machine learning ability. In recent years, many financial experts have focused on machine learning, seeing the possibility of combining with machine learning and the financial field where vast amounts of financial data exist. Machine learning techniques have been proved to be powerful in describing the non-stationary and chaotic stock price dynamics. A lot of researches have been successfully conducted on forecasting of stock prices using machine learning algorithms. Recently, financial companies have begun to provide Robo-Advisor service, a compound word of Robot and Advisor, which can perform various financial tasks through advanced algorithms using rapidly changing huge amount of data. Robo-Adviser's main task is to advise the investors about the investor's personal investment propensity and to provide the service to manage the portfolio automatically. In this study, we propose a method of forecasting the Korean volatility index, VKOSPI, using the SVM model, which is one of the machine learning methods, and applying it to real option trading to increase the trading performance. VKOSPI is a measure of the future volatility of the KOSPI 200 index based on KOSPI 200 index option prices. VKOSPI is similar to the VIX index, which is based on S&P 500 option price in the United States. The Korea Exchange(KRX) calculates and announce the real-time VKOSPI index. VKOSPI is the same as the usual volatility and affects the option prices. The direction of VKOSPI and option prices show positive relation regardless of the option type (call and put options with various striking prices). If the volatility increases, all of the call and put option premium increases because the probability of the option's exercise possibility increases. The investor can know the rising value of the option price with respect to the volatility rising value in real time through Vega, a Black-Scholes's measurement index of an option's sensitivity to changes in the volatility. Therefore, accurate forecasting of VKOSPI movements is one of the important factors that can generate profit in option trading. In this study, we verified through real option data that the accurate forecast of VKOSPI is able to make a big profit in real option trading. To the best of our knowledge, there have been no studies on the idea of predicting the direction of VKOSPI based on machine learning and introducing the idea of applying it to actual option trading. In this study predicted daily VKOSPI changes through SVM model and then made intraday option strangle position, which gives profit as option prices reduce, only when VKOSPI is expected to decline during daytime. We analyzed the results and tested whether it is applicable to real option trading based on SVM's prediction. The results showed the prediction accuracy of VKOSPI was 57.83% on average, and the number of position entry times was 43.2 times, which is less than half of the benchmark (100 times). A small number of trading is an indicator of trading efficiency. In addition, the experiment proved that the trading performance was significantly higher than the benchmark.