• Title/Summary/Keyword: algorithms

Search Result 16,314, Processing Time 0.048 seconds

IPC Multi-label Classification based on Functional Characteristics of Fields in Patent Documents (특허문서 필드의 기능적 특성을 활용한 IPC 다중 레이블 분류)

  • Lim, Sora;Kwon, YongJin
    • Journal of Internet Computing and Services
    • /
    • v.18 no.1
    • /
    • pp.77-88
    • /
    • 2017
  • Recently, with the advent of knowledge based society where information and knowledge make values, patents which are the representative form of intellectual property have become important, and the number of the patents follows growing trends. Thus, it needs to classify the patents depending on the technological topic of the invention appropriately in order to use a vast amount of the patent information effectively. IPC (International Patent Classification) is widely used for this situation. Researches about IPC automatic classification have been studied using data mining and machine learning algorithms to improve current IPC classification task which categorizes patent documents by hand. However, most of the previous researches have focused on applying various existing machine learning methods to the patent documents rather than considering on the characteristics of the data or the structure of patent documents. In this paper, therefore, we propose to use two structural fields, technical field and background, considered as having impacts on the patent classification, where the two field are selected by applying of the characteristics of patent documents and the role of the structural fields. We also construct multi-label classification model to reflect what a patent document could have multiple IPCs. Furthermore, we propose a method to classify patent documents at the IPC subclass level comprised of 630 categories so that we investigate the possibility of applying the IPC multi-label classification model into the real field. The effect of structural fields of patent documents are examined using 564,793 registered patents in Korea, and 87.2% precision is obtained in the case of using title, abstract, claims, technical field and background. From this sequence, we verify that the technical field and background have an important role in improving the precision of IPC multi-label classification in IPC subclass level.

Development of Control Algorithm for Greenhouse Cooling Using Two-fluid Fogging System (이류체 포그 냉방시스템의 제어알고리즘 개발)

  • Nam, Sang-Woon;Kim, Young-Shik;Sung, In-Mo
    • Journal of Bio-Environment Control
    • /
    • v.22 no.2
    • /
    • pp.138-145
    • /
    • 2013
  • In order to develop the efficient control algorithm of the two-fluid fogging system, cooling experiments for the many different types of fogging cycles were conducted in tomato greenhouses. It showed that the cooling effect was 1.2 to $4.0^{\circ}C$ and the cooling efficiency was 8.2 to 32.9% on average. The cooling efficiency with fogging interval was highest in the case of the fogging cycle of 90 seconds. The cooling efficiency showed a tendency to increase as the fogging time increased and the stopping time decreased. As the spray rate of fog in the two-fluid fogging system increased, there was a tendency for the cooling efficiency to improve. However, as the inside air approaches its saturation level, even though the spray rate of fog increases, it does not lead to further evaporation. Thus, it can be inferred that increasing the spray rate of fog before the inside air reaches the saturation level could make higher the cooling efficiency. As cooling efficiency increases, the saturation deficit of inside air decreased and the difference between absolute humidity of inside and outside air increased. The more fog evaporated, the difference between absolute humidity of inside and outside air tended to increase and as the result, the discharge of vapor due to ventilation occurs more easily, which again lead to an increase in the evaporation rate and ultimately increase in the cooling efficiency. Regression analysis result on the saturation deficit of inside air showed that the fogging time needed to change of saturation deficit of $10g{\cdot}kg^{-1}$ was 120 seconds and stopping time was 60 seconds. But in order to decrease the amplitude of temperature and to increase the cooling efficiency, the fluctuation range of saturation deficit was set to $5g{\cdot}kg^{-1}$ and we decided that the fogging-stopping time of 60-30 seconds was more appropriate. Control types of two-fluid fogging systems were classified as computer control or simple control, and their control algorithms were derived. We recommend that if the two-fluid fogging system is controlled by manipulating only the set point of temperature, humidity, and on-off time, it would be best to set up the on-off time at 60-30 seconds in time control, the lower limit of air temperature at 30 to $32^{\circ}C$ and the upper limit of relative humidity at 85 to 90%.

Quantitative Conductivity Estimation Error due to Statistical Noise in Complex $B_1{^+}$ Map (정량적 도전율측정의 오차와 $B_1{^+}$ map의 노이즈에 관한 분석)

  • Shin, Jaewook;Lee, Joonsung;Kim, Min-Oh;Choi, Narae;Seo, Jin Keun;Kim, Dong-Hyun
    • Investigative Magnetic Resonance Imaging
    • /
    • v.18 no.4
    • /
    • pp.303-313
    • /
    • 2014
  • Purpose : In-vivo conductivity reconstruction using transmit field ($B_1{^+}$) information of MRI was proposed. We assessed the accuracy of conductivity reconstruction in the presence of statistical noise in complex $B_1{^+}$ map and provided a parametric model of the conductivity-to-noise ratio value. Materials and Methods: The $B_1{^+}$ distribution was simulated for a cylindrical phantom model. By adding complex Gaussian noise to the simulated $B_1{^+}$ map, quantitative conductivity estimation error was evaluated. The quantitative evaluation process was repeated over several different parameters such as Larmor frequency, object radius and SNR of $B_1{^+}$ map. A parametric model for the conductivity-to-noise ratio was developed according to these various parameters. Results: According to the simulation results, conductivity estimation is more sensitive to statistical noise in $B_1{^+}$ phase than to noise in $B_1{^+}$ magnitude. The conductivity estimate of the object of interest does not depend on the external object surrounding it. The conductivity-to-noise ratio is proportional to the signal-to-noise ratio of the $B_1{^+}$ map, Larmor frequency, the conductivity value itself and the number of averaged pixels. To estimate accurate conductivity value of the targeted tissue, SNR of $B_1{^+}$ map and adequate filtering size have to be taken into account for conductivity reconstruction process. In addition, the simulation result was verified at 3T conventional MRI scanner. Conclusion: Through all these relationships, quantitative conductivity estimation error due to statistical noise in $B_1{^+}$ map is modeled. By using this model, further issues regarding filtering and reconstruction algorithms can be investigated for MREPT.

A Study on Shape Optimization of Plane Truss Structures (평면(平面) 트러스 구조물(構造物)의 형상최적화(形狀最適化)에 관한 구연(究研))

  • Lee, Gyu won;Byun, Keun Joo;Hwang, Hak Joo
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.5 no.3
    • /
    • pp.49-59
    • /
    • 1985
  • Formulation of the geometric optimization for truss structures based on the elasticity theory turn out to be the nonlinear programming problem which has to deal with the Cross sectional area of the member and the coordinates of its nodes simultaneously. A few techniques have been proposed and adopted for the analysis of this nonlinear programming problem for the time being. These techniques, however, bear some limitations on truss shapes loading conditions and design criteria for the practical application to real structures. A generalized algorithm for the geometric optimization of the truss structures which can eliminate the above mentioned limitations, is developed in this study. The algorithm developed utilizes the two-phases technique. In the first phase, the cross sectional area of the truss member is optimized by transforming the nonlinear problem into SUMT, and solving SUMT utilizing the modified Newton-Raphson method. In the second phase, the geometric shape is optimized utilizing the unidirctional search technique of the Rosenbrock method which make it possible to minimize only the objective function. The algorithm developed in this study is numerically tested for several truss structures with various shapes, loading conditions and design criteria, and compared with the results of the other algorithms to examme its applicability and stability. The numerical comparisons show that the two-phases algorithm developed in this study is safely applicable to any design criteria, and the convergency rate is very fast and stable compared with other iteration methods for the geometric optimization of truss structures.

  • PDF

NEAR REAL-TIME ESTIMATION OF GEOMAGNETIC LOCAL K INDEX FROM GYEONGZU MAGNETOMETER (경주 지자기관측소 자료를 이용한 준실시간 K 지수 산출에 관한 연구)

  • Choi, K.C.;Cho, K.S.;Moon, Y.J.;Kim, K.H.;Lee, D.Y.;Park, Y.D.;Lim, M.T.;Park, Y.S.;Lim, H.R.
    • Journal of Astronomy and Space Sciences
    • /
    • v.22 no.4
    • /
    • pp.431-440
    • /
    • 2005
  • Local K-index is an indicator representing local geomagnetic activity in every 3 hour. For estimation of the local K-index, a reasonable determination of solar quiet curve (undisturbed daily variation of geomagnetic field) is quiet essential. To derive the solar quiet curve, the FMI method, which is one of representative algorithms, uses horizontal components (H and D) of 3 days magnetometer data from the previous day to the next day for a specific day. However, this method is not applicable to real time forecast since it always requires the next day data. In this study, we have devised a new method to estimate local K-index in near real-time by modifying the FMI method. The new method selects a recent quiet day whose $K_p$ indices, reported by NOAA/SEC are all lower than 3, and replace the previous day and the next day data by the recent quiet day data. We estimated 2,672 local K indices from Gyeongzu magnetometer in 2003, and then compared the indices with those from the conventional FMI method. We also compared the K indices with those from Kakioka observatory. As a result, we found that (1) K indices from the new method are nearly consistent with those of the conventional FMI method with a very high correlation (R=0.96); (2) onr local K indices also have a relatively high correlation (R=0.81) with those from Kakioka station. Our results show that the new method can be used for near real-time estimation of local K indices from Gyeongzu magnetometer.

Monitoring Ground-level SO2 Concentrations Based on a Stacking Ensemble Approach Using Satellite Data and Numerical Models (위성 자료와 수치모델 자료를 활용한 스태킹 앙상블 기반 SO2 지상농도 추정)

  • Choi, Hyunyoung;Kang, Yoojin;Im, Jungho;Shin, Minso;Park, Seohui;Kim, Sang-Min
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1053-1066
    • /
    • 2020
  • Sulfur dioxide (SO2) is primarily released through industrial, residential, and transportation activities, and creates secondary air pollutants through chemical reactions in the atmosphere. Long-term exposure to SO2 can result in a negative effect on the human body causing respiratory or cardiovascular disease, which makes the effective and continuous monitoring of SO2 crucial. In South Korea, SO2 monitoring at ground stations has been performed, but this does not provide spatially continuous information of SO2 concentrations. Thus, this research estimated spatially continuous ground-level SO2 concentrations at 1 km resolution over South Korea through the synergistic use of satellite data and numerical models. A stacking ensemble approach, fusing multiple machine learning algorithms at two levels (i.e., base and meta), was adopted for ground-level SO2 estimation using data from January 2015 to April 2019. Random forest and extreme gradient boosting were used as based models and multiple linear regression was adopted for the meta-model. The cross-validation results showed that the meta-model produced the improved performance by 25% compared to the base models, resulting in the correlation coefficient of 0.48 and root-mean-square-error of 0.0032 ppm. In addition, the temporal transferability of the approach was evaluated for one-year data which were not used in the model development. The spatial distribution of ground-level SO2 concentrations based on the proposed model agreed with the general seasonality of SO2 and the temporal patterns of emission sources.

Wildfire Severity Mapping Using Sentinel Satellite Data Based on Machine Learning Approaches (Sentinel 위성영상과 기계학습을 이용한 국내산불 피해강도 탐지)

  • Sim, Seongmun;Kim, Woohyeok;Lee, Jaese;Kang, Yoojin;Im, Jungho;Kwon, Chunguen;Kim, Sungyong
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1109-1123
    • /
    • 2020
  • In South Korea with forest as a major land cover class (over 60% of the country), many wildfires occur every year. Wildfires weaken the shear strength of the soil, forming a layer of soil that is vulnerable to landslides. It is important to identify the severity of a wildfire as well as the burned area to sustainably manage the forest. Although satellite remote sensing has been widely used to map wildfire severity, it is often difficult to determine the severity using only the temporal change of satellite-derived indices such as Normalized Difference Vegetation Index (NDVI) and Normalized Burn Ratio (NBR). In this study, we proposed an approach for determining wildfire severity based on machine learning through the synergistic use of Sentinel-1A Synthetic Aperture Radar-C data and Sentinel-2A Multi Spectral Instrument data. Three wildfire cases-Samcheok in May 2017, Gangreung·Donghae in April 2019, and Gosung·Sokcho in April 2019-were used for developing wildfire severity mapping models with three machine learning algorithms (i.e., Random Forest, Logistic Regression, and Support Vector Machine). The results showed that the random forest model yielded the best performance, resulting in an overall accuracy of 82.3%. The cross-site validation to examine the spatiotemporal transferability of the machine learning models showed that the models were highly sensitive to temporal differences between the training and validation sites, especially in the early growing season. This implies that a more robust model with high spatiotemporal transferability can be developed when more wildfire cases with different seasons and areas are added in the future.

Application of an empirical method to improve radar rainfall estimation using cross governmental dual-pol. radars (범부처 이중편파레이더의 강우 추정 향상을 위한 경험적 방법의 적용)

  • Yoon, Jungsoo;Suk, Mi-Kyung;Nam, Kyung-Yeub;Park, Jong-Sook
    • Journal of Korea Water Resources Association
    • /
    • v.49 no.7
    • /
    • pp.625-634
    • /
    • 2016
  • Three leading agencies under different ministries - Korea Meteorological Administration (KMA) in the ministry of Environment, Han river control office in the Ministry of Land, Infrastructure and Transport (MOLIT) and Weather Group of ROK Air Force in the Ministry of National Defense (MND) - have been operated radars in the purpose of observing weather, hydrology and military operational weather in Korea. Eight S-band dual-pol. radars have been newly installed or replaced by these ministries over different places by 2015. However each ministry has different aims of operating radars, observation strategies, data processing algorithms, etc. Due to the differences, there is a wide level of accuracy on observed radar data as well as the composite images made of the cross governmental radar measurement. Gaining fairly high level of accuracy on radar data obtained by different agencies has been shared as a great concern by the ministries. Thus, "an agreement of harmonizing weather and hydrological radar products" was made by the three ministries in 2010. Particularly, this is very important to produce better rainfall estimation using the cross governmental radar measurement. Weather Radar Center(WRC) in KMA has been developed an empirical method using measurements observed by Yongin testbed radar. This study is aiming to examine the efficiency of the empirical method to improve the accuracies of radar rainfalls estimated from cross governmental dual-pol. radar measurements. As a result, the radar rainfalls of three radars (Baengnyeongdo, Biseulsan, and, Sobaeksan Radar) were shown improvement in accuracy (1-NE) up to 70% using data from May to October in 2015. Also, the range of the accuracies in radar rainfall estimation, which were from 30% to 60% before adjusting polarimetric variables, were decreased from 65% to 70% after adjusting polarimetric variables.

The Effects of Discrepancy in Reconstruction Algorithm between Patient Data and Normal Database in AutoQuant Evaluation: Focusing on Half-Time Scan Algorithm in Myocardial SPECT (심근 관류 스펙트에서 Half-Time Scan과 새로운 재구성법이 적용된 정상군 데이터를 기반으로 한 정량적 분석 결과의 차이 비교)

  • Lee, Hyung-Jin;Do, Yong-Ho;Cho, Seong-Wook;Kim, Jin-Eui
    • The Korean Journal of Nuclear Medicine Technology
    • /
    • v.18 no.1
    • /
    • pp.122-126
    • /
    • 2014
  • Purpose: The new reconstruction algorithms (NRA) provided by vendor aim to shorten the acquisition scan time. Whereas depending on the installed version AutoQuant program used for myocardial SPECT quantitative analysis did not contain the normal data that NRA is applied. Thus, the purpose of this paper is to compare the results according to AutoQuant versions in myocardial SPECT applied NRA and half-time scan (HT). Materials and Methods: Rest Tl and stress MIBI data of total 80 (40 men, 40 women) patients were gathered. Data were applied HT acquisition and ASTONISH (Philips) software which is NRA. Modified autoquant of SNUH and old version of AutoQuant (full-time scan) provided by company were compared. Comparison groups were classified as coronary artery disease (CAD), 24 hrs delay and almost normal patients who have a simple pain patient. Perfusion distribution aspect, summed stress score (SSS), summed rest score (SRS), extent and total perfusion deficit (TPD) of each 25 patient who have above diseases were compared and evaluated. Results: The case of CAD, when using re-edited AutoQuant (HT) SSS and SRS showed about 30% reduction (P<0.0001), Extent showed about 38% reduction and TPD showed about 30% reduction in the tendency (P<0.0001). In the score of the perfusion, especially on the part of infero-medium, infero-apical, lateral-medium and lateral-apical regions were the biggest change. The case of the 24 hrs delay patient SRS (P=0.042), Extent (P=0.018) and TPD (P=0.0024) showed about 13-18% reduction. And the case of simple pain patient, comparison of 4 results showed about 5-7% reduction. Conclusion: This study was started based on expectation that results could be affected by normal patient data. Normal patient data is possible to change by race and gender. It was proved that combination of new reconstruction algorithm for reducing scan time and analysis program according to scan protocol with NRA could also be affected to results. Clinical usefulness of gated myocardial SPECT is possibly increased if each hospital properly collects normal patient data for their scan acquisition protocol.

  • PDF

The PRISM-based Rainfall Mapping at an Enhanced Grid Cell Resolution in Complex Terrain (복잡지형 고해상도 격자망에서의 PRISM 기반 강수추정법)

  • Chung, U-Ran;Yun, Kyung-Dahm;Cho, Kyung-Sook;Yi, Jae-Hyun;Yun, Jin-I.
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.11 no.2
    • /
    • pp.72-78
    • /
    • 2009
  • The demand for rainfall data in gridded digital formats has increased in recent years due to the close linkage between hydrological models and decision support systems using the geographic information system. One of the most widely used tools for digital rainfall mapping is the PRISM (parameter-elevation regressions on independent slopes model) which uses point data (rain gauge stations), a digital elevation model (DEM), and other spatial datasets to generate repeatable estimates of monthly and annual precipitation. In the PRISM, rain gauge stations are assigned with weights that account for other climatically important factors besides elevation, and aspects and the topographic exposure are simulated by dividing the terrain into topographic facets. The size of facet or grid cell resolution is determined by the density of rain gauge stations and a $5{\times}5km$ grid cell is considered as the lowest limit under the situation in Korea. The PRISM algorithms using a 270m DEM for South Korea were implemented in a script language environment (Python) and relevant weights for each 270m grid cell were derived from the monthly data from 432 official rain gauge stations. Weighted monthly precipitation data from at least 5 nearby stations for each grid cell were regressed to the elevation and the selected linear regression equations with the 270m DEM were used to generate a digital precipitation map of South Korea at 270m resolution. Among 1.25 million grid cells, precipitation estimates at 166 cells, where the measurements were made by the Korea Water Corporation rain gauge network, were extracted and the monthly estimation errors were evaluated. An average of 10% reduction in the root mean square error (RMSE) was found for any months with more than 100mm monthly precipitation compared to the RMSE associated with the original 5km PRISM estimates. This modified PRISM may be used for rainfall mapping in rainy season (May to September) at much higher spatial resolution than the original PRISM without losing the data accuracy.