• Title/Summary/Keyword: 데이터매칭

Search Result 683, Processing Time 0.029 seconds

Analyzing The Economic Impact of The Fire Risk Reduction at Regional Level in Goyang City (지역단위 화재 위험도 저감의 고양시 경제적 파급효과 분석)

  • Son, Minsu;Cho, Dongin;Park, Chang Keun;Ko, Hyun A;Jung, Seunghyun
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.22 no.2
    • /
    • pp.685-693
    • /
    • 2021
  • This study examined the fire risk of the region in Goyang City using the spatial information data of buildings. The economic damage by industry was assessed according to the probability of fire risk. The study area was confined to Goyang-si, Gyeonggi-do, and the same fire risk reduction rate was applied to each region for the convenience of analysis. The possibility of fire was derived based on the buildings' density and usage in the area by National GIS building-integrated information standard data. The calculation of economic damage by industry in Goyang City due to the fire risk was calculated by combining the Goyang-si industry-related model produced by matching with 30 industrial categories in Input-Output Statistics of Korea Bank and 20 industrial categories in the Goyang-si business survey and the possibility of fire. The basic scenario of production impossibility during six months and business loss due to fire was established and analyzed based on the supply model. The analysis showed that Ilsan-dong-gu, Ilsan-seo-gu, and Deokyang-gu suffered the most economic damage. The "electricity, gas, steam, and water business" showed the greatest loss by industry.

Research on Longitudinal Slope Estimation Using Digital Elevation Model (수치표고모델 정보를 활용한 도로 종단경사 산출 연구)

  • Han, Yohee;Jung, Yeonghun;Chun, Uibum;Kim, Youngchan;Park, Shin Hyoung
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.20 no.6
    • /
    • pp.84-99
    • /
    • 2021
  • As the micro-mobility market grows, the demand for route guidance, that includes uphill information as well, is increasing. Since the climbing angle depends on the electric motor uesed, it is necessary to establish an uphill road DB according to the threshold standard. Although road alignment information is a very important element in the basic information of the roads, there is no information currently on the longitudinal slope in the road digital map. The High Definition(HD) map which is being built as a preparation for the era of autonomous vehicles has the altitude value, unlike the existing standard node link system. However, the HD map is very insufficient because it has the altitude value only for some sections of the road network. This paper, hence, intends to propose a method to generate the road longitudinal slope using currently available data. We developed a method of computing the longitudinal slope by combining the digital elevation model and the standard link system. After creating an altitude at the road link point divided by 4m based on the Seoul road network, we calculated individual slope per unit distance of the road. After designating a representative slope for each road link, we have extracted the very steep road that cannot be climbed with personal mobility and the slippery roads that cannot be used during heavy snowfall. We additionally described errors in the altitude values due to surrounding terrain and the issues related to the slope calculation method. In the future, we expect that the road longitudinal slope information will be used as basic data that can be used for various convergence analyses.

The Moderating Effect of Self-rated Health on the Association between Grandparenting and Depressive Symptoms among Grandparents (손자녀 양육이 조부모의 우울감에 미치는 영향에 대한 주관적 건강상태의 조절효과)

  • Song, Si Young;Jun, Hey Jung;Joo, Susanna
    • 한국노년학
    • /
    • v.40 no.3
    • /
    • pp.459-475
    • /
    • 2020
  • This study examined the moderating effect self-rated health has on the association between grandparenting and depressive symptoms. The fourth wave (in 2012) of the Korean Longitudinal Study of Ageing (KLoSA) was used for the analyses (N=650). The Coarsened Exact Matching (CEM) method was applied in order to ensure randomness in the selection of grandparents who participated in grandparenting and those who did not. The dependent variable was depressive symptoms, the independent variable was grandparenting, and the moderating variable was self-rated health. Control variables were education level, age, household income, gender, marital status and life satisfaction. After matching data by CEM, weighted multiple regression was applied with STATA 13.0. Also, simple slope analysis and region of significance were performed to interpret the interaction terms. The results showed that self-rated health had a significant moderating effect. Specifically, for grandparents with high self-rated health, grandparenting was associated with a lower level of depressive symptoms. On the contrary, for grandparents with low self-rated health, there was no difference depending on whether they participated in raising grandchildren. Based on these results, we discussed the need for consideration of self-rated health for grandparents who participated in grandparenting.

An Analysis of 3D Mesh Accuracy and Completeness of Combination of Drone and Smartphone Images for Building 3D Modeling (건물3D모델링을 위한 드론과 스마트폰영상 조합의 3D메쉬 정확도 및 완성도 분석)

  • Han, Seung-Hee;Yoo, Sang-Hyeon
    • Journal of Cadastre & Land InformatiX
    • /
    • v.52 no.1
    • /
    • pp.69-80
    • /
    • 2022
  • Drone photogrammetry generally acquires images vertically or obliquely from above, so when photographing for the purpose of three-dimensional modeling, image matching for the ground of a building and spatial accuracy of point cloud data are poor, resulting in poor 3D mesh completeness. Therefore, to overcome this, this study analyzed the spatial accuracy of each drone image by acquiring smartphone images from the ground, and evaluated the accuracy improvement and completeness of 3D mesh when the smartphone image is not combined with the drone image. As a result of the study, the horizontal (x,y) accuracy of drone photogrammetry was about 1/200,000, similar to that of traditional photogrammetry. In addition, it was analyzed that the accuracy according to the photographing method was more affected by the photographing angle of the object than the increase in the number of photos. In the case of the smartphone image combination, the accuracy was not significantly affected, but the completeness of the 3D mesh was able to obtain a 3D mesh of about LoD3 that satisfies the digital twin city standard. Therefore, it is judged that it can be sufficiently used to build a 3D model for digital twin city by combining drone images and smartphones or DSLR images taken on the ground.

Deep Learning Similarity-based 1:1 Matching Method for Real Product Image and Drawing Image

  • Han, Gi-Tae
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.12
    • /
    • pp.59-68
    • /
    • 2022
  • This paper presents a method for 1:1 verification by comparing the similarity between the given real product image and the drawing image. The proposed method combines two existing CNN-based deep learning models to construct a Siamese Network. After extracting the feature vector of the image through the FC (Fully Connected) Layer of each network and comparing the similarity, if the real product image and the drawing image (front view, left and right side view, top view, etc) are the same product, the similarity is set to 1 for learning and, if it is a different product, the similarity is set to 0. The test (inference) model is a deep learning model that queries the real product image and the drawing image in pairs to determine whether the pair is the same product or not. In the proposed model, through a comparison of the similarity between the real product image and the drawing image, if the similarity is greater than or equal to a threshold value (Threshold: 0.5), it is determined that the product is the same, and if it is less than or equal to, it is determined that the product is a different product. The proposed model showed an accuracy of about 71.8% for a query to a product (positive: positive) with the same drawing as the real product, and an accuracy of about 83.1% for a query to a different product (positive: negative). In the future, we plan to conduct a study to improve the matching accuracy between the real product image and the drawing image by combining the parameter optimization study with the proposed model and adding processes such as data purification.

A Study on the Development of an Indoor Positioning Support System for Providing Landmark Information (랜드마크 정보 제공을 위한 실내위치측위 지원 시스템 구축에 관한 연구)

  • Ock-Woo NAM;Chang-Soo SHIN;Yun-Soo CHOI
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.26 no.4
    • /
    • pp.130-144
    • /
    • 2023
  • Recently, various positioning technologies are being researched based on signal-based positioning and image-based positioning to obtain accurate indoor location information. Among these, various studies are being conducted on image positioning technology that determines the location of a mobile terminal using images acquired through cameras and sensor data collected as needed. For video-based positioning, a method of determining indoor location is used by matching mobile terminal photos with virtual landmark images, and for this purpose, it is necessary to build indoor spatial information about various landmarks such as billboards, vending machines, and ATM machines. In order to construct indoor spatial information on various landmarks, a panoramic image in the form of a road view and accurate 3D survey results were obtained through c 13 buildings of the Electronics and Telecommunications Research Institute(ETRI). When comparing the 3D total station final result and the terrestrial lidar panoramic image coordinates, the coordinates and distance performance were obtained within about 0.10m, confirming that accurate landmark construction for use in indoor positioning was possible. By utilizing these terrestrial lidar achievements to perform 3D landmark modeling necessary for image positioning, it was possible to more quickly model landmark information that could not be constructed only through 3D modeling using existing as-built drawings.

Development of Intelligent Job Classification System based on Job Posting on Job Sites (구인구직사이트의 구인정보 기반 지능형 직무분류체계의 구축)

  • Lee, Jung Seung
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.123-139
    • /
    • 2019
  • The job classification system of major job sites differs from site to site and is different from the job classification system of the 'SQF(Sectoral Qualifications Framework)' proposed by the SW field. Therefore, a new job classification system is needed for SW companies, SW job seekers, and job sites to understand. The purpose of this study is to establish a standard job classification system that reflects market demand by analyzing SQF based on job offer information of major job sites and the NCS(National Competency Standards). For this purpose, the association analysis between occupations of major job sites is conducted and the association rule between SQF and occupation is conducted to derive the association rule between occupations. Using this association rule, we proposed an intelligent job classification system based on data mapping the job classification system of major job sites and SQF and job classification system. First, major job sites are selected to obtain information on the job classification system of the SW market. Then We identify ways to collect job information from each site and collect data through open API. Focusing on the relationship between the data, filtering only the job information posted on each job site at the same time, other job information is deleted. Next, we will map the job classification system between job sites using the association rules derived from the association analysis. We will complete the mapping between these market segments, discuss with the experts, further map the SQF, and finally propose a new job classification system. As a result, more than 30,000 job listings were collected in XML format using open API in 'WORKNET,' 'JOBKOREA,' and 'saramin', which are the main job sites in Korea. After filtering out about 900 job postings simultaneously posted on multiple job sites, 800 association rules were derived by applying the Apriori algorithm, which is a frequent pattern mining. Based on 800 related rules, the job classification system of WORKNET, JOBKOREA, and saramin and the SQF job classification system were mapped and classified into 1st and 4th stages. In the new job taxonomy, the first primary class, IT consulting, computer system, network, and security related job system, consisted of three secondary classifications, five tertiary classifications, and five fourth classifications. The second primary classification, the database and the job system related to system operation, consisted of three secondary classifications, three tertiary classifications, and four fourth classifications. The third primary category, Web Planning, Web Programming, Web Design, and Game, was composed of four secondary classifications, nine tertiary classifications, and two fourth classifications. The last primary classification, job systems related to ICT management, computer and communication engineering technology, consisted of three secondary classifications and six tertiary classifications. In particular, the new job classification system has a relatively flexible stage of classification, unlike other existing classification systems. WORKNET divides jobs into third categories, JOBKOREA divides jobs into second categories, and the subdivided jobs into keywords. saramin divided the job into the second classification, and the subdivided the job into keyword form. The newly proposed standard job classification system accepts some keyword-based jobs, and treats some product names as jobs. In the classification system, not only are jobs suspended in the second classification, but there are also jobs that are subdivided into the fourth classification. This reflected the idea that not all jobs could be broken down into the same steps. We also proposed a combination of rules and experts' opinions from market data collected and conducted associative analysis. Therefore, the newly proposed job classification system can be regarded as a data-based intelligent job classification system that reflects the market demand, unlike the existing job classification system. This study is meaningful in that it suggests a new job classification system that reflects market demand by attempting mapping between occupations based on data through the association analysis between occupations rather than intuition of some experts. However, this study has a limitation in that it cannot fully reflect the market demand that changes over time because the data collection point is temporary. As market demands change over time, including seasonal factors and major corporate public recruitment timings, continuous data monitoring and repeated experiments are needed to achieve more accurate matching. The results of this study can be used to suggest the direction of improvement of SQF in the SW industry in the future, and it is expected to be transferred to other industries with the experience of success in the SW industry.

A Study on Automatic Classification Model of Documents Based on Korean Standard Industrial Classification (한국표준산업분류를 기준으로 한 문서의 자동 분류 모델에 관한 연구)

  • Lee, Jae-Seong;Jun, Seung-Pyo;Yoo, Hyoung Sun
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.221-241
    • /
    • 2018
  • As we enter the knowledge society, the importance of information as a new form of capital is being emphasized. The importance of information classification is also increasing for efficient management of digital information produced exponentially. In this study, we tried to automatically classify and provide tailored information that can help companies decide to make technology commercialization. Therefore, we propose a method to classify information based on Korea Standard Industry Classification (KSIC), which indicates the business characteristics of enterprises. The classification of information or documents has been largely based on machine learning, but there is not enough training data categorized on the basis of KSIC. Therefore, this study applied the method of calculating similarity between documents. Specifically, a method and a model for presenting the most appropriate KSIC code are proposed by collecting explanatory texts of each code of KSIC and calculating the similarity with the classification object document using the vector space model. The IPC data were collected and classified by KSIC. And then verified the methodology by comparing it with the KSIC-IPC concordance table provided by the Korean Intellectual Property Office. As a result of the verification, the highest agreement was obtained when the LT method, which is a kind of TF-IDF calculation formula, was applied. At this time, the degree of match of the first rank matching KSIC was 53% and the cumulative match of the fifth ranking was 76%. Through this, it can be confirmed that KSIC classification of technology, industry, and market information that SMEs need more quantitatively and objectively is possible. In addition, it is considered that the methods and results provided in this study can be used as a basic data to help the qualitative judgment of experts in creating a linkage table between heterogeneous classification systems.

Development of a TBM Advance Rate Model and Its Field Application Based on Full-Scale Shield TBM Tunneling Tests in 70 MPa of Artificial Rock Mass (70 MPa급 인공암반 내 실대형 쉴드TBM 굴진실험을 통한 굴진율 모델 및 활용방안 제안)

  • Kim, Jungjoo;Kim, Kyoungyul;Ryu, Heehwan;Hwan, Jung Ju;Hong, Sungyun;Jo, Seonah;Bae, Dusan
    • KEPCO Journal on Electric Power and Energy
    • /
    • v.6 no.3
    • /
    • pp.305-313
    • /
    • 2020
  • The use of cable tunnels for electric power transmission as well as their construction in difficult conditions such as in subsea terrains and large overburden areas has increased. So, in order to efficiently operate the small diameter shield TBM (Tunnel Boring Machine), the estimation of advance rate and development of a design model is necessary. However, due to limited scope of survey and face mapping, it is very difficult to match the rock mass characteristics and TBM operational data in order to achieve their mutual relationships and to develop an advance rate model. Also, the working mechanism of previously utilized linear cutting machine is slightly different than the real excavation mechanism owing to the penetration of a number of disc cutters taking place at the same time in the rock mass in conjunction with rotation of the cutterhead. So, in order to suggest the advance rate and machine design models for small diameter TBMs, an EPB (Earth Pressure Balance) shield TBM having 3.54 m diameter cutterhead was manufactured and 19 cases of full-scale tunneling tests were performed each in 87.5 ㎥ volume of artificial rock mass. The relationships between advance rate and machine data were effectively analyzed by performing the tests in homogeneous rock mass with 70 MPa uniaxial compressive strength according to the TBM operational parameters such as thrust force and RPM of cutterhead. The utilization of the recorded penetration depth and torque values in the development of models is more accurate and realistic since they were derived through real excavation mechanism. The relationships between normal force on single disc cutter and penetration depth as well as between normal force and rolling force were suggested in this study. The prediction of advance rate and design of TBM can be performed in rock mass having 70 MPa strength using these relationships. An effort was made to improve the application of the developed model by applying the FPI (Field Penetration Index) concept which can overcome the limitation of 100% RQD (Rock Quality Designation) in artificial rock mass.

Machine learning-based corporate default risk prediction model verification and policy recommendation: Focusing on improvement through stacking ensemble model (머신러닝 기반 기업부도위험 예측모델 검증 및 정책적 제언: 스태킹 앙상블 모델을 통한 개선을 중심으로)

  • Eom, Haneul;Kim, Jaeseong;Choi, Sangok
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.105-129
    • /
    • 2020
  • This study uses corporate data from 2012 to 2018 when K-IFRS was applied in earnest to predict default risks. The data used in the analysis totaled 10,545 rows, consisting of 160 columns including 38 in the statement of financial position, 26 in the statement of comprehensive income, 11 in the statement of cash flows, and 76 in the index of financial ratios. Unlike most previous prior studies used the default event as the basis for learning about default risk, this study calculated default risk using the market capitalization and stock price volatility of each company based on the Merton model. Through this, it was able to solve the problem of data imbalance due to the scarcity of default events, which had been pointed out as the limitation of the existing methodology, and the problem of reflecting the difference in default risk that exists within ordinary companies. Because learning was conducted only by using corporate information available to unlisted companies, default risks of unlisted companies without stock price information can be appropriately derived. Through this, it can provide stable default risk assessment services to unlisted companies that are difficult to determine proper default risk with traditional credit rating models such as small and medium-sized companies and startups. Although there has been an active study of predicting corporate default risks using machine learning recently, model bias issues exist because most studies are making predictions based on a single model. Stable and reliable valuation methodology is required for the calculation of default risk, given that the entity's default risk information is very widely utilized in the market and the sensitivity to the difference in default risk is high. Also, Strict standards are also required for methods of calculation. The credit rating method stipulated by the Financial Services Commission in the Financial Investment Regulations calls for the preparation of evaluation methods, including verification of the adequacy of evaluation methods, in consideration of past statistical data and experiences on credit ratings and changes in future market conditions. This study allowed the reduction of individual models' bias by utilizing stacking ensemble techniques that synthesize various machine learning models. This allows us to capture complex nonlinear relationships between default risk and various corporate information and maximize the advantages of machine learning-based default risk prediction models that take less time to calculate. To calculate forecasts by sub model to be used as input data for the Stacking Ensemble model, training data were divided into seven pieces, and sub-models were trained in a divided set to produce forecasts. To compare the predictive power of the Stacking Ensemble model, Random Forest, MLP, and CNN models were trained with full training data, then the predictive power of each model was verified on the test set. The analysis showed that the Stacking Ensemble model exceeded the predictive power of the Random Forest model, which had the best performance on a single model. Next, to check for statistically significant differences between the Stacking Ensemble model and the forecasts for each individual model, the Pair between the Stacking Ensemble model and each individual model was constructed. Because the results of the Shapiro-wilk normality test also showed that all Pair did not follow normality, Using the nonparametric method wilcoxon rank sum test, we checked whether the two model forecasts that make up the Pair showed statistically significant differences. The analysis showed that the forecasts of the Staging Ensemble model showed statistically significant differences from those of the MLP model and CNN model. In addition, this study can provide a methodology that allows existing credit rating agencies to apply machine learning-based bankruptcy risk prediction methodologies, given that traditional credit rating models can also be reflected as sub-models to calculate the final default probability. Also, the Stacking Ensemble techniques proposed in this study can help design to meet the requirements of the Financial Investment Business Regulations through the combination of various sub-models. We hope that this research will be used as a resource to increase practical use by overcoming and improving the limitations of existing machine learning-based models.