• Title/Summary/Keyword: Generate Data

Search Result 3,065, Processing Time 0.035 seconds

A Qualitative Study on Facilitating Factors of User-Created Contents: Based on Theories of Folklore (사용자 제작 콘텐츠의 활성화 요인에 대한 정성적 연구: 구비문학 이론을 중심으로)

  • Jung, Seung-Ki;Lee, Ki-Ho;Lee, In-Seong;Kim, Jin-Woo
    • Asia pacific journal of information systems
    • /
    • v.19 no.2
    • /
    • pp.43-72
    • /
    • 2009
  • Recently, user-created content (UCC) have emerged as popular medium of on-line participation among users. The Internet environment has been constantly evolving, attracting active participation and information sharing among common users. This tendency is a significant deviation from the earlier Internet use as an one-way information channel through which users passively received information or contents from contents providers. Thanks to UCCs online users can now more freely generate and exchange contents; therefore, identifying the critical factors that affect content-generating activities has increasingly become an important issue. This paper proposes a set of critical factors for stimulating contents generation and sharing activities by Internet users. These factors were derived from the theories of folklores such as tales and songs. Based on some shared traits of folklores and UCC content, we found four critical elements which should be heeded in constructing UCC contents, which are: context of culture, context of situation, skill of generator, and response of audience. In addition, we selected three major UCC websites: a specialized contents portal, a general internet portal, and an official contents service site, They have different use environments, user interfaces, and service policies, To identify critical factors for generating, sharing and transferring UCC, we traced user activities, interactions and flows of content in the three UCC websites. Moreover, we conducted extensive interviews with users and operators as well as policy makers in each site. Based on qualitative and quantitative analyses of the data, this research identifies nine critical factors that facilitate contents generation and sharing activities among users. In the context of culture, we suggest voluntary community norms, proactive use of copyrights, strong user relationships, and a fair monetary reward system as critical elements in facilitating the process of contents generation and sharing activities. Norms which were established by users themselves regulate user behavior and influence content format. Strong relationships of users stimulate content generation activities by enhancing collaborative content generation. Particularly, users generate contents through collaboration with others, based on their enhanced relationship and specialized skills. They send and receive contents by leaving messages on website or blogs, using instant messenger or SMS. It is an interesting and important phenomenon, because the quality of contents can be constantly improved and revised, depending on the specialized abilities of those engaged in a particular content. In this process, the reward system is an essential driving factor. Yet, monetary reward should be considered only after some fair criterion is established. In terms of the context of the situation, the quality of contents uploading system was proposed to have strong influence on the content generating activities. Among other influential factors on contents generation activities are generators' specialized skills and involvement of the users were proposed. In addition, the audience response, especially effective development of shared interests as well as feedback, was suggested to have significant influence on contents generation activities. Content generators usually reflect the shared interest of others. Shared interest is a distinct characteristic of UCC and observed in all the three websites, in which common interest is formed by the "threads" embedded with content. Through such threads of information and contents users discuss and share ideas while continuously extending and updating shared contents in the process. Evidently, UCC is a new paradigm representing the next generation of the Internet. In order to fully utilize this innovative paradigm, we need to understand how users take advantage of this medium in generating contents, and what affects their content generation activities. Based on these findings, UCC service providers should design their websites as common playground where users freely interact and share their common interests. As such this paper makes an important first step to gaining better understand about this new communication paradigm created by UCC.

Sentiment Analysis of Korean Reviews Using CNN: Focusing on Morpheme Embedding (CNN을 적용한 한국어 상품평 감성분석: 형태소 임베딩을 중심으로)

  • Park, Hyun-jung;Song, Min-chae;Shin, Kyung-shik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.59-83
    • /
    • 2018
  • With the increasing importance of sentiment analysis to grasp the needs of customers and the public, various types of deep learning models have been actively applied to English texts. In the sentiment analysis of English texts by deep learning, natural language sentences included in training and test datasets are usually converted into sequences of word vectors before being entered into the deep learning models. In this case, word vectors generally refer to vector representations of words obtained through splitting a sentence by space characters. There are several ways to derive word vectors, one of which is Word2Vec used for producing the 300 dimensional Google word vectors from about 100 billion words of Google News data. They have been widely used in the studies of sentiment analysis of reviews from various fields such as restaurants, movies, laptops, cameras, etc. Unlike English, morpheme plays an essential role in sentiment analysis and sentence structure analysis in Korean, which is a typical agglutinative language with developed postpositions and endings. A morpheme can be defined as the smallest meaningful unit of a language, and a word consists of one or more morphemes. For example, for a word '예쁘고', the morphemes are '예쁘(= adjective)' and '고(=connective ending)'. Reflecting the significance of Korean morphemes, it seems reasonable to adopt the morphemes as a basic unit in Korean sentiment analysis. Therefore, in this study, we use 'morpheme vector' as an input to a deep learning model rather than 'word vector' which is mainly used in English text. The morpheme vector refers to a vector representation for the morpheme and can be derived by applying an existent word vector derivation mechanism to the sentences divided into constituent morphemes. By the way, here come some questions as follows. What is the desirable range of POS(Part-Of-Speech) tags when deriving morpheme vectors for improving the classification accuracy of a deep learning model? Is it proper to apply a typical word vector model which primarily relies on the form of words to Korean with a high homonym ratio? Will the text preprocessing such as correcting spelling or spacing errors affect the classification accuracy, especially when drawing morpheme vectors from Korean product reviews with a lot of grammatical mistakes and variations? We seek to find empirical answers to these fundamental issues, which may be encountered first when applying various deep learning models to Korean texts. As a starting point, we summarized these issues as three central research questions as follows. First, which is better effective, to use morpheme vectors from grammatically correct texts of other domain than the analysis target, or to use morpheme vectors from considerably ungrammatical texts of the same domain, as the initial input of a deep learning model? Second, what is an appropriate morpheme vector derivation method for Korean regarding the range of POS tags, homonym, text preprocessing, minimum frequency? Third, can we get a satisfactory level of classification accuracy when applying deep learning to Korean sentiment analysis? As an approach to these research questions, we generate various types of morpheme vectors reflecting the research questions and then compare the classification accuracy through a non-static CNN(Convolutional Neural Network) model taking in the morpheme vectors. As for training and test datasets, Naver Shopping's 17,260 cosmetics product reviews are used. To derive morpheme vectors, we use data from the same domain as the target one and data from other domain; Naver shopping's about 2 million cosmetics product reviews and 520,000 Naver News data arguably corresponding to Google's News data. The six primary sets of morpheme vectors constructed in this study differ in terms of the following three criteria. First, they come from two types of data source; Naver news of high grammatical correctness and Naver shopping's cosmetics product reviews of low grammatical correctness. Second, they are distinguished in the degree of data preprocessing, namely, only splitting sentences or up to additional spelling and spacing corrections after sentence separation. Third, they vary concerning the form of input fed into a word vector model; whether the morphemes themselves are entered into a word vector model or with their POS tags attached. The morpheme vectors further vary depending on the consideration range of POS tags, the minimum frequency of morphemes included, and the random initialization range. All morpheme vectors are derived through CBOW(Continuous Bag-Of-Words) model with the context window 5 and the vector dimension 300. It seems that utilizing the same domain text even with a lower degree of grammatical correctness, performing spelling and spacing corrections as well as sentence splitting, and incorporating morphemes of any POS tags including incomprehensible category lead to the better classification accuracy. The POS tag attachment, which is devised for the high proportion of homonyms in Korean, and the minimum frequency standard for the morpheme to be included seem not to have any definite influence on the classification accuracy.

A Study on the Application of Outlier Analysis for Fraud Detection: Focused on Transactions of Auction Exception Agricultural Products (부정 탐지를 위한 이상치 분석 활용방안 연구 : 농수산 상장예외품목 거래를 대상으로)

  • Kim, Dongsung;Kim, Kitae;Kim, Jongwoo;Park, Steve
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.3
    • /
    • pp.93-108
    • /
    • 2014
  • To support business decision making, interests and efforts to analyze and use transaction data in different perspectives are increasing. Such efforts are not only limited to customer management or marketing, but also used for monitoring and detecting fraud transactions. Fraud transactions are evolving into various patterns by taking advantage of information technology. To reflect the evolution of fraud transactions, there are many efforts on fraud detection methods and advanced application systems in order to improve the accuracy and ease of fraud detection. As a case of fraud detection, this study aims to provide effective fraud detection methods for auction exception agricultural products in the largest Korean agricultural wholesale market. Auction exception products policy exists to complement auction-based trades in agricultural wholesale market. That is, most trades on agricultural products are performed by auction; however, specific products are assigned as auction exception products when total volumes of products are relatively small, the number of wholesalers is small, or there are difficulties for wholesalers to purchase the products. However, auction exception products policy makes several problems on fairness and transparency of transaction, which requires help of fraud detection. In this study, to generate fraud detection rules, real huge agricultural products trade transaction data from 2008 to 2010 in the market are analyzed, which increase more than 1 million transactions and 1 billion US dollar in transaction volume. Agricultural transaction data has unique characteristics such as frequent changes in supply volumes and turbulent time-dependent changes in price. Since this was the first trial to identify fraud transactions in this domain, there was no training data set for supervised learning. So, fraud detection rules are generated using outlier detection approach. We assume that outlier transactions have more possibility of fraud transactions than normal transactions. The outlier transactions are identified to compare daily average unit price, weekly average unit price, and quarterly average unit price of product items. Also quarterly averages unit price of product items of the specific wholesalers are used to identify outlier transactions. The reliability of generated fraud detection rules are confirmed by domain experts. To determine whether a transaction is fraudulent or not, normal distribution and normalized Z-value concept are applied. That is, a unit price of a transaction is transformed to Z-value to calculate the occurrence probability when we approximate the distribution of unit prices to normal distribution. The modified Z-value of the unit price in the transaction is used rather than using the original Z-value of it. The reason is that in the case of auction exception agricultural products, Z-values are influenced by outlier fraud transactions themselves because the number of wholesalers is small. The modified Z-values are called Self-Eliminated Z-scores because they are calculated excluding the unit price of the specific transaction which is subject to check whether it is fraud transaction or not. To show the usefulness of the proposed approach, a prototype of fraud transaction detection system is developed using Delphi. The system consists of five main menus and related submenus. First functionalities of the system is to import transaction databases. Next important functions are to set up fraud detection parameters. By changing fraud detection parameters, system users can control the number of potential fraud transactions. Execution functions provide fraud detection results which are found based on fraud detection parameters. The potential fraud transactions can be viewed on screen or exported as files. The study is an initial trial to identify fraud transactions in Auction Exception Agricultural Products. There are still many remained research topics of the issue. First, the scope of analysis data was limited due to the availability of data. It is necessary to include more data on transactions, wholesalers, and producers to detect fraud transactions more accurately. Next, we need to extend the scope of fraud transaction detection to fishery products. Also there are many possibilities to apply different data mining techniques for fraud detection. For example, time series approach is a potential technique to apply the problem. Even though outlier transactions are detected based on unit prices of transactions, however it is possible to derive fraud detection rules based on transaction volumes.

The Effect of Supply Chain Dynamic Capabilities, Open Innovation and Supply Uncertainty on Supply Chain Performance (공급사슬 동적역량, 개방형 혁신, 공급 불확실성이 공급사슬 성과에 미치는 영향)

  • Lee, Sang-Yeol
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.19 no.4
    • /
    • pp.481-491
    • /
    • 2018
  • As the global business environment is dynamic, uncertain, and complex, supply chain management determines the performance of the supply chain in terms of the utilization of resources and capabilities of companies involved in the supply chain. Companies pursuing open innovation gain greater access to the external environment and accumulate knowledge flows and learning experiences, and may generate better business performance from dynamic capabilities. This study analyzed the effects of supply chain dynamic capabilities, open innovation, and supply uncertainty on supply chain performance. Through questionnaires on 178 companies listed on KOSDAQ, empirical results are as follows: First, integration and reactivity capabilities among supply chain dynamic capabilities have a positive effect on supply chain performance. Second, the moderating effect of open innovation showed a negative correlation in the case of information exchange, and a positive correlation in the cases of integration, cooperation and reactivity. Third, two of the 3-way interaction terms, "information exchange*open innovation*supply uncertainty" and "integration*open innovation*supply uncertainty" were statistically significant. The implications of this study are as follows: First, as the supply chain needs to achieve optimization of the whole process between supply chain components rather than individual companies, dynamic capabilities play an important role in improving performance. Second, for KOSDAQ companies featuring limited capital resources, open innovation that integrates external knowledge is valuable. In order to increase synergistic effects, it is necessary to develop dynamic capabilities accordingly. Third, since resources are constrained, managers must determine the type or level of capabilities and open innovation in accordance with supply uncertainty. Since this study has limitations in analyzing survey data, it is necessary to collect secondary data or longitudinal data. It is also necessary to further analyze the internal and external factors that have a significant impact on supply chain performance.

Quantification of Brain Images Using Korean Standard Templates and Structural and Cytoarchitectonic Probabilistic Maps (한국인 뇌 표준판과 해부학적 및 세포구축학적 확률뇌지도를 이용한 뇌영상 정량화)

  • Lee, Jae-Sung;Lee, Dong-Soo;Kim, Yu-Kyeong;Kim, Jin-Su;Lee, Jong-Min;Koo, Bang-Bon;Kim, Jae-Jin;Kwon, Jun-Soo;Yoo, Tae-Woo;Chang, Ki-Hyun;Kim, Sun-I.;Kang, Hye-Jin;Kang, Eun-Joo
    • The Korean Journal of Nuclear Medicine
    • /
    • v.38 no.3
    • /
    • pp.241-252
    • /
    • 2004
  • Purpose: Population based structural and functional maps of the brain provide effective tools for the analysis and interpretation of complex and individually variable brain data. Brain MRI and PET standard templates and statistical probabilistic maps based on image data of Korean normal volunteers have been developed and probabilistic maps based on cytoarchitectonic data have been introduced. A quantification method using these data was developed for the objective assessment of regional intensity in the brain images. Materials and Methods: Age, gender and ethnic specific anatomical and functional brain templates based on MR and PET images of Korean normal volunteers were developed. Korean structural probabilistic maps for 89 brain regions and cytoarchitectonic probabilistic maps for 13 Brodmann areas were transformed onto the standard templates. Brain FDG PET and SPGR MR images of normal volunteers were spatially normalized onto the template of each modality and gender. Regional uptake of radiotracers in PET and gray matter concentration in MR images were then quantified by averaging (or summing) regional intensities weighted using the probabilistic maps of brain regions. Regionally specific effects of aging on glucose metabolism in cingulate cortex were also examined. Results: Quantification program could generate quantification results for single spatially normalized images per 20 seconds. Glucose metabolism change in cingulate gyrus was regionally specific: ratios of glucose metabolism in the rostral anterior cingulate vs. posterior cingulate and the caudal anterior cingulate vs. posterior cingulate were significantly decreased as the age increased. 'Rostral anterior'/'posterior' was decreased by 3.1% per decade of age ($P<10^{-11}$, r=0.81) and 'caudal anterior'/'posterior' was decreased by 1.7% ($P<10^{-8}$, r=0.72). Conclusion: Ethnic specific standard templates and probabilistic maps and quantification program developed in this study will be useful for the analysis of brain image of Korean people since the difference in shape of the hemispheres and the sulcal pattern of brain relative to age, gender, races, and diseases cannot be fully overcome by the nonlinear spatial normalization techniques.

A Study on groundwater and pollutant recharge in urban area: use of hydrochemical data

  • Lee, Ju-Hee;Kwon, Jang-Soon;Yun, Seong-Taek;Chae, Gi-Tak;Park, Seong-Sook
    • Proceedings of the Korean Society of Soil and Groundwater Environment Conference
    • /
    • 2004.09a
    • /
    • pp.119-120
    • /
    • 2004
  • Urban groundwater has a unique hydrologic system because of the complex surface and subsurface infrastructures such as deep foundation of many high buildings, subway systems, and sewers and public water supply systems. It generally has been considered that increased surface impermeability reduces the amount of groundwater recharge. On the other hand, leaks from sewers and public water supply systems may generate the large amounts of recharges. All of these urban facilities also may change the groundwater quality by the recharge of a myriad of contaminants. This study was performed to determine the factors controlling the recharge of deep groundwater in an urban area, based on the hydrogeochemical characteristics. The term ‘contamination’ in this study means any kind of inflow of shallow groundwater regardless of clean or contaminated. For this study, urban groundwater samples were collected from a total of 310 preexisting wells with the depth over 100 m. Random sampling method was used to select the wells for this study. Major cations together with Si, Al, Fe, Pb, Hg and Mn were analyzed by ICP-AES, and Cl, N $O_3$, N $H_4$, F, Br, S $O_4$and P $O_4$ were analyzed by IC. There are two groups of groundwater, based on hydrochemical characteristics. The first group is distributed broadly from Ca-HC $O_3$ type to Ca-C1+N $O_3$ type; the other group is the Na+K-HC $O_3$ type. The latter group is considered to represent the baseline quality of deep groundwater in the study area. Using the major ions data for the Na+K-HC $O_3$ type water, we evaluated the extent of groundwater contamination, assuming that if subtract the baseline composition from acquired data for a specific water, the remaining concentrations may indicate the degree of contamination. The remainder of each solute for each sample was simply averaged. The results showed that both Ca and HC $O_3$ represent the typical solutes which are quite enriched in urban groundwater. In particular, the P$CO_2$ values calculated using PHREEQC (version 2.8) showed a correlation with the concentrations of maior inorganic components (Na, Mg, Ca, N $O_3$, S $O_4$, etc.). The p$CO_2$ values for the first group waters widely ranged between about 10$^{-3.0}$ atm to 10$^{-1.0}$ atm and differed from those of the background water samples belonging to the Na+K-HC $O_3$ type (<10$^{-3.5}$ atm). Considering that the p$CO_2$ of soil water (near 10$^{-1.5}$ atm), this indicates that inflow of shallow water is very significant in deep groundwaters in the study area. Furthermore, the P$CO_2$ values can be used as an effective parameter to estimate the relative recharge of shallow water and thus the contamination susceptibility. The results of our present study suggest that down to considerable depth, urban groundwater in crystalline aquifer may be considerably affected by the recharge of shallow water (and pollutants) from an adjacent area. We also suggest that for such evaluation, careful examination of systematically collected hydrochemical data is requisite as an effective tool, in addition to hydrologic and hydrogeologic interpretation.ion.ion.

  • PDF

Analysis and Performance Evaluation of Pattern Condensing Techniques used in Representative Pattern Mining (대표 패턴 마이닝에 활용되는 패턴 압축 기법들에 대한 분석 및 성능 평가)

  • Lee, Gang-In;Yun, Un-Il
    • Journal of Internet Computing and Services
    • /
    • v.16 no.2
    • /
    • pp.77-83
    • /
    • 2015
  • Frequent pattern mining, which is one of the major areas actively studied in data mining, is a method for extracting useful pattern information hidden from large data sets or databases. Moreover, frequent pattern mining approaches have been actively employed in a variety of application fields because the results obtained from them can allow us to analyze various, important characteristics within databases more easily and automatically. However, traditional frequent pattern mining methods, which simply extract all of the possible frequent patterns such that each of their support values is not smaller than a user-given minimum support threshold, have the following problems. First, traditional approaches have to generate a numerous number of patterns according to the features of a given database and the degree of threshold settings, and the number can also increase in geometrical progression. In addition, such works also cause waste of runtime and memory resources. Furthermore, the pattern results excessively generated from the methods also lead to troubles of pattern analysis for the mining results. In order to solve such issues of previous traditional frequent pattern mining approaches, the concept of representative pattern mining and its various related works have been proposed. In contrast to the traditional ones that find all the possible frequent patterns from databases, representative pattern mining approaches selectively extract a smaller number of patterns that represent general frequent patterns. In this paper, we describe details and characteristics of pattern condensing techniques that consider the maximality or closure property of generated frequent patterns, and conduct comparison and analysis for the techniques. Given a frequent pattern, satisfying the maximality for the pattern signifies that all of the possible super sets of the pattern must have smaller support values than a user-specific minimum support threshold; meanwhile, satisfying the closure property for the pattern means that there is no superset of which the support is equal to that of the pattern with respect to all the possible super sets. By mining maximal frequent patterns or closed frequent ones, we can achieve effective pattern compression and also perform mining operations with much smaller time and space resources. In addition, compressed patterns can be converted into the original frequent pattern forms again if necessary; especially, the closed frequent pattern notation has the ability to convert representative patterns into the original ones again without any information loss. That is, we can obtain a complete set of original frequent patterns from closed frequent ones. Although the maximal frequent pattern notation does not guarantee a complete recovery rate in the process of pattern conversion, it has an advantage that can extract a smaller number of representative patterns more quickly compared to the closed frequent pattern notation. In this paper, we show the performance results and characteristics of the aforementioned techniques in terms of pattern generation, runtime, and memory usage by conducting performance evaluation with respect to various real data sets collected from the real world. For more exact comparison, we also employ the algorithms implementing these techniques on the same platform and Implementation level.

Optimal Selection of Classifier Ensemble Using Genetic Algorithms (유전자 알고리즘을 이용한 분류자 앙상블의 최적 선택)

  • Kim, Myung-Jong
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.4
    • /
    • pp.99-112
    • /
    • 2010
  • Ensemble learning is a method for improving the performance of classification and prediction algorithms. It is a method for finding a highly accurateclassifier on the training set by constructing and combining an ensemble of weak classifiers, each of which needs only to be moderately accurate on the training set. Ensemble learning has received considerable attention from machine learning and artificial intelligence fields because of its remarkable performance improvement and flexible integration with the traditional learning algorithms such as decision tree (DT), neural networks (NN), and SVM, etc. In those researches, all of DT ensemble studies have demonstrated impressive improvements in the generalization behavior of DT, while NN and SVM ensemble studies have not shown remarkable performance as shown in DT ensembles. Recently, several works have reported that the performance of ensemble can be degraded where multiple classifiers of an ensemble are highly correlated with, and thereby result in multicollinearity problem, which leads to performance degradation of the ensemble. They have also proposed the differentiated learning strategies to cope with performance degradation problem. Hansen and Salamon (1990) insisted that it is necessary and sufficient for the performance enhancement of an ensemble that the ensemble should contain diverse classifiers. Breiman (1996) explored that ensemble learning can increase the performance of unstable learning algorithms, but does not show remarkable performance improvement on stable learning algorithms. Unstable learning algorithms such as decision tree learners are sensitive to the change of the training data, and thus small changes in the training data can yield large changes in the generated classifiers. Therefore, ensemble with unstable learning algorithms can guarantee some diversity among the classifiers. To the contrary, stable learning algorithms such as NN and SVM generate similar classifiers in spite of small changes of the training data, and thus the correlation among the resulting classifiers is very high. This high correlation results in multicollinearity problem, which leads to performance degradation of the ensemble. Kim,s work (2009) showedthe performance comparison in bankruptcy prediction on Korea firms using tradition prediction algorithms such as NN, DT, and SVM. It reports that stable learning algorithms such as NN and SVM have higher predictability than the unstable DT. Meanwhile, with respect to their ensemble learning, DT ensemble shows the more improved performance than NN and SVM ensemble. Further analysis with variance inflation factor (VIF) analysis empirically proves that performance degradation of ensemble is due to multicollinearity problem. It also proposes that optimization of ensemble is needed to cope with such a problem. This paper proposes a hybrid system for coverage optimization of NN ensemble (CO-NN) in order to improve the performance of NN ensemble. Coverage optimization is a technique of choosing a sub-ensemble from an original ensemble to guarantee the diversity of classifiers in coverage optimization process. CO-NN uses GA which has been widely used for various optimization problems to deal with the coverage optimization problem. The GA chromosomes for the coverage optimization are encoded into binary strings, each bit of which indicates individual classifier. The fitness function is defined as maximization of error reduction and a constraint of variance inflation factor (VIF), which is one of the generally used methods to measure multicollinearity, is added to insure the diversity of classifiers by removing high correlation among the classifiers. We use Microsoft Excel and the GAs software package called Evolver. Experiments on company failure prediction have shown that CO-NN is effectively applied in the stable performance enhancement of NNensembles through the choice of classifiers by considering the correlations of the ensemble. The classifiers which have the potential multicollinearity problem are removed by the coverage optimization process of CO-NN and thereby CO-NN has shown higher performance than a single NN classifier and NN ensemble at 1% significance level, and DT ensemble at 5% significance level. However, there remain further research issues. First, decision optimization process to find optimal combination function should be considered in further research. Secondly, various learning strategies to deal with data noise should be introduced in more advanced further researches in the future.

A Study on the Possibility of Short-term Monitoring of Coastal Topography Changes Using GOCI-II (GOCI-II를 활용한 단기 연안지형변화 모니터링 가능성 평가 연구)

  • Lee, Jingyo;Kim, Keunyong;Ryu, Joo-Hyung
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.5_2
    • /
    • pp.1329-1340
    • /
    • 2021
  • The intertidal zone, which is a transitional zone between the ocean and the land, requires continuous monitoring as various changes occur rapidly due to artificial activity and natural disturbance. Monitoring of coastal topography changes using remote sensing method is evaluated to be effective in overcoming the limitations of intertidal zone accessibility and observing long-term topographic changes in intertidal zone. Most of the existing coastal topographic monitoring studies using remote sensing were conducted through high spatial resolution images such as Landsat and Sentinel. This study extracted the waterline using the NDWI from the GOCI-II (Geostationary Ocean Color Satellite-II) data, identified the changes in the intertidal area in Gyeonggi Bay according to various tidal heights, and examined the utility of DEM generation and topography altitude change observation over a short period of time. GOCI-II (249 scenes), Sentinel-2A/B (39 scenes), Landsat 8 OLI (7 scenes) images were obtained around Gyeonggi Bay from October 8, 2020 to August 16, 2021. If generating intertidal area DEM, Sentinel and Landsat images required at least 3 months to 1 year of data collection, but the GOCI-II satellite was able to generate intertidal area DEM in Gyeonggi Bay using only one day of data according to tidal heights, and the topography altitude was also observed through exposure frequency. When observing coastal topography changes using the GOCI-II satellite, it would be a good idea to detect topography changes early through a short cycle and to accurately interpolate and utilize insufficient spatial resolutions using multi-remote sensing data of high resolution. Based on the above results, it is expected that it will be possible to quickly provide information necessary for the latest topographic map and coastal management of the Korean Peninsula by expanding the research area and developing technologies that can be automatically analyzed and detected.

Key Foods selection using data from the 7th Korea National Health and Nutrition Examination Survey (2016-2018) (제7기 국민건강영양조사 (2016-2018) 자료를 활용한 한국인의 주요 식품 (Key Foods) 선정에 관한 연구)

  • Lee, Jung-Sug;Shim, Jee-Seon;Kim, Ki Nam;Lee, Hyun Sook;Chang, Moon-Jeong;Kim, Hye-Young
    • Journal of Nutrition and Health
    • /
    • v.54 no.1
    • /
    • pp.10-22
    • /
    • 2021
  • Purpose: Key Foods refers to foods that have a high contribution in the nutrient intake of individuals, and exert important effects on their health. This study was undertaken to identify Korean Key Foods, using data from the 7th Korea National Health and Nutrition Examination Survey (KNHNES). Methods: The data source for the extraction of Key Foods was the 24-hour dietary survey data obtained from the 7th KNHNES (2016-2018), and 21,271 subjects were evaluated. A total of 17 nutrients were selected as the key nutrients for identifying the Key Foods, including energy, carbohydrates, protein, lipid, dietary fiber, calcium, phosphorus, iron, sodium, potassium, vitamin A, thiamin, riboflavin, niacin, vitamin C, cholesterol, and sugars. The nutrient consumption approach was applied to generate a list of potential Key Foods. Foods included in 85% of the cumulative intake contribution from one or more key nutrients, were subsequently selected as Key Foods. Results: Of the 1,728 foods consumed by survey respondents, we extracted 728 Key Foods. These Key Foods explained 94% key nutrient intakes of the subjects. Based on the contribution rate to key nutrient intake, the top 10 Key Foods identified were multigrain rice (5.32%), plain white rice (4.23%), milk (3.3%), cabbage kimchi (2.82%), grilled pork belly (1.56%), apples (1.52%), fried eggs (1.49%), cereal (1.36%), instant coffee mix (1.21%), and sweet potatoes (1.12%). These 10 foods accounted for 23.93% total key nutrient intake of the survey respondents. Conclusion: Seven hundred and twenty-eight foods were extracted and identified as the 2020 Korean Key Foods. These Key Foods can be considered the priority foods to be analyzed for establishing a national nutrient database.