• Title/Summary/Keyword: Big Data Based Modeling

Search Result 181, Processing Time 0.026 seconds

A Study on Global Value Chains(GVCs) Research Trends Based on Keyword Network Analysis (키워드 네트워크 분석을 활용한 글로벌가치사슬(GVCs) 연구동향 분석)

  • Hyun-Yong Park;Young-Jun Choi;Li Jia-En
    • Korea Trade Review
    • /
    • v.45 no.5
    • /
    • pp.239-260
    • /
    • 2020
  • This research was conducted on 176 GVCs-related research papers listed in the Index of Korean Academic Writers. The analysis methodology used the keyword network analysis methodology of big data analysis. For the comprehensive analysis of research trends, the research trends through word frequency (TF), important topic (TF-IDF), and topical modeling were analyzed in 176 papers. In addition, the research period of GVCs was divided into the early stages of the first study (2003-2014), the second phase of the study (2015-2017), and the third phase of the study (2018-2020). According to the comprehensive analysis, the GVCs research was conducted with the keyword 'value added' as the center, focusing on the keywords of export (trade), Korea, business, influence, and production. Major research topics were 'supporting corporate cooperation and capacity building' and 'comparative advantage with added value of overseas direct investment'. According to the analysis of major period-specific research trends, GVCs were studied in the early stages of the first phase of the study with global value chain trends and corporate production strategies. In the second research propulsion period, research was done in terms of trade value added. In the recent third phase of the study, small and medium-sized enterprises actively participated in the global value chain and actively researched ways to support the government. Through this study, the importance of the global value chain has been confirmed quantitatively and qualitatively, and it is recognized as an important factor to be considered in the strategy of enhancing industrial competitiveness and entering overseas markets. In particular, small and medium-sized companies' participation in the global value chain and support measures are being presented as important research topics in the future.

Case Analysis on High Concentration of SO2 and Review on Its Reduction Policy in the Ulsan Metropolitan Area since 2001 (울산 지역에서 2001년 이후 이산화황(SO2)의 고농도 사례 분석과 저감 정책 방안의 검토)

  • Moon, Yun-Seob
    • Journal of Environmental Science International
    • /
    • v.17 no.4
    • /
    • pp.423-437
    • /
    • 2008
  • Until comparatively lately, the annual time series of the $SO_2$ concentration had been shown in a decreasing trend in Ulsan as well as other Korean cities. However, the high concentration of $SO_2$ was frequently found in the specific countermeasure region including the national industrial complex such as Mipo and Onsan in the Ulsan city for the period of $2001{\sim}2004$. There are many conditions that can influence the high concentration of $SO_2$ at monitoring sites in Ulsan, such as: First, annual usage of the fuel including sulfur increased in comparison with the year before in spite of the fuel conversion policy which wants to use low sulfur oil less than 3% and LNG. Second, point source, such as the power plants and the petroleum and chemistry stacks, was the biggest contributor in $SO_2$ emission, as a analyzed result of both the air quality modeling and the stack tole-monitoring system (TMS) data. And third, the air pollutants that occurred in processes of homing and manufacturing of the fuel including sulfur were transported slow into a special monitoring site by accumulating along the frontal area of see-breeze. It was concluded that Ulsan's current environmental policy together with control methods should be changed into the regulation on total amount of emission, including a market-based emission trading with calculating of atmospheric environmental critical loads, for the $SO_2$ reduction like the specific countermeasure for the $O_3$ and PM10 reduction in the Seoul metropolitan area. And this change should be started in the big point sources of $1{\sim}3$ species because they are big contributors of Ulsan's $SO_2$ pollution. Especially it is necessary to revitalize of the self-regulation environmental management. Other control methods for sustaining the $SO_2$ reduction are as follows: maintenance of the fuel conversion policy, reinforcement of the regional stationary source emission standard, and enlargement of the stack TMS.

3D Product digital camera Model on the Web and study about developing 3D shopping mall (Web 상에서 3차원 디지털카메라제품모델과 3차원 쇼핑몰 개발에 관한 연구)

  • 조진희;이규옥
    • Archives of design research
    • /
    • v.14 no.1
    • /
    • pp.63-70
    • /
    • 2001
  • Thanks to the inter-connection of information servers throughout the world based on the internet technology, the new sphere which actual transaction can be made like in the visible market has become conspicuous as the virtual space. The movement to realize the new business through the cyber space has been actively ongoing. In the domestic market, a lot of corporations knowing the needs of internet shopping malls have entered into this e-business but they have not made a big success comparing with internet's potentials. And, it can be attributed to the simple planes and the limitations of information provided by the cyber malls, which means that the needs of better information transfer we apparent Accordingly, in this thesis, the research on the 3-D based products and shopping malls has been made through the inter-complementary composition between the 2-D shopping malls and 3-B ones. This research consists of 3 parts. Firstly, through the research on references and existing data, it presents the analysis on consumer's characteristics and sales limits of the internet shopping mall's products. Secondly, the background of 3-D shopping mall's advent and the virtual reality technology data are put together. Finally, it presents how the development of 3-D based product modeling and shopping malls can increase the consumer's purchase power and furthermore the directions of shopping malls to go.

  • PDF

Outliers and Level Shift Detection of the Mean-sea Level, Extreme Highest and Lowest Tide Level Data (평균 해수면 및 최극조위 자료의 이상자료 및 기준고도 변화(Level Shift) 진단)

  • Lee, Gi-Seop;Cho, Hong-Yeon
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.32 no.5
    • /
    • pp.322-330
    • /
    • 2020
  • Modeling for outliers in time series was carried out using the MSL and extreme high, low tide levels (EHL, HLL) data set in the Busan and Mokpo stations. The time-series model is seasonal ARIMA model including the components of the AO (additive outliers) and LS (level shift). The optimal model was selected based on the AIC value and the model parameters were estimated using the 'tso' function (in 'tsoutliers' package of R). The main results by the model application, i.e.. outliers and level shift detections, are as follows. (1) The two AO are detected in the Busan monthly EHL data and the AO magnitudes were estimated to 65.5 cm (by typhoon MAEMI) and 29.5 cm (by typhoon SANBA), respectively. (2) The one level shift in 1983 is detected in Mokpo monthly MSL data, and the LS magnitude was estimated to 21.2 cm by the Youngsan River tidal estuary barrier construction. On the other hand, the RMS errors are computed about 1.95 cm (MSL), 5.11 cm (EHL), and 6.50 cm (ELL) in Busan station, and about 2.10 cm (MSL), 11.80 cm (EHL), and 9.14 cm (ELL) in Mokpo station, respectively.

Clickstream Big Data Mining for Demographics based Digital Marketing (인구통계특성 기반 디지털 마케팅을 위한 클릭스트림 빅데이터 마이닝)

  • Park, Jiae;Cho, Yoonho
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.3
    • /
    • pp.143-163
    • /
    • 2016
  • The demographics of Internet users are the most basic and important sources for target marketing or personalized advertisements on the digital marketing channels which include email, mobile, and social media. However, it gradually has become difficult to collect the demographics of Internet users because their activities are anonymous in many cases. Although the marketing department is able to get the demographics using online or offline surveys, these approaches are very expensive, long processes, and likely to include false statements. Clickstream data is the recording an Internet user leaves behind while visiting websites. As the user clicks anywhere in the webpage, the activity is logged in semi-structured website log files. Such data allows us to see what pages users visited, how long they stayed there, how often they visited, when they usually visited, which site they prefer, what keywords they used to find the site, whether they purchased any, and so forth. For such a reason, some researchers tried to guess the demographics of Internet users by using their clickstream data. They derived various independent variables likely to be correlated to the demographics. The variables include search keyword, frequency and intensity for time, day and month, variety of websites visited, text information for web pages visited, etc. The demographic attributes to predict are also diverse according to the paper, and cover gender, age, job, location, income, education, marital status, presence of children. A variety of data mining methods, such as LSA, SVM, decision tree, neural network, logistic regression, and k-nearest neighbors, were used for prediction model building. However, this research has not yet identified which data mining method is appropriate to predict each demographic variable. Moreover, it is required to review independent variables studied so far and combine them as needed, and evaluate them for building the best prediction model. The objective of this study is to choose clickstream attributes mostly likely to be correlated to the demographics from the results of previous research, and then to identify which data mining method is fitting to predict each demographic attribute. Among the demographic attributes, this paper focus on predicting gender, age, marital status, residence, and job. And from the results of previous research, 64 clickstream attributes are applied to predict the demographic attributes. The overall process of predictive model building is compose of 4 steps. In the first step, we create user profiles which include 64 clickstream attributes and 5 demographic attributes. The second step performs the dimension reduction of clickstream variables to solve the curse of dimensionality and overfitting problem. We utilize three approaches which are based on decision tree, PCA, and cluster analysis. We build alternative predictive models for each demographic variable in the third step. SVM, neural network, and logistic regression are used for modeling. The last step evaluates the alternative models in view of model accuracy and selects the best model. For the experiments, we used clickstream data which represents 5 demographics and 16,962,705 online activities for 5,000 Internet users. IBM SPSS Modeler 17.0 was used for our prediction process, and the 5-fold cross validation was conducted to enhance the reliability of our experiments. As the experimental results, we can verify that there are a specific data mining method well-suited for each demographic variable. For example, age prediction is best performed when using the decision tree based dimension reduction and neural network whereas the prediction of gender and marital status is the most accurate by applying SVM without dimension reduction. We conclude that the online behaviors of the Internet users, captured from the clickstream data analysis, could be well used to predict their demographics, thereby being utilized to the digital marketing.

A Robust Object Detection and Tracking Method using RGB-D Model (RGB-D 모델을 이용한 강건한 객체 탐지 및 추적 방법)

  • Park, Seohee;Chun, Junchul
    • Journal of Internet Computing and Services
    • /
    • v.18 no.4
    • /
    • pp.61-67
    • /
    • 2017
  • Recently, CCTV has been combined with areas such as big data, artificial intelligence, and image analysis to detect various abnormal behaviors and to detect and analyze the overall situation of objects such as people. Image analysis research for this intelligent video surveillance function is progressing actively. However, CCTV images using 2D information generally have limitations such as object misrecognition due to lack of topological information. This problem can be solved by adding the depth information of the object created by using two cameras to the image. In this paper, we perform background modeling using Mixture of Gaussian technique and detect whether there are moving objects by segmenting the foreground from the modeled background. In order to perform the depth information-based segmentation using the RGB information-based segmentation results, stereo-based depth maps are generated using two cameras. Next, the RGB-based segmented region is set as a domain for extracting depth information, and depth-based segmentation is performed within the domain. In order to detect the center point of a robustly segmented object and to track the direction, the movement of the object is tracked by applying the CAMShift technique, which is the most basic object tracking method. From the experiments, we prove the efficiency of the proposed object detection and tracking method using the RGB-D model.

A Study on Risk Issues and Policy for Future Society of Digital Transformation: Focusing on Artificial Intelligence (디지털 전환의 미래사회 위험이슈 및 정책적 대응 방향: 인공지능을 중심으로)

  • Koo, Bonjin
    • Journal of Technology Innovation
    • /
    • v.30 no.1
    • /
    • pp.1-20
    • /
    • 2022
  • Digital transformation refers to the economic and social effects of digitisation and digitalisation. Although digital transformation acts as a useful tool for economic/social development and enhancing the convenience of life, it can have negative effects (misuse of personal information, ethical problems, deepening social gaps, etc.). The government is actively establishing policies to promote digital transformation to secure competitiveness and technological hegemony, however, understanding of digital transformation-related risk issues and implementing policies to prevent them are relatively slow. Thus, this study systematically identifies risk issues of the future society that can be caused by digital transformation based on quantitative analysis of media articles big data through the Embedded Topic Modeling method. Specifically, first, detailed issues of negative effects of digital transformation in major countries were identified. Then detailed issues of negative effects of artificial intelligence in major countries and Korea were identified. Further, by synthesizing the results, future direction of the government's digital transformation policies for responding the negative effects was proposed. The policy implications are as follows. First, since the negative effects of digital transformation does not only affect technological fields but also affect the overall society, such as national security, social issues, and fairness issues. Therefore, the government should not only promote the positive functions of digital transformation, but also prepare policies to counter the negative functions of digital transformation. Second, the detailed issues of future social risks of digital transformation appear differently depending on contexts, so the government should establish a policy to respond to the negative effects of digital transformation in consideration of the national and social context. Third, the government should set a major direction for responding negative effects of digital transformation to minimize confusion among stakeholders, and prepare effective policy measures.

In-service teacher's perception on the mathematical modeling tasks and competency for designing the mathematical modeling tasks: Focused on reality (현직 수학 교사들의 수학적 모델링 과제에 대한 인식과 과제 개발 역량: 현실성을 중심으로)

  • Hwang, Seonyoung;Han, Sunyoung
    • The Mathematical Education
    • /
    • v.62 no.3
    • /
    • pp.381-400
    • /
    • 2023
  • As the era of solving various and complex problems in the real world using artificial intelligence and big data appears, problem-solving competencies that can solve realistic problems through a mathematical approach are required. In fact, the 2015 revised mathematics curriculum and the 2022 revised mathematics curriculum emphasize mathematical modeling as an activity and competency to solve real-world problems. However, the real-world problems presented in domestic and international textbooks have a high proportion of artificial problems that rarely occur in real-world. Accordingly, domestic and international countries are paying attention to the reality of mathematical modeling tasks and suggesting the need for authentic tasks that reflect students' daily lives. However, not only did previous studies focus on theoretical proposals for reality, but studies analyzing teachers' perceptions of reality and their competency to reflect reality in the task are insufficient. Accordingly, this study aims to analyze in-service mathematics teachers' perception of reality among the characteristics of tasks for mathematical modeling and the in-service mathematics teachers' competency for designing the mathematical modeling tasks. First of all, five criteria for satisfying the reality were established by analyzing literatures. Afterward, teacher training was conducted under the theme of mathematical modeling. Pre- and post-surveys for 41 in-service mathematics teachers who participated in the teacher training was conducted to confirm changes in perception of reality. The pre- and post- surveys provided a task that did not reflect reality, and in-service mathematics teachers determined whether the task given in surveys reflected reality and selected one reason for the judgment among five criteria for reality. Afterwards, frequency analysis was conducted by coding the results of the survey answered by in-service mathematics teachers in the pre- and post- survey, and frequencies were compared to confirm in-service mathematics teachers' perception changes on reality. In addition, the mathematical modeling tasks designed by in-service teachers were evaluated with the criteria for reality to confirm the teachers' competency for designing mathematical modeling tasks reflecting the reality. As a result, it was shown that in-service mathematics teachers changed from insufficient perception that only considers fragmentary criterion for reality to perceptions that consider all the five criteria of reality. In particular, as a result of analyzing the basis for judgment among in-service mathematics teachers whose judgment on reality was reversed in the pre- and post-survey, changes in the perception of in-service mathematics teachers was confirmed, who did not consider certain criteria as a criterion for reality in the pre-survey, but considered them as a criterion for reality in the post-survey. In addition, as a result of evaluating the tasks designed by in-service mathematics teachers for mathematical modeling, in-service mathematics teachers showed the competency to reflect reality in their tasks. However, among the five criteria for reality, the criterion for "situations that can occur in students' daily lives," "need to solve the task," and "require conclusions in a real-world situation" were relatively less reflected. In addition, it was found that the proportion of teachers with low task development competencies was higher in the teacher group who could not make the right judgment than in the teacher group who could make the right judgment on the reality of the task. Based on the results of these studies, this study provides implications for teacher education to enable mathematics teachers to apply mathematical modeling lesson in their classes.

Investigating Dynamic Mutation Process of Issues Using Unstructured Text Analysis (비정형 텍스트 분석을 활용한 이슈의 동적 변이과정 고찰)

  • Lim, Myungsu;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.1
    • /
    • pp.1-18
    • /
    • 2016
  • Owing to the extensive use of Web media and the development of the IT industry, a large amount of data has been generated, shared, and stored. Nowadays, various types of unstructured data such as image, sound, video, and text are distributed through Web media. Therefore, many attempts have been made in recent years to discover new value through an analysis of these unstructured data. Among these types of unstructured data, text is recognized as the most representative method for users to express and share their opinions on the Web. In this sense, demand for obtaining new insights through text analysis is steadily increasing. Accordingly, text mining is increasingly being used for different purposes in various fields. In particular, issue tracking is being widely studied not only in the academic world but also in industries because it can be used to extract various issues from text such as news, (SocialNetworkServices) to analyze the trends of these issues. Conventionally, issue tracking is used to identify major issues sustained over a long period of time through topic modeling and to analyze the detailed distribution of documents involved in each issue. However, because conventional issue tracking assumes that the content composing each issue does not change throughout the entire tracking period, it cannot represent the dynamic mutation process of detailed issues that can be created, merged, divided, and deleted between these periods. Moreover, because only keywords that appear consistently throughout the entire period can be derived as issue keywords, concrete issue keywords such as "nuclear test" and "separated families" may be concealed by more general issue keywords such as "North Korea" in an analysis over a long period of time. This implies that many meaningful but short-lived issues cannot be discovered by conventional issue tracking. Note that detailed keywords are preferable to general keywords because the former can be clues for providing actionable strategies. To overcome these limitations, we performed an independent analysis on the documents of each detailed period. We generated an issue flow diagram based on the similarity of each issue between two consecutive periods. The issue transition pattern among categories was analyzed by using the category information of each document. In this study, we then applied the proposed methodology to a real case of 53,739 news articles. We derived an issue flow diagram from the articles. We then proposed the following useful application scenarios for the issue flow diagram presented in the experiment section. First, we can identify an issue that actively appears during a certain period and promptly disappears in the next period. Second, the preceding and following issues of a particular issue can be easily discovered from the issue flow diagram. This implies that our methodology can be used to discover the association between inter-period issues. Finally, an interesting pattern of one-way and two-way transitions was discovered by analyzing the transition patterns of issues through category analysis. Thus, we discovered that a pair of mutually similar categories induces two-way transitions. In contrast, one-way transitions can be recognized as an indicator that issues in a certain category tend to be influenced by other issues in another category. For practical application of the proposed methodology, high-quality word and stop word dictionaries need to be constructed. In addition, not only the number of documents but also additional meta-information such as the read counts, written time, and comments of documents should be analyzed. A rigorous performance evaluation or validation of the proposed methodology should be performed in future works.

Development of Topic Trend Analysis Model for Industrial Intelligence using Public Data (텍스트마이닝을 활용한 공개데이터 기반 기업 및 산업 토픽추이분석 모델 제안)

  • Park, Sunyoung;Lee, Gene Moo;Kim, You-Eil;Seo, Jinny
    • Journal of Technology Innovation
    • /
    • v.26 no.4
    • /
    • pp.199-232
    • /
    • 2018
  • There are increasing needs for understanding and fathoming of business management environment through big data analysis at industrial and corporative level. The research using the company disclosure information, which is comprehensively covering the business performance and the future plan of the company, is getting attention. However, there is limited research on developing applicable analytical models leveraging such corporate disclosure data due to its unstructured nature. This study proposes a text-mining-based analytical model for industrial and firm level analyses using publicly available company disclousre data. Specifically, we apply LDA topic model and word2vec word embedding model on the U.S. SEC data from the publicly listed firms and analyze the trends of business topics at the industrial and corporate levels. Using LDA topic modeling based on SEC EDGAR 10-K document, whole industrial management topics are figured out. For comparison of different pattern of industries' topic trend, software and hardware industries are compared in recent 20 years. Also, the changes of management subject at firm level are observed with comparison of two companies in software industry. The changes of topic trends provides lens for identifying decreasing and growing management subjects at industrial and firm level. Mapping companies and products(or services) based on dimension reduction after using word2vec word embedding model and principal component analysis of 10-K document at firm level in software industry, companies and products(services) that have similar management subjects are identified and also their changes in decades. For suggesting methodology to develop analysis model based on public management data at industrial and corporate level, there may be contributions in terms of making ground of practical methodology to identifying changes of managements subjects. However, there are required further researches to provide microscopic analytical model with regard to relation of technology management strategy between management performance in case of related to various pattern of management topics as of frequent changes of management subject or their momentum. Also more studies are needed for developing competitive context analysis model with product(service)-portfolios between firms.