• Title/Summary/Keyword: semantic network

Search Result 735, Processing Time 0.026 seconds

Exploring Preservice Teachers' Science PCK and the Role of Argumentation Structure as a Pedagogical Reasoning Tool (교수적 추론 도구로서 논증구조를 활용한 과학과 예비교사들의 가족유사성 PCK 특성 탐색)

  • Youngsun Kwak
    • Journal of the Korean Society of Earth Science Education
    • /
    • v.16 no.1
    • /
    • pp.56-71
    • /
    • 2023
  • The purpose of this study is to explore the role and effectiveness of argumentation structure and the developmental characteristics of science PCK with Earth science preservice teachers who used argumentation structure as a pedagogical reasoning tool. Since teachers demonstrate PCK in a series of pedagogical reasoning processes using argumentation structures, we explored the characteristics of future-oriented family resemblance-PCK shown by preservice science teachers using argumentation structures. At the end of the semester, we conducted in-depth interviews with 15 earth science preservice teachers who had experienced lesson design and teaching practice using the argumentation structure. Qualitative analysis including a semantic network analysis was conducted based on the in-depth interview to analyze the characteristics of preservice teachers' family resemblance-PCK. Results include that preservice teachers organized their classes systematically by applying the argumentation structure, and structured classes by differentiating argumentation elements from facts to conclusions. Regarding the characteristics of each component of the argumentation structure, preservice teachers had difficulty finding warrant, rebuttal, and qualifier. The area of PCK most affected by the argumentation structure is the science teaching practice, and preservice teachers emphasized the selection of a instructional model suitable for lesson content, the use of various teaching methods and inquiry activities to persuade lesson content, and developing of data literacy and digital competency. Discussed in the conclusion are the potential and usability of argument structure as a pedagogical reasoning tool, the possibility of developing science inquiry and reasoning competency of secondary school students who experience science classes using argumentation structure, and the need for developing a teacher education protocol using argumentation structure as a pedagogical reasoning tool.

Analysis entrepreneurship trends using keyword analysis of news article Big Data :2013~2022 (뉴스기사 빅데이터의 키워드분석을 활용한 창업 트렌드 분석:2013~2022 )

  • Jaeeog Kim;Byunghoon Jeon
    • Journal of Platform Technology
    • /
    • v.11 no.3
    • /
    • pp.83-97
    • /
    • 2023
  • This research aims to identify startup trends by analyzing a large number of news articles through semantic network analysis. Using the BIGKinds article analysis service provided by the Korea Press Foundation, 330,628 news articles from 19 newspapers from January 2013 to December 2022 were comprehensively analyzed. The study focused on exploring the changes in key issues over the past decade, considering the impact of the social environment and global economic trends on entrepreneurship. We compared the number of news articles and changes in issues before and after the COVID-19 pandemic, and visualized entrepreneurship trends through frequency analysis, relationship analysis, and correlation analysis. The results of the study showed that the top keywords for entrepreneurship-related words are startup activation and commercialization, and the correlation between COVID-19 and entrepreneurship keywords is almost negligible in a linear sense, but the number of news articles decreased during the pandemic, which has an impact. In particular, the most frequently mentioned keywords are Ministry of SMEs and Startups, place is the United States, and person is limited. The agency was the SBA, and the entrepreneurship sector is more affected by social issues than any other sector, with the important characteristics of increased frequency of prompt access. This study supplies essential basic data for understanding and exploring issues and events related to entrepreneurship and suggests future research topics in the field.

  • PDF

A Study on the Perception of Quality of Care Services by Care Workers using Big Data (빅데이터를 활용한 요양보호사의 서비스질 인식에 관한 연구)

  • Han-A Cho
    • Journal of Korean Dental Hygiene Science
    • /
    • v.6 no.1
    • /
    • pp.13-25
    • /
    • 2023
  • Background: This study was conducted to confirm the service quality management of care workers, who are direct service personnel of long-term care insurance for the elderly, using unstructured big data. Methods: Using a textome, this study collected and analyzed unstructured social data related to care workers' service quality. Frequency, TF-IDF, centrality, semantic network, and CONCOR analyses were conducted on the top 50 keywords collected by crawling the data. Results: As a result of frequency analysis, the top-ranked keywords were 'Long-term care services,' 'Care workers,' 'Quality of care services,' 'Long term care,' 'Long term care facilities,' 'Enhancement,' 'Elderly,' 'Treatment,' 'Improvement,' and 'Necessity.' The results of degree centrality and eigenvector centrality were almost the same as those of the frequency analysis. As a result of the CONCOR analysis, it was found that the improvement in the quality of long-term care services, the operation of the long-term care services, the long-term care services system, and the perception of the psychological aspects of the care workers were of high concern. Conclusion: This study contributes to setting various directions for improving the service quality of care workers by presenting perceptions related to the service quality of care workers as a meaningful group.

Semantic Segmentation for Roof Extraction using Official Buildings Information (건물 통합 정보를 이용한 지붕 추출 의미론적 분류)

  • Youm, Sungkwan;Lee, Heekwon;Shin, Kwang-Seong
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.582-583
    • /
    • 2021
  • As the production of new and renewable energy such as solar and wind power has diversified, microgrid systems that can simultaneously produce and consume have been introduced. . In general, a decrease in electricity prices through solar power is expected in summer, so producer protection is required. In this paper, we propose a transparent and safe gift power transaction system between users using blockchain in a microgrid environment. A futures is simply a contract in which the buyer is obligated to buy electricity or the seller is obliged to sell electricity at a fixed price and a predetermined futures price. This system proposes a futures trading algorithm that searches for futures prices and concludes power transactions with automated operations without user intervention by using a smart contract, a reliable executable code within the blockchain network. If a power producer thinks that the price during the peak production period (Hajj) is likely to decrease during production planning, it sells futures first in the futures market and buys back futures during the peak production period (Haj) to make a profit in the spot market. losses can be compensated. In addition, if there is a risk that the price of electricity will rise when a sales contract is concluded, a broker can compensate for a loss in the spot market by first buying futures in the futures market and liquidating futures when the sales contract is fulfilled.

  • PDF

Brain Activation in Generating Hypothesis about Biological Phenomena and the Processing of Mental Arithmetic: An fMRI Study (생명 현상에 대한 과학적 가설 생성과 수리 연산에서 나타나는 두뇌 활성: fMRI 연구)

  • Kwon, Yong-Ju;Shin, Dong-Hoon;Lee, Jun-Ki;Yang, Il-Ho
    • Journal of The Korean Association For Science Education
    • /
    • v.27 no.1
    • /
    • pp.93-104
    • /
    • 2007
  • The purpose of this study is to investigate brain activity both during the processing of a scientific hypothesis about biological phenomena and mental arithmetic using 3.0T fMRI at the KAIST. For this study, 16 healthy male subjects participated voluntarily. Each subject's functional brain images by performing a scientific hypothesis task and a mental arithmetic task for 684 seconds were measured. After the fMRI measuring, verbal reports were collected to ensure the reliability of brain image data. This data, which were found to be adequate based on the results of analyzing verbal reports, were all included in the statistical analysis. When the data were statistically analyzed using SPM2 software, the scientific hypothesis generating process was found to have independent brain network different from the mental arithmetic process. In the scientific hypothesis process, we can infer that there is the process of encoding semantic derived from the fusiform gyrus through question-situation analysis in the pre-frontal lobe. In the mental arithmetic process, the area combining pre-frontal and parietal lobes plays an important role, and the parietal lobe is considered to be involved in skillfulness. In addition, the scientific hypothesis process was found to be accompanied by scientific emotion. These results enabled the examination of the scientific hypothesis process from the cognitive neuroscience perspective, and may be used as basic materials for developing a learning program for scientific hypothesis generation. In addition, this program can be proposed as a model of scientific brain-based learning.

Personalized Recommendation System for IPTV using Ontology and K-medoids (IPTV환경에서 온톨로지와 k-medoids기법을 이용한 개인화 시스템)

  • Yun, Byeong-Dae;Kim, Jong-Woo;Cho, Yong-Seok;Kang, Sang-Gil
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.3
    • /
    • pp.147-161
    • /
    • 2010
  • As broadcasting and communication are converged recently, communication is jointed to TV. TV viewing has brought about many changes. The IPTV (Internet Protocol Television) provides information service, movie contents, broadcast, etc. through internet with live programs + VOD (Video on demand) jointed. Using communication network, it becomes an issue of new business. In addition, new technical issues have been created by imaging technology for the service, networking technology without video cuts, security technologies to protect copyright, etc. Through this IPTV network, users can watch their desired programs when they want. However, IPTV has difficulties in search approach, menu approach, or finding programs. Menu approach spends a lot of time in approaching programs desired. Search approach can't be found when title, genre, name of actors, etc. are not known. In addition, inserting letters through remote control have problems. However, the bigger problem is that many times users are not usually ware of the services they use. Thus, to resolve difficulties when selecting VOD service in IPTV, a personalized service is recommended, which enhance users' satisfaction and use your time, efficiently. This paper provides appropriate programs which are fit to individuals not to save time in order to solve IPTV's shortcomings through filtering and recommendation-related system. The proposed recommendation system collects TV program information, the user's preferred program genres and detailed genre, channel, watching program, and information on viewing time based on individual records of watching IPTV. To look for these kinds of similarities, similarities can be compared by using ontology for TV programs. The reason to use these is because the distance of program can be measured by the similarity comparison. TV program ontology we are using is one extracted from TV-Anytime metadata which represents semantic nature. Also, ontology expresses the contents and features in figures. Through world net, vocabulary similarity is determined. All the words described on the programs are expanded into upper and lower classes for word similarity decision. The average of described key words was measured. The criterion of distance calculated ties similar programs through K-medoids dividing method. K-medoids dividing method is a dividing way to divide classified groups into ones with similar characteristics. This K-medoids method sets K-unit representative objects. Here, distance from representative object sets temporary distance and colonize it. Through algorithm, when the initial n-unit objects are tried to be divided into K-units. The optimal object must be found through repeated trials after selecting representative object temporarily. Through this course, similar programs must be colonized. Selecting programs through group analysis, weight should be given to the recommendation. The way to provide weight with recommendation is as the follows. When each group recommends programs, similar programs near representative objects will be recommended to users. The formula to calculate the distance is same as measure similar distance. It will be a basic figure which determines the rankings of recommended programs. Weight is used to calculate the number of watching lists. As the more programs are, the higher weight will be loaded. This is defined as cluster weight. Through this, sub-TV programs which are representative of the groups must be selected. The final TV programs ranks must be determined. However, the group-representative TV programs include errors. Therefore, weights must be added to TV program viewing preference. They must determine the finalranks.Based on this, our customers prefer proposed to recommend contents. So, based on the proposed method this paper suggested, experiment was carried out in controlled environment. Through experiment, the superiority of the proposed method is shown, compared to existing ways.

A Study on Market Size Estimation Method by Product Group Using Word2Vec Algorithm (Word2Vec을 활용한 제품군별 시장규모 추정 방법에 관한 연구)

  • Jung, Ye Lim;Kim, Ji Hui;Yoo, Hyoung Sun
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.1
    • /
    • pp.1-21
    • /
    • 2020
  • With the rapid development of artificial intelligence technology, various techniques have been developed to extract meaningful information from unstructured text data which constitutes a large portion of big data. Over the past decades, text mining technologies have been utilized in various industries for practical applications. In the field of business intelligence, it has been employed to discover new market and/or technology opportunities and support rational decision making of business participants. The market information such as market size, market growth rate, and market share is essential for setting companies' business strategies. There has been a continuous demand in various fields for specific product level-market information. However, the information has been generally provided at industry level or broad categories based on classification standards, making it difficult to obtain specific and proper information. In this regard, we propose a new methodology that can estimate the market sizes of product groups at more detailed levels than that of previously offered. We applied Word2Vec algorithm, a neural network based semantic word embedding model, to enable automatic market size estimation from individual companies' product information in a bottom-up manner. The overall process is as follows: First, the data related to product information is collected, refined, and restructured into suitable form for applying Word2Vec model. Next, the preprocessed data is embedded into vector space by Word2Vec and then the product groups are derived by extracting similar products names based on cosine similarity calculation. Finally, the sales data on the extracted products is summated to estimate the market size of the product groups. As an experimental data, text data of product names from Statistics Korea's microdata (345,103 cases) were mapped in multidimensional vector space by Word2Vec training. We performed parameters optimization for training and then applied vector dimension of 300 and window size of 15 as optimized parameters for further experiments. We employed index words of Korean Standard Industry Classification (KSIC) as a product name dataset to more efficiently cluster product groups. The product names which are similar to KSIC indexes were extracted based on cosine similarity. The market size of extracted products as one product category was calculated from individual companies' sales data. The market sizes of 11,654 specific product lines were automatically estimated by the proposed model. For the performance verification, the results were compared with actual market size of some items. The Pearson's correlation coefficient was 0.513. Our approach has several advantages differing from the previous studies. First, text mining and machine learning techniques were applied for the first time on market size estimation, overcoming the limitations of traditional sampling based- or multiple assumption required-methods. In addition, the level of market category can be easily and efficiently adjusted according to the purpose of information use by changing cosine similarity threshold. Furthermore, it has a high potential of practical applications since it can resolve unmet needs for detailed market size information in public and private sectors. Specifically, it can be utilized in technology evaluation and technology commercialization support program conducted by governmental institutions, as well as business strategies consulting and market analysis report publishing by private firms. The limitation of our study is that the presented model needs to be improved in terms of accuracy and reliability. The semantic-based word embedding module can be advanced by giving a proper order in the preprocessed dataset or by combining another algorithm such as Jaccard similarity with Word2Vec. Also, the methods of product group clustering can be changed to other types of unsupervised machine learning algorithm. Our group is currently working on subsequent studies and we expect that it can further improve the performance of the conceptually proposed basic model in this study.

Context Sharing Framework Based on Time Dependent Metadata for Social News Service (소셜 뉴스를 위한 시간 종속적인 메타데이터 기반의 컨텍스트 공유 프레임워크)

  • Ga, Myung-Hyun;Oh, Kyeong-Jin;Hong, Myung-Duk;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.4
    • /
    • pp.39-53
    • /
    • 2013
  • The emergence of the internet technology and SNS has increased the information flow and has changed the way people to communicate from one-way to two-way communication. Users not only consume and share the information, they also can create and share it among their friends across the social network service. It also changes the Social Media behavior to become one of the most important communication tools which also includes Social TV. Social TV is a form which people can watch a TV program and at the same share any information or its content with friends through Social media. Social News is getting popular and also known as a Participatory Social Media. It creates influences on user interest through Internet to represent society issues and creates news credibility based on user's reputation. However, the conventional platforms in news services only focus on the news recommendation domain. Recent development in SNS has changed this landscape to allow user to share and disseminate the news. Conventional platform does not provide any special way for news to be share. Currently, Social News Service only allows user to access the entire news. Nonetheless, they cannot access partial of the contents which related to users interest. For example user only have interested to a partial of the news and share the content, it is still hard for them to do so. In worst cases users might understand the news in different context. To solve this, Social News Service must provide a method to provide additional information. For example, Yovisto known as an academic video searching service provided time dependent metadata from the video. User can search and watch partial of video content according to time dependent metadata. They also can share content with a friend in social media. Yovisto applies a method to divide or synchronize a video based whenever the slides presentation is changed to another page. However, we are not able to employs this method on news video since the news video is not incorporating with any power point slides presentation. Segmentation method is required to separate the news video and to creating time dependent metadata. In this work, In this paper, a time dependent metadata-based framework is proposed to segment news contents and to provide time dependent metadata so that user can use context information to communicate with their friends. The transcript of the news is divided by using the proposed story segmentation method. We provide a tag to represent the entire content of the news. And provide the sub tag to indicate the segmented news which includes the starting time of the news. The time dependent metadata helps user to track the news information. It also allows them to leave a comment on each segment of the news. User also may share the news based on time metadata as segmented news or as a whole. Therefore, it helps the user to understand the shared news. To demonstrate the performance, we evaluate the story segmentation accuracy and also the tag generation. For this purpose, we measured accuracy of the story segmentation through semantic similarity and compared to the benchmark algorithm. Experimental results show that the proposed method outperforms benchmark algorithms in terms of the accuracy of story segmentation. It is important to note that sub tag accuracy is the most important as a part of the proposed framework to share the specific news context with others. To extract a more accurate sub tags, we have created stop word list that is not related to the content of the news such as name of the anchor or reporter. And we applied to framework. We have analyzed the accuracy of tags and sub tags which represent the context of news. From the analysis, it seems that proposed framework is helpful to users for sharing their opinions with context information in Social media and Social news.

Analysis of media trends related to spent nuclear fuel treatment technology using text mining techniques (텍스트마이닝 기법을 활용한 사용후핵연료 건식처리기술 관련 언론 동향 분석)

  • Jeong, Ji-Song;Kim, Ho-Dong
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.2
    • /
    • pp.33-54
    • /
    • 2021
  • With the fourth industrial revolution and the arrival of the New Normal era due to Corona, the importance of Non-contact technologies such as artificial intelligence and big data research has been increasing. Convergent research is being conducted in earnest to keep up with these research trends, but not many studies have been conducted in the area of nuclear research using artificial intelligence and big data-related technologies such as natural language processing and text mining analysis. This study was conducted to confirm the applicability of data science analysis techniques to the field of nuclear research. Furthermore, the study of identifying trends in nuclear spent fuel recognition is critical in terms of being able to determine directions to nuclear industry policies and respond in advance to changes in industrial policies. For those reasons, this study conducted a media trend analysis of pyroprocessing, a spent nuclear fuel treatment technology. We objectively analyze changes in media perception of spent nuclear fuel dry treatment techniques by applying text mining analysis techniques. Text data specializing in Naver's web news articles, including the keywords "Pyroprocessing" and "Sodium Cooled Reactor," were collected through Python code to identify changes in perception over time. The analysis period was set from 2007 to 2020, when the first article was published, and detailed and multi-layered analysis of text data was carried out through analysis methods such as word cloud writing based on frequency analysis, TF-IDF and degree centrality calculation. Analysis of the frequency of the keyword showed that there was a change in media perception of spent nuclear fuel dry treatment technology in the mid-2010s, which was influenced by the Gyeongju earthquake in 2016 and the implementation of the new government's energy conversion policy in 2017. Therefore, trend analysis was conducted based on the corresponding time period, and word frequency analysis, TF-IDF, degree centrality values, and semantic network graphs were derived. Studies show that before the 2010s, media perception of spent nuclear fuel dry treatment technology was diplomatic and positive. However, over time, the frequency of keywords such as "safety", "reexamination", "disposal", and "disassembly" has increased, indicating that the sustainability of spent nuclear fuel dry treatment technology is being seriously considered. It was confirmed that social awareness also changed as spent nuclear fuel dry treatment technology, which was recognized as a political and diplomatic technology, became ambiguous due to changes in domestic policy. This means that domestic policy changes such as nuclear power policy have a greater impact on media perceptions than issues of "spent nuclear fuel processing technology" itself. This seems to be because nuclear policy is a socially more discussed and public-friendly topic than spent nuclear fuel. Therefore, in order to improve social awareness of spent nuclear fuel processing technology, it would be necessary to provide sufficient information about this, and linking it to nuclear policy issues would also be a good idea. In addition, the study highlighted the importance of social science research in nuclear power. It is necessary to apply the social sciences sector widely to the nuclear engineering sector, and considering national policy changes, we could confirm that the nuclear industry would be sustainable. However, this study has limitations that it has applied big data analysis methods only to detailed research areas such as "Pyroprocessing," a spent nuclear fuel dry processing technology. Furthermore, there was no clear basis for the cause of the change in social perception, and only news articles were analyzed to determine social perception. Considering future comments, it is expected that more reliable results will be produced and efficiently used in the field of nuclear policy research if a media trend analysis study on nuclear power is conducted. Recently, the development of uncontact-related technologies such as artificial intelligence and big data research is accelerating in the wake of the recent arrival of the New Normal era caused by corona. Convergence research is being conducted in earnest in various research fields to follow these research trends, but not many studies have been conducted in the nuclear field with artificial intelligence and big data-related technologies such as natural language processing and text mining analysis. The academic significance of this study is that it was possible to confirm the applicability of data science analysis technology in the field of nuclear research. Furthermore, due to the impact of current government energy policies such as nuclear power plant reductions, re-evaluation of spent fuel treatment technology research is undertaken, and key keyword analysis in the field can contribute to future research orientation. It is important to consider the views of others outside, not just the safety technology and engineering integrity of nuclear power, and further reconsider whether it is appropriate to discuss nuclear engineering technology internally. In addition, if multidisciplinary research on nuclear power is carried out, reasonable alternatives can be prepared to maintain the nuclear industry.

Evaluation of Web Service Similarity Assessment Methods (웹서비스 유사성 평가 방법들의 실험적 평가)

  • Hwang, You-Sub
    • Journal of Intelligence and Information Systems
    • /
    • v.15 no.4
    • /
    • pp.1-22
    • /
    • 2009
  • The World Wide Web is transitioning from being a mere collection of documents that contain useful information toward providing a collection of services that perform useful tasks. The emerging Web service technology has been envisioned as the next technological wave and is expected to play an important role in this recent transformation of the Web. By providing interoperable interface standards for application-to-application communication, Web services can be combined with component based software development to promote application interaction and integration both within and across enterprises. To make Web services for service-oriented computing operational, it is important that Web service repositories not only be well-structured but also provide efficient tools for developers to find reusable Web service components that meet their needs. As the potential of Web services for service-oriented computing is being widely recognized, the demand for effective Web service discovery mechanisms is concomitantly growing. A number of techniques for Web service discovery have been proposed, but the discovery challenge has not been satisfactorily addressed. Unfortunately, most existing solutions are either too rudimentary to be useful or too domain dependent to be generalizable. In this paper, we propose a Web service organizing framework that combines clustering techniques with string matching and leverages the semantics of the XML-based service specification in WSDL documents. We believe that this is one of the first attempts at applying data mining techniques in the Web service discovery domain. Our proposed approach has several appealing features : (1) It minimizes the requirement of prior knowledge from both service consumers and publishers; (2) It avoids exploiting domain dependent ontologies; and (3) It is able to visualize the semantic relationships among Web services. We have developed a prototype system based on the proposed framework using an unsupervised artificial neural network and empirically evaluated the proposed approach and tool using real Web service descriptions drawn from operational Web service registries. We report on some preliminary results demonstrating the efficacy of the proposed approach.

  • PDF