• Title/Summary/Keyword: Data convert

Search Result 661, Processing Time 0.024 seconds

Multi-Vector Document Embedding Using Semantic Decomposition of Complex Documents (복합 문서의 의미적 분해를 통한 다중 벡터 문서 임베딩 방법론)

  • Park, Jongin;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.19-41
    • /
    • 2019
  • According to the rapidly increasing demand for text data analysis, research and investment in text mining are being actively conducted not only in academia but also in various industries. Text mining is generally conducted in two steps. In the first step, the text of the collected document is tokenized and structured to convert the original document into a computer-readable form. In the second step, tasks such as document classification, clustering, and topic modeling are conducted according to the purpose of analysis. Until recently, text mining-related studies have been focused on the application of the second steps, such as document classification, clustering, and topic modeling. However, with the discovery that the text structuring process substantially influences the quality of the analysis results, various embedding methods have actively been studied to improve the quality of analysis results by preserving the meaning of words and documents in the process of representing text data as vectors. Unlike structured data, which can be directly applied to a variety of operations and traditional analysis techniques, Unstructured text should be preceded by a structuring task that transforms the original document into a form that the computer can understand before analysis. It is called "Embedding" that arbitrary objects are mapped to a specific dimension space while maintaining algebraic properties for structuring the text data. Recently, attempts have been made to embed not only words but also sentences, paragraphs, and entire documents in various aspects. Particularly, with the demand for analysis of document embedding increases rapidly, many algorithms have been developed to support it. Among them, doc2Vec which extends word2Vec and embeds each document into one vector is most widely used. However, the traditional document embedding method represented by doc2Vec generates a vector for each document using the whole corpus included in the document. This causes a limit that the document vector is affected by not only core words but also miscellaneous words. Additionally, the traditional document embedding schemes usually map each document into a single corresponding vector. Therefore, it is difficult to represent a complex document with multiple subjects into a single vector accurately using the traditional approach. In this paper, we propose a new multi-vector document embedding method to overcome these limitations of the traditional document embedding methods. This study targets documents that explicitly separate body content and keywords. In the case of a document without keywords, this method can be applied after extract keywords through various analysis methods. However, since this is not the core subject of the proposed method, we introduce the process of applying the proposed method to documents that predefine keywords in the text. The proposed method consists of (1) Parsing, (2) Word Embedding, (3) Keyword Vector Extraction, (4) Keyword Clustering, and (5) Multiple-Vector Generation. The specific process is as follows. all text in a document is tokenized and each token is represented as a vector having N-dimensional real value through word embedding. After that, to overcome the limitations of the traditional document embedding method that is affected by not only the core word but also the miscellaneous words, vectors corresponding to the keywords of each document are extracted and make up sets of keyword vector for each document. Next, clustering is conducted on a set of keywords for each document to identify multiple subjects included in the document. Finally, a Multi-vector is generated from vectors of keywords constituting each cluster. The experiments for 3.147 academic papers revealed that the single vector-based traditional approach cannot properly map complex documents because of interference among subjects in each vector. With the proposed multi-vector based method, we ascertained that complex documents can be vectorized more accurately by eliminating the interference among subjects.

Energy Budgets for the Developmental Stages of Palaemon macrodactylus (Palaemon macrodactylus의 생활사에 따른 에너지 수지)

  • CHIN Pyung;KIM Heung-Yun;SIN Yun-Kyong
    • Korean Journal of Fisheries and Aquatic Sciences
    • /
    • v.25 no.5
    • /
    • pp.341-358
    • /
    • 1992
  • In order to estimate energy budgets of Palaemon macrodactylus, larvae of the shrimp were reared in the laboratory at constant conditions $(25^{\circ}C: 31-32\%o),$ and then juvenile to adult of the shrimp were reared at $15^{\circ}C\;and\;25^{\circ}C$ in the laboratory. Energy used by the reared shrimps were calculated from estimates of data on feeding, growth, molting, metabolism, nitrogen excretion, and energy content. Juveniles and adults reared in the laboratory, which fed on Artemia nauplii, had an average daily growth rates of 0.079 mm/day at $15^{\circ}C\;and\;of\;0.122mm/day\;at\;25^{\circ}C$. The average growth factor* of P. macrodactylus males and females ranged from $3.2\%$ for adult to $13.2\%$ for juveniles individuals, respectively. Intermolt periods were related to body size of the shrimp and to temperature. Average laboratory growth curves were calculated from data on growth factors and intermolt periods to body size of the shrimp at $15^{\circ}C\;and\;25^{\circ}C$. The calorie contents of the shrimp, their molts, eggs and larvae were determined by biochemical composition and oxygen bomb calorimetry. The average amount of energy used in growth for larvae and juvenile to adult were 4.94 cal and 4.55 cal per dry weight in milligram, respectively. The ammount of oxygen used in metabolism was calculated from size, temperature-specific respiration rate. To convert the ammount of oxygen used in respiration into the equivalent energy lost heat was estimated from the data on chemical composition for the larvae and adult, the values was 4.58 cal/ml $O_2$. The energy content per egg was 0.078 cal. The assimilation efficiency estimated by nitrogen content of food and egested faeces gave $61.5\%$ for the larvae. The efficiencies for juvenile to adult ranged between $79.4\%$ and $90.1\%$ The gross growth efficiencies $(K_1)$ and net growth efficiencies $(K_2)$ of P macrodactylus showed $18.33\%\;and 32.63\%$ for total larval stages, ranged from $21.30\%\;to\;31.04\%\;and\;from\;30.03\%\;to\;39.34\%$ for juvenile to adult, respectively.

  • PDF

Knowledge Extraction Methodology and Framework from Wikipedia Articles for Construction of Knowledge-Base (지식베이스 구축을 위한 한국어 위키피디아의 학습 기반 지식추출 방법론 및 플랫폼 연구)

  • Kim, JaeHun;Lee, Myungjin
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.43-61
    • /
    • 2019
  • Development of technologies in artificial intelligence has been rapidly increasing with the Fourth Industrial Revolution, and researches related to AI have been actively conducted in a variety of fields such as autonomous vehicles, natural language processing, and robotics. These researches have been focused on solving cognitive problems such as learning and problem solving related to human intelligence from the 1950s. The field of artificial intelligence has achieved more technological advance than ever, due to recent interest in technology and research on various algorithms. The knowledge-based system is a sub-domain of artificial intelligence, and it aims to enable artificial intelligence agents to make decisions by using machine-readable and processible knowledge constructed from complex and informal human knowledge and rules in various fields. A knowledge base is used to optimize information collection, organization, and retrieval, and recently it is used with statistical artificial intelligence such as machine learning. Recently, the purpose of the knowledge base is to express, publish, and share knowledge on the web by describing and connecting web resources such as pages and data. These knowledge bases are used for intelligent processing in various fields of artificial intelligence such as question answering system of the smart speaker. However, building a useful knowledge base is a time-consuming task and still requires a lot of effort of the experts. In recent years, many kinds of research and technologies of knowledge based artificial intelligence use DBpedia that is one of the biggest knowledge base aiming to extract structured content from the various information of Wikipedia. DBpedia contains various information extracted from Wikipedia such as a title, categories, and links, but the most useful knowledge is from infobox of Wikipedia that presents a summary of some unifying aspect created by users. These knowledge are created by the mapping rule between infobox structures and DBpedia ontology schema defined in DBpedia Extraction Framework. In this way, DBpedia can expect high reliability in terms of accuracy of knowledge by using the method of generating knowledge from semi-structured infobox data created by users. However, since only about 50% of all wiki pages contain infobox in Korean Wikipedia, DBpedia has limitations in term of knowledge scalability. This paper proposes a method to extract knowledge from text documents according to the ontology schema using machine learning. In order to demonstrate the appropriateness of this method, we explain a knowledge extraction model according to the DBpedia ontology schema by learning Wikipedia infoboxes. Our knowledge extraction model consists of three steps, document classification as ontology classes, proper sentence classification to extract triples, and value selection and transformation into RDF triple structure. The structure of Wikipedia infobox are defined as infobox templates that provide standardized information across related articles, and DBpedia ontology schema can be mapped these infobox templates. Based on these mapping relations, we classify the input document according to infobox categories which means ontology classes. After determining the classification of the input document, we classify the appropriate sentence according to attributes belonging to the classification. Finally, we extract knowledge from sentences that are classified as appropriate, and we convert knowledge into a form of triples. In order to train models, we generated training data set from Wikipedia dump using a method to add BIO tags to sentences, so we trained about 200 classes and about 2,500 relations for extracting knowledge. Furthermore, we evaluated comparative experiments of CRF and Bi-LSTM-CRF for the knowledge extraction process. Through this proposed process, it is possible to utilize structured knowledge by extracting knowledge according to the ontology schema from text documents. In addition, this methodology can significantly reduce the effort of the experts to construct instances according to the ontology schema.

Resolving the 'Gray sheep' Problem Using Social Network Analysis (SNA) in Collaborative Filtering (CF) Recommender Systems (소셜 네트워크 분석 기법을 활용한 협업필터링의 특이취향 사용자(Gray Sheep) 문제 해결)

  • Kim, Minsung;Im, Il
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.2
    • /
    • pp.137-148
    • /
    • 2014
  • Recommender system has become one of the most important technologies in e-commerce in these days. The ultimate reason to shop online, for many consumers, is to reduce the efforts for information search and purchase. Recommender system is a key technology to serve these needs. Many of the past studies about recommender systems have been devoted to developing and improving recommendation algorithms and collaborative filtering (CF) is known to be the most successful one. Despite its success, however, CF has several shortcomings such as cold-start, sparsity, gray sheep problems. In order to be able to generate recommendations, ordinary CF algorithms require evaluations or preference information directly from users. For new users who do not have any evaluations or preference information, therefore, CF cannot come up with recommendations (Cold-star problem). As the numbers of products and customers increase, the scale of the data increases exponentially and most of the data cells are empty. This sparse dataset makes computation for recommendation extremely hard (Sparsity problem). Since CF is based on the assumption that there are groups of users sharing common preferences or tastes, CF becomes inaccurate if there are many users with rare and unique tastes (Gray sheep problem). This study proposes a new algorithm that utilizes Social Network Analysis (SNA) techniques to resolve the gray sheep problem. We utilize 'degree centrality' in SNA to identify users with unique preferences (gray sheep). Degree centrality in SNA refers to the number of direct links to and from a node. In a network of users who are connected through common preferences or tastes, those with unique tastes have fewer links to other users (nodes) and they are isolated from other users. Therefore, gray sheep can be identified by calculating degree centrality of each node. We divide the dataset into two, gray sheep and others, based on the degree centrality of the users. Then, different similarity measures and recommendation methods are applied to these two datasets. More detail algorithm is as follows: Step 1: Convert the initial data which is a two-mode network (user to item) into an one-mode network (user to user). Step 2: Calculate degree centrality of each node and separate those nodes having degree centrality values lower than the pre-set threshold. The threshold value is determined by simulations such that the accuracy of CF for the remaining dataset is maximized. Step 3: Ordinary CF algorithm is applied to the remaining dataset. Step 4: Since the separated dataset consist of users with unique tastes, an ordinary CF algorithm cannot generate recommendations for them. A 'popular item' method is used to generate recommendations for these users. The F measures of the two datasets are weighted by the numbers of nodes and summed to be used as the final performance metric. In order to test performance improvement by this new algorithm, an empirical study was conducted using a publically available dataset - the MovieLens data by GroupLens research team. We used 100,000 evaluations by 943 users on 1,682 movies. The proposed algorithm was compared with an ordinary CF algorithm utilizing 'Best-N-neighbors' and 'Cosine' similarity method. The empirical results show that F measure was improved about 11% on average when the proposed algorithm was used

    . Past studies to improve CF performance typically used additional information other than users' evaluations such as demographic data. Some studies applied SNA techniques as a new similarity metric. This study is novel in that it used SNA to separate dataset. This study shows that performance of CF can be improved, without any additional information, when SNA techniques are used as proposed. This study has several theoretical and practical implications. This study empirically shows that the characteristics of dataset can affect the performance of CF recommender systems. This helps researchers understand factors affecting performance of CF. This study also opens a door for future studies in the area of applying SNA to CF to analyze characteristics of dataset. In practice, this study provides guidelines to improve performance of CF recommender systems with a simple modification.

  • A study on the Standardization of Design Guidelines for Geographic Information Databases (지리정보 DB 설계 지침의 표준화 연구)

    • Lim, Duk-Sung;Moon, Sang-Ho;Si, Jong-Ik;Hong, Bong-Hee
      • Journal of Korea Spatial Information System Society
      • /
      • v.5 no.1 s.9
      • /
      • pp.49-63
      • /
      • 2003
    • Recently, two international standard organizations, ISO and OGC, have done the work of standardization for GIS. Current standardization work for providing interoperability among GIS DB focuses on the design of open interfaces. But, this work has not considered procedures and methods for designing GIS DB. Eventually, GIS DB has its own model. When we share the data by open interface among heterogeneous GIS DB, differences between models result in the loss of information. Our aim in this paper is to revise the design guidelines for geographic information databases in order to make consistent spatial data models, logical structures, and semantic structure of populated geographical databases. In details, we propose standard guidelines which convert ISO abstract schema into relation model, object-relation model, object-centered model, and geometry-centered model. Furthermore, we provide sample models for applying these guidelines in commercial GIS S/Ws. Building GIS DB based on design guidelines proposed in the paper has the following advantages: the interoperability among databases, the standardization of schema definitions, and the catalogue of GIS databases through.

    • PDF

    The feasibility evaluation of Respiratory Gated radiation therapy simulation according to the Respiratory Training with lung cancer (폐암 환자의 호흡훈련에 의한 호흡동조 방사선치료계획의 유용성 평가)

    • Hong, mi ran;Kim, cheol jong;Park, soo yeon;Choi, jae won;Pyo, hong ryeol
      • The Journal of Korean Society for Radiation Therapy
      • /
      • v.28 no.2
      • /
      • pp.149-159
      • /
      • 2016
    • Purpose : To evaluate the usefulness of the breathing exercise,we analyzed the change in the RPM signal and the diaphragm imagebefore 4D respiratory gated radiation therapy planning of lung cancer patients. Materials and Methods : The breathing training was enforced on 11 patients getting the 4D respiratory gated radiation therapy from April, 2016 until August. At the same time, RPM signal and diaphragm image was obtained respiration training total three steps in step 1 signal acquisition of free-breathing state, 2 steps respiratory signal acquisition through the guide of the respiratory signal, 3 steps, won the regular respiration signal to the description and repeat training. And then, acquired the minimum value, maximum value, average value, and a standard deviation of the inspiration and expiration in RPM signal and diaphragm image in each steps. Were normalized by the value of the step 1, to convert the 2,3 steps to the other distribution ratio (%), by evaluating the change in the interior of the respiratory motion of the patient, it was evaluated breathing exercise usefulness of each patient. Results : The mean value and the standard deviation of each step were obtained with the procedure 1 of the RPM signal and the diaphragm amplitude as a 100% reference. In the RPM signal, the amplitudes and standard deviations of four patients (36.4%, eleven) decreased by 18.1%, 27.6% on average in 3 steps, and 2 patients (18.2%, 11 people) had standard deviation, It decreased by an average of 36.5%. Meanwhile, the other four patients (36.4%, eleven) decreased by an average of only amplitude 13.1%. In Step 3, the amplitude of the diaphragm image decreased by 30% on average of 9 patients (81.8%, 11 people), and the average of 2 patients (18.2%, 11 people) increased by 7.3%. However, the amplitudes of RPM signals and diaphragm image in 3steps were reduced by 52.6% and 42.1% on average from all patients, respectively, compared to the 2 steps. Relationship between RPM signal and diaphragm image amplitude difference was consistent with patterns of movement 1, 2 and 3steps, respectively, except for No. 2 No. 10 patients. Conclusion : It is possible to induce an optimized respiratory cycle when respiratory training is done. By conducting respiratory training before treatment, it was possible to expect the effect of predicting the movement of the lung which could control the patient's respiration. Ultimately, it can be said that breathing exercises are useful because it is possible to minimize the systematic error of radiotherapy, expect more accurate treatment. In this study, it is limited to research analyzed based on data on respiratory training before treatment, and it will be necessary to verify with the actual CT plan and the data acquired during treatment in the future.

    • PDF

    Corporate Credit Rating based on Bankruptcy Probability Using AdaBoost Algorithm-based Support Vector Machine (AdaBoost 알고리즘기반 SVM을 이용한 부실 확률분포 기반의 기업신용평가)

    • Shin, Taek-Soo;Hong, Tae-Ho
      • Journal of Intelligence and Information Systems
      • /
      • v.17 no.3
      • /
      • pp.25-41
      • /
      • 2011
    • Recently, support vector machines (SVMs) are being recognized as competitive tools as compared with other data mining techniques for solving pattern recognition or classification decision problems. Furthermore, many researches, in particular, have proved them more powerful than traditional artificial neural networks (ANNs) (Amendolia et al., 2003; Huang et al., 2004, Huang et al., 2005; Tay and Cao, 2001; Min and Lee, 2005; Shin et al., 2005; Kim, 2003).The classification decision, such as a binary or multi-class decision problem, used by any classifier, i.e. data mining techniques is so cost-sensitive particularly in financial classification problems such as the credit ratings that if the credit ratings are misclassified, a terrible economic loss for investors or financial decision makers may happen. Therefore, it is necessary to convert the outputs of the classifier into wellcalibrated posterior probabilities-based multiclass credit ratings according to the bankruptcy probabilities. However, SVMs basically do not provide such probabilities. So it required to use any method to create the probabilities (Platt, 1999; Drish, 2001). This paper applied AdaBoost algorithm-based support vector machines (SVMs) into a bankruptcy prediction as a binary classification problem for the IT companies in Korea and then performed the multi-class credit ratings of the companies by making a normal distribution shape of posterior bankruptcy probabilities from the loss functions extracted from the SVMs. Our proposed approach also showed that their methods can minimize the misclassification problems by adjusting the credit grade interval ranges on condition that each credit grade for credit loan borrowers has its own credit risk, i.e. bankruptcy probability.

    Digital Hologram Compression Technique By Hybrid Video Coding (하이브리드 비디오 코팅에 의한 디지털 홀로그램 압축기술)

    • Seo, Young-Ho;Choi, Hyun-Jun;Kang, Hoon-Jong;Lee, Seung-Hyun;Kim, Dong-Wook
      • Journal of the Institute of Electronics Engineers of Korea SP
      • /
      • v.42 no.5 s.305
      • /
      • pp.29-40
      • /
      • 2005
    • According as base of digital hologram has been magnified, discussion of compression technology is expected as a international standard which defines the compression technique of 3D image and video has been progressed in form of 3DAV which is a part of MPEG. As we can identify in case of 3DAV, the coding technique has high possibility to be formed into the hybrid type which is a merged, refined, or mixid with the various previous technique. Therefore, we wish to present the relationship between various image/video coding techniques and digital hologram In this paper, we propose an efficient coding method of digital hologram using standard compression tools for video and image. At first, we convert fringe patterns into video data using a principle of CGH(Computer Generated Hologram), and then encode it. In this research, we propose a compression algorithm is made up of various method such as pre-processing for transform, local segmentation with global information of object image, frequency transform for coding, scanning to make fringe to video stream, classification of coefficients, and hybrid video coding. Finally the proposed hybrid compression algorithm is all of these methods. The tool for still image coding is JPEG2000, and the toots for video coding include various international compression algorithm such as MPEG-2, MPEG-4, and H.264 and various lossless compression algorithm. The proposed algorithm illustrated that it have better properties for reconstruction than the previous researches on far greater compression rate above from four times to eight times as much. Therefore we expect that the proposed technique for digital hologram coding is to be a good preceding research.

    Analysis of Changes in Pine Forests According to Natural Forest Dynamics Using Time-series NFI Data (시계열 국가산림자원조사 자료 기반 자연적 임분동태 변화에 따른 소나무림의 감소 특성 평가)

    • Eun-Sook Kim;Jong Bin Jung;Sinyoung Park
      • Journal of Korean Society of Forest Science
      • /
      • v.113 no.1
      • /
      • pp.40-50
      • /
      • 2024
    • Pine forests are continuously declining due to competition with broadleaf trees, such as oaks, as a consequence of changes in the natural dynamics of forest ecosystem. This natural decline creates a risk of losing the various benefits pine trees have provided to people in the past. Therefore, it is necessary to prepare future forest management directions by considering the state of pine tree decline in each region. The goal of this study is to understand the characteristics of pine forest changes according to forest dynamics and to predict future regional changes. For this purpose, we evaluated the trend of change in pine forests and extracted various variables(topography, forest stand type, disturbance, and climate) that affect the change, using time-series National Forest Inventory (NFI) data. Also, using selected key variables, a model was developed to predict future changes in pine forests. As a results, it showed that the importance of pine trees in forests across the country has decreased overall over the past 10 years. Also, 75% of the sample points representing pine trees remained unchanged, while the remaining 25% had changed to mixed forests. It was found that these changes mainly occurred in areas with good moisture conditions or disturbance factors inside and outside the forest. In the next 10 years, approximately 14.2% of current pine forests was predicted to convert to mixed forests due to changes in natural forest dynamics. Regionally, the rate of pine forest change was highest in Jeju(42.8%) and Gyeonggi(26.9%) and lowest in Gyeongbuk(8.8%) and Gangwon(13.8%). It was predicted that pine forests would be at a high risk of decline in western areas of the Korean Peninsula, including Gyeonggi, Chungcheong, and Jeonnam. This results can be used to make a management plan for pine forests throughout the country.

    Summative Evaluation of 1993, 1994 Discussion Contest of Scientific Investigation (제 1, 2회 학생 과학 공동탐구 토론대회의 종합적 평가)

    • Kim, Eun-Sook;Yoon, Hye-Gyoung
      • Journal of The Korean Association For Science Education
      • /
      • v.16 no.4
      • /
      • pp.376-388
      • /
      • 1996
    • The first and the second "Discussion Contest of Scientific Investigation" was evaluated in this study. This contest was a part of 'Korean Youth Science Festival' held in 1993 and 1994. The evaluation was based on the data collected from the middle school students of final teams, their teachers, a large number of middle school students and college students who were audience of the final competition. Questionnaires, interviews, reports of final teams, and video tape of final competition were used to collect data. The study focussed on three research questions. The first was about the preparation and the research process of students of final teams. The second was about the format and the proceeding of the Contest. The third was whether participating the Contest was useful experience for the students and the teachers of the final teams. The first area, the preparation and the research process of students, were investigated in three aspects. One was the level of cooperation, participation, support and the role of teachers. The second was the information search and experiment, and the third was the report writing. The students of the final teams from both years, had positive opinion about the cooperation, students' active involvement, and support from family and school. Students considered their teachers to be a guide or a counsellor, showing their level of active participation. On the other hand, the interview of 1993 participants showed that there were times that teachers took strong leading role. Therefore one can conclude that students took active roles most of the time while the room for improvement still exists. To search the information they need during the period of the preparation, student visited various places such as libraries, bookstores, universities, and research institutes. Their search was not limited to reading the books, although the books were primary source of information. Students also learned how to organize the information they found and considered leaning of organizing skill useful and fun. Variety of experiments was an important part of preparation and students had positive opinion about it. Understanding related theory was considered most difficult and important, while designing and building proper equipments was considered difficult but not important. This reflects the students' school experience where the equipments were all set in advance and students were asked to confirm the theories presented in the previous class hours. About the reports recording the research process, students recognize the importance and the necessity of the report but had difficulty in writing it. Their reports showed tendency to list everything they did without clear connection to the problem to be solved. Most of the reports did not record the references and some of them confused report writing with story telling. Therefore most of them need training in writing the reports. It is also desirable to describe the process of student learning when theory or mathematics that are beyond the level of middle school curriculum were used because it is part of their investigation. The second area of evaluation was about the format and the proceeding of the Contest, the problems given to students, and the process of student discussion. The format of the Contests, which consisted of four parts, presentation, refutation, debate and review, received good evaluation from students because it made students think more and gave more difficult time but was meaningful and helped to remember longer time according to students. On the other hand, students said the time given to each part of the contest was too short. The problems given to students were short and open ended to stimulate students' imagination and to offer various possible routes to the solution. This type of problem was very unfamiliar and gave a lot of difficulty to students. Student had positive opinion about the research process they experienced but did not recognize the fact that such a process was possible because of the oneness of the task. The level of the problems was rated as too difficult by teachers and college students but as appropriate by the middle school students in audience and participating students. This suggests that it is possible for student to convert the problems to be challengeable and intellectually satisfactory appropriate for their level of understanding even when the problems were difficult for middle school students. During the process of student discussion, a few problems were observed. Some problems were related to the technics of the discussion, such as inappropriate behavior for the role he/she was taking, mismatching answers to the questions. Some problems were related to thinking. For example, students thinking was off balanced toward deductive reasoning, and reasoning based on experimental data was weak. The last area of evaluation was the effect of the Contest. It was measured through the change of the attitude toward science and science classes, and willingness to attend the next Contest. According to the result of the questionnaire, no meaningful change in attitude was observed. However, through the interview several students were observed to have significant positive change in attitude while no student with negative change was observed. Most of the students participated in Contest said they would participate again or recommend their friend to participate. Most of the teachers agreed that the Contest should continue and they would recommend their colleagues or students to participate. As described above, the "Discussion Contest of Scientific Investigation", which was developed and tried as a new science contest, had positive response from participating students and teachers, and the audience. Two among the list of results especially demonstrated that the goal of the Contest, "active and cooperative science learning experience", was reached. One is the fact that students recognized the experience of cooperation, discussion, information search, variety of experiments to be fun and valuable. The other is the fact that the students recognized the format of the contest consisting of presentation, refutation, discussion and review, required more thinking and was challenging, but was more meaningful. Despite a few problems such as, unfamiliarity with the technics of discussion, weakness in inductive and/or experiment based reasoning, and difficulty in report writing, The Contest demonstrated the possibility of new science learning environment and science contest by offering the chance to challenge open tasks by utilizing student science knowledge and ability to inquire and to discuss rationally and critically with other students.

    • PDF

    (34141) Korea Institute of Science and Technology Information, 245, Daehak-ro, Yuseong-gu, Daejeon
    Copyright (C) KISTI. All Rights Reserved.