• Title/Summary/Keyword: model processing

Search Result 8,687, Processing Time 0.039 seconds

Prediction of Correct Answer Rate and Identification of Significant Factors for CSAT English Test Based on Data Mining Techniques (데이터마이닝 기법을 활용한 대학수학능력시험 영어영역 정답률 예측 및 주요 요인 분석)

  • Park, Hee Jin;Jang, Kyoung Ye;Lee, Youn Ho;Kim, Woo Je;Kang, Pil Sung
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.4 no.11
    • /
    • pp.509-520
    • /
    • 2015
  • College Scholastic Ability Test(CSAT) is a primary test to evaluate the study achievement of high-school students and used by most universities for admission decision in South Korea. Because its level of difficulty is a significant issue to both students and universities, the government makes a huge effort to have a consistent difficulty level every year. However, the actual levels of difficulty have significantly fluctuated, which causes many problems with university admission. In this paper, we build two types of data-driven prediction models to predict correct answer rate and to identify significant factors for CSAT English test through accumulated test data of CSAT, unlike traditional methods depending on experts' judgments. Initially, we derive candidate question-specific factors that can influence the correct answer rate, such as the position, EBS-relation, readability, from the annual CSAT practices and CSAT for 10 years. In addition, we drive context-specific factors by employing topic modeling which identify the underlying topics over the text. Then, the correct answer rate is predicted by multiple linear regression and level of difficulty is predicted by classification tree. The experimental results show that 90% of accuracy can be achieved by the level of difficulty (difficult/easy) classification model, whereas the error rate for correct answer rate is below 16%. Points and problem category are found to be critical to predict the correct answer rate. In addition, the correct answer rate is also influenced by some of the topics discovered by topic modeling. Based on our study, it will be possible to predict the range of expected correct answer rate for both question-level and entire test-level, which will help CSAT examiners to control the level of difficulties.

A Study on Establishment of the Optimum Mountain Meteorological Observation Network System for Forest Fire Prevention (산불 방지를 위한 산악기상관측시스템 구축방안)

  • Lee, Si-Young;Chung, Il-Ung;Kim, Sang-Kook
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.8 no.1
    • /
    • pp.36-44
    • /
    • 2006
  • In this study, we constructed a forest fire danger map in the Yeongdong area of Gangwon-do and Northeastern area of Gyeongsangbuk-do using a forest fire rating model and geographical information system (GIS). We investigated the appropriate positions of the automatic weather station (AWS) and a comprehensive network solution (a system including measurement, communication and data processing) for the establishment of an optimum mountain meteorological observation network system (MMONS). Also, we suggested a possible plan for combining the MMONS with unmanned monitoring camera systems and wireless relay towers operated by local governments and the Korea Forest Service for prevention of forest fire.

Estimation of the Optimum Number of Machines and Equipments for Professional Dairy Farm (낙농 전업농의 기계장치 최적 규모 추정)

  • 유병기;이용범;장진택;이동현;권두중;기광석;성시흥;이대원
    • Journal of Animal Environmental Science
    • /
    • v.2 no.1
    • /
    • pp.27-39
    • /
    • 1996
  • A survey was conducted for dairy farmer to estimate the optimum number of machine and equipment in 1994. Labor hours, operation costs and operation methods for each dairy processing were investigated and analyzed for the farmers to find the expected numbers of machine and equipment on the basis of the desired farm scale. And also, the estimated models were compared and analyzed with the conventional models which more than half dairy farmers used bucket milker in tie stall barn. Some of the results are as follows : 1. Analysis results of conventional model showed that a dairy farm could raise to 15 heads of dairy cow with family labor of 1.5 men, labor hours of 2, 700 in you and total operation costs of 734 thousand won per head. 2. The result, used in conjunction with minimum operation costs in tie stall barn, showed that 28 dairy cows could be raised by using concentrates feeding by hoppers, water supply by water cups, milking by pipeline milker, and manure cleaning by barn cleaner with total operation costs of 520 thousands won per head. 3. The total operation costs of a loose barn system is higher than those of tie stall barn system to raise about 30 heads. For the loose barn system, the herringbone parlour was used for milking, concentrate feeding by automatic concentrate feeder, water supply by thermal insulation feeder, and manure cleaning by scraper with total operation costs of 582 thousands won per head every year.

  • PDF

Design and Implementation of Quality Broker Architecture to Web Service Selection based on Autonomic Feedback (자율적 피드백 기반 웹 서비스 선정을 위한 품질 브로커 아키텍처의 설계 및 구현)

  • Seo, Young-Jun;Song, Young-Jae
    • The KIPS Transactions:PartD
    • /
    • v.15D no.2
    • /
    • pp.223-234
    • /
    • 2008
  • Recently the web service area provides the efficient integrated environment of the internal and external of corporation and enterprise that wants the introduction of it is increasing. Also the web service develops and the new business model appears, the domestic enterprise environment and e-business environment are changing caused by web service. The web service which provides the similar function increases, most the method which searches the suitable service in demand of the user is more considered seriously. When it needs to choose one among the similar web services, service consumer generally needs quality information of web service. The problem, however, is that the advertised QoS information of a web service is not always trustworthy. A service provider may publish inaccurate QoS information to attract more customers, or the published QoS information may be out of date. Allowing current customers to rate the QoS they receive from a web service, and making these ratings public, can provide new customers with valuable information on how to rank services. This paper suggests the agent-based quality broker architecture which helps to find a service providing the optimum quality that the consumer needs in a position of service consumer. It is able to solve problem which modify quality requirements of the consumer from providing the architecture it selects a web service to consumer dynamically. Namely, the consumer is able to search the service which provides the optimal quality criteria through UDDI browser which is connected in quality broker server. To quality criteria value decision of each service the user intervention is excluded the maximum. In the existing selection architecture, the objective evaluation was difficult in subjective class of service selecting of the consumer. But the proposal architecture is able to secure an objectivity with the quality criteria value decision where the agent monitors binding information in consumer location. Namely, it solves QoS information of service which provider does not provide with QoS information sharing which is caused by with feedback of consumer side agents.

A hybrid algorithm for the synthesis of computer-generated holograms

  • Nguyen The Anh;An Jun Won;Choe Jae Gwang;Kim Nam
    • Proceedings of the Optical Society of Korea Conference
    • /
    • 2003.07a
    • /
    • pp.60-61
    • /
    • 2003
  • A new approach to reduce the computation time of genetic algorithm (GA) for making binary phase holograms is described. Synthesized holograms having diffraction efficiency of 75.8% and uniformity of 5.8% are proven in computer simulation and experimentally demonstrated. Recently, computer-generated holograms (CGHs) having high diffraction efficiency and flexibility of design have been widely developed in many applications such as optical information processing, optical computing, optical interconnection, etc. Among proposed optimization methods, GA has become popular due to its capability of reaching nearly global. However, there exits a drawback to consider when we use the genetic algorithm. It is the large amount of computation time to construct desired holograms. One of the major reasons that the GA' s operation may be time intensive results from the expense of computing the cost function that must Fourier transform the parameters encoded on the hologram into the fitness value. In trying to remedy this drawback, Artificial Neural Network (ANN) has been put forward, allowing CGHs to be created easily and quickly (1), but the quality of reconstructed images is not high enough to use in applications of high preciseness. For that, we are in attempt to find a new approach of combiningthe good properties and performance of both the GA and ANN to make CGHs of high diffraction efficiency in a short time. The optimization of CGH using the genetic algorithm is merely a process of iteration, including selection, crossover, and mutation operators [2]. It is worth noting that the evaluation of the cost function with the aim of selecting better holograms plays an important role in the implementation of the GA. However, this evaluation process wastes much time for Fourier transforming the encoded parameters on the hologram into the value to be solved. Depending on the speed of computer, this process can even last up to ten minutes. It will be more effective if instead of merely generating random holograms in the initial process, a set of approximately desired holograms is employed. By doing so, the initial population will contain less trial holograms equivalent to the reduction of the computation time of GA's. Accordingly, a hybrid algorithm that utilizes a trained neural network to initiate the GA's procedure is proposed. Consequently, the initial population contains less random holograms and is compensated by approximately desired holograms. Figure 1 is the flowchart of the hybrid algorithm in comparison with the classical GA. The procedure of synthesizing a hologram on computer is divided into two steps. First the simulation of holograms based on ANN method [1] to acquire approximately desired holograms is carried. With a teaching data set of 9 characters obtained from the classical GA, the number of layer is 3, the number of hidden node is 100, learning rate is 0.3, and momentum is 0.5, the artificial neural network trained enables us to attain the approximately desired holograms, which are fairly good agreement with what we suggested in the theory. The second step, effect of several parameters on the operation of the hybrid algorithm is investigated. In principle, the operation of the hybrid algorithm and GA are the same except the modification of the initial step. Hence, the verified results in Ref [2] of the parameters such as the probability of crossover and mutation, the tournament size, and the crossover block size are remained unchanged, beside of the reduced population size. The reconstructed image of 76.4% diffraction efficiency and 5.4% uniformity is achieved when the population size is 30, the iteration number is 2000, the probability of crossover is 0.75, and the probability of mutation is 0.001. A comparison between the hybrid algorithm and GA in term of diffraction efficiency and computation time is also evaluated as shown in Fig. 2. With a 66.7% reduction in computation time and a 2% increase in diffraction efficiency compared to the GA method, the hybrid algorithm demonstrates its efficient performance. In the optical experiment, the phase holograms were displayed on a programmable phase modulator (model XGA). Figures 3 are pictures of diffracted patterns of the letter "0" from the holograms generated using the hybrid algorithm. Diffraction efficiency of 75.8% and uniformity of 5.8% are measured. We see that the simulation and experiment results are fairly good agreement with each other. In this paper, Genetic Algorithm and Neural Network have been successfully combined in designing CGHs. This method gives a significant reduction in computation time compared to the GA method while still allowing holograms of high diffraction efficiency and uniformity to be achieved. This work was supported by No.mOl-2001-000-00324-0 (2002)) from the Korea Science & Engineering Foundation.

  • PDF

Comparison of internal and marginal fit of crown according to milling order in a single machinable wax disc (단일 절삭가공용 왁스 디스크 내에서 순차적 절삭가공 순서에 따른 크라운의 내면 및 변연 적합도 비교)

  • Song, Jun-Beom;Lee, Jonghyuk;Ha, Seung-Ryong;Choi, Yu-Sung
    • The Journal of Korean Academy of Prosthodontics
    • /
    • v.59 no.4
    • /
    • pp.395-404
    • /
    • 2021
  • Purpose. The purpose of present study was to evaluate the effect of changing structural stability of wax disc on the fit of prosthesis when the milling proceeded in order. Materials and methods. Prepared maxillary left first molar was used to fabricate a Ni-Cr alloy reference model. This was scanned to design crown and then wax pattern was milled, invested and cast to fabricate prosthesis. The wax patterns located in a row centrally within a single wax disc were set into a total of five groups ranging from WM1 group that was first milled to WM5 group that was last milled and the number of each group was set as 10. Silicone replica technique was used to measure the marginal gap, axial internal gap, line angle internal gap, occlusal internal gap. Data was evaluated with one-way ANOVA with significance level set at α = .05 and then Tukey HSD test was conducted for post analysis. Results. Marginal gap measured in each group, it was 40.41 ± 2.15 ㎛ in WM1 group, 40.44 ± 2.23 ㎛ in WM2 group, 39.96 ± 2.25 ㎛ in WM3 group, 39.96 ± 2.48 ㎛ in WM4 group, and 40.57 ± 2.53 ㎛ in WM5 group. No significant difference was found between groups. The significant difference between the groups was also not found in the axial internal gap, line angle internal gap, and occlusal internal gap. Conclusion. Internal and marginal fit of single crown to the sequential order of milling processing in the single machinable wax disc did not seem to be affected by the sequence.

Utility Evaluation of Supportive Devices for Interventional Lower Extremity Angiography (인터벤션 하지 혈관조영검사를 위한 보조기구의 유용성 평가)

  • Kong, Chang gi;Song, Jong Nam;Jeong, Moon Taek;Han, Jae Bok
    • Journal of the Korean Society of Radiology
    • /
    • v.13 no.4
    • /
    • pp.613-621
    • /
    • 2019
  • The purpose of this study is to evaluate the effectiveness of supportive devices which are for minimizing the patient's movement during lower extremity angiography and to verify image quality of phantom by analyzing of Mask image, DSA image and Roadmap image into SNR and CNR. As a result of comparing SNR with CNR of mask image obtained by DSA technique using the phantom alone and phantom placed on the supportive devices, there was no significant difference between about 0~0.06 for SNR and about 0~0.003 for CNR. The study showed about 0.11~0.35 for SNR and 0.016~0.031 for CNR of DSA imaging by DSA technique about only water phantom of the blood vessel model and the water phantom placed on the device. Analyzing SNR and CNR of Roadmap technique about water phantom on the auxiliary device (hardboard paper, pomax, polycarbonate, acrylic) and water phantom alone, there was no significant difference between 0.02~0.05 for SNR and 0.002~0.004 for CNR. In conclusion, there was no significant difference on image quality by using supportive devices made by hardboard paper, pomax, polycarbonate or acryl regardless of whether using supportive devices or not. Supportive devices to minimize of the patient's movement may reduce the total amount of contrast, exam-time, radiation exposure and eliminate risk factors during angiogram. Supportive devices made by hardboard paper can be applied easily during angiogram due to advantages of reasonable price and simple processing. It is considered that will be useful to consider cost efficiency and types of materials and their properties in accordance with purpose and method of the study when the operator makes and uses supportive devices.

A Machine Learning-based Total Production Time Prediction Method for Customized-Manufacturing Companies (주문생산 기업을 위한 기계학습 기반 총생산시간 예측 기법)

  • Park, Do-Myung;Choi, HyungRim;Park, Byung-Kwon
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.1
    • /
    • pp.177-190
    • /
    • 2021
  • Due to the development of the fourth industrial revolution technology, efforts are being made to improve areas that humans cannot handle by utilizing artificial intelligence techniques such as machine learning. Although on-demand production companies also want to reduce corporate risks such as delays in delivery by predicting total production time for orders, they are having difficulty predicting this because the total production time is all different for each order. The Theory of Constraints (TOC) theory was developed to find the least efficient areas to increase order throughput and reduce order total cost, but failed to provide a forecast of total production time. Order production varies from order to order due to various customer needs, so the total production time of individual orders can be measured postmortem, but it is difficult to predict in advance. The total measured production time of existing orders is also different, which has limitations that cannot be used as standard time. As a result, experienced managers rely on persimmons rather than on the use of the system, while inexperienced managers use simple management indicators (e.g., 60 days total production time for raw materials, 90 days total production time for steel plates, etc.). Too fast work instructions based on imperfections or indicators cause congestion, which leads to productivity degradation, and too late leads to increased production costs or failure to meet delivery dates due to emergency processing. Failure to meet the deadline will result in compensation for delayed compensation or adversely affect business and collection sectors. In this study, to address these problems, an entity that operates an order production system seeks to find a machine learning model that estimates the total production time of new orders. It uses orders, production, and process performance for materials used for machine learning. We compared and analyzed OLS, GLM Gamma, Extra Trees, and Random Forest algorithms as the best algorithms for estimating total production time and present the results.

An Outlier Detection Using Autoencoder for Ocean Observation Data (해양 이상 자료 탐지를 위한 오토인코더 활용 기법 최적화 연구)

  • Kim, Hyeon-Jae;Kim, Dong-Hoon;Lim, Chaewook;Shin, Yongtak;Lee, Sang-Chul;Choi, Youngjin;Woo, Seung-Buhm
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.33 no.6
    • /
    • pp.265-274
    • /
    • 2021
  • Outlier detection research in ocean data has traditionally been performed using statistical and distance-based machine learning algorithms. Recently, AI-based methods have received a lot of attention and so-called supervised learning methods that require classification information for data are mainly used. This supervised learning method requires a lot of time and costs because classification information (label) must be manually designated for all data required for learning. In this study, an autoencoder based on unsupervised learning was applied as an outlier detection to overcome this problem. For the experiment, two experiments were designed: one is univariate learning, in which only SST data was used among the observation data of Deokjeok Island and the other is multivariate learning, in which SST, air temperature, wind direction, wind speed, air pressure, and humidity were used. Period of data is 25 years from 1996 to 2020, and a pre-processing considering the characteristics of ocean data was applied to the data. An outlier detection of actual SST data was tried with a learned univariate and multivariate autoencoder. We tried to detect outliers in real SST data using trained univariate and multivariate autoencoders. To compare model performance, various outlier detection methods were applied to synthetic data with artificially inserted errors. As a result of quantitatively evaluating the performance of these methods, the multivariate/univariate accuracy was about 96%/91%, respectively, indicating that the multivariate autoencoder had better outlier detection performance. Outlier detection using an unsupervised learning-based autoencoder is expected to be used in various ways in that it can reduce subjective classification errors and cost and time required for data labeling.

Analyzing Different Contexts for Energy Terms through Text Mining of Online Science News Articles (온라인 과학 기사 텍스트 마이닝을 통해 분석한 에너지 용어 사용의 맥락)

  • Oh, Chi Yeong;Kang, Nam-Hwa
    • Journal of Science Education
    • /
    • v.45 no.3
    • /
    • pp.292-303
    • /
    • 2021
  • This study identifies the terms frequently used together with energy in online science news articles and topics of the news reports to find out how the term energy is used in everyday life and to draw implications for science curriculum and instruction about energy. A total of 2,171 online news articles in science category published by 11 major newspaper companies in Korea for one year from March 1, 2018 were selected by using energy as a search term. As a result of natural language processing, a total of 51,224 sentences consisting of 507,901 words were compiled for analysis. Using the R program, term frequency analysis, semantic network analysis, and structural topic modeling were performed. The results show that the terms with exceptionally high frequencies were technology, research, and development, which reflected the characteristics of news articles that report new findings. On the other hand, terms used more than once per two articles were industry-related terms (industry, product, system, production, market) and terms that were sufficiently expected as energy-related terms such as 'electricity' and 'environment.' Meanwhile, 'sun', 'heat', 'temperature', and 'power generation', which are frequently used in energy-related science classes, also appeared as terms belonging to the highest frequency. From a network analysis, two clusters were found including terms related to industry and technology and terms related to basic science and research. From the analysis of terms paired with energy, it was also found that terms related to the use of energy such as 'energy efficiency,' 'energy saving,' and 'energy consumption' were the most frequently used. Out of 16 topics found, four contexts of energy were drawn including 'high-tech industry,' 'industry,' 'basic science,' and 'environment and health.' The results suggest that the introduction of the concept of energy degradation as a starting point for energy classes can be effective. It also shows the need to introduce high-tech industries or the context of environment and health into energy learning.