• Title/Summary/Keyword: I-벡터

Search Result 380, Processing Time 0.023 seconds

Elimination of Redundant Input Information and Parameters during Neural Network Training (신경망 학습 과정중 불필요한 입력 정보 및 파라미터들의 제거)

  • Won, Yong-Gwan;Park, Gwang-Gyu
    • The Transactions of the Korea Information Processing Society
    • /
    • v.3 no.3
    • /
    • pp.439-448
    • /
    • 1996
  • Extraction and selection of the informative features play a central role in pattern recognition. This paper describes a modified back-propagation algorithm that performs selection of the informative features and trains a neural network simultaneously. The algorithm is mainly composed of three repetitive steps : training, connection pruning, and input unit elimination. Afer initial training, the connections that have small magnitude are first pruned. Any unit that has a small number of connections to the hidden units is deleted,which is equivalent to excluding the feature corresponding to that unit.If the error increases,the network is retraned,again followed by connection pruning and input unit elimination.As a result,the algorithm selects the most im-portant features in the measurement space without a transformation to another space.Also,the selected features are the most-informative ones for the classification,because feature selection is tightly coupled with the classifi-cation performance.This algorithm helps avoid measurement of redundant or less informative features,which may be expensive.Furthermore,the final network does not include redundant parameters,i.e.,weights and biases,that may cause degradation of classification performance.In applications,the algorithm preserves the most informative features and significantly reduces the dimension of the feature vectors whiout performance degradation.

  • PDF

Hand Motion Recognition Algorithm Using Skin Color and Center of Gravity Profile (피부색과 무게중심 프로필을 이용한 손동작 인식 알고리즘)

  • Park, Youngmin
    • The Journal of the Convergence on Culture Technology
    • /
    • v.7 no.2
    • /
    • pp.411-417
    • /
    • 2021
  • The field that studies human-computer interaction is called HCI (Human-computer interaction). This field is an academic field that studies how humans and computers communicate with each other and recognize information. This study is a study on hand gesture recognition for human interaction. This study examines the problems of existing recognition methods and proposes an algorithm to improve the recognition rate. The hand region is extracted based on skin color information for the image containing the shape of the human hand, and the center of gravity profile is calculated using principal component analysis. I proposed a method to increase the recognition rate of hand gestures by comparing the obtained information with predefined shapes. We proposed a method to increase the recognition rate of hand gestures by comparing the obtained information with predefined shapes. The existing center of gravity profile has shown the result of incorrect hand gesture recognition for the deformation of the hand due to rotation, but in this study, the center of gravity profile is used and the point where the distance between the points of all contours and the center of gravity is the longest is the starting point. Thus, a robust algorithm was proposed by re-improving the center of gravity profile. No gloves or special markers attached to the sensor are used for hand gesture recognition, and a separate blue screen is not installed. For this result, find the feature vector at the nearest distance to solve the misrecognition, and obtain an appropriate threshold to distinguish between success and failure.

Improved Estimation of Hourly Surface Ozone Concentrations using Stacking Ensemble-based Spatial Interpolation (스태킹 앙상블 모델을 이용한 시간별 지상 오존 공간내삽 정확도 향상)

  • KIM, Ye-Jin;KANG, Eun-Jin;CHO, Dong-Jin;LEE, Si-Woo;IM, Jung-Ho
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.25 no.3
    • /
    • pp.74-99
    • /
    • 2022
  • Surface ozone is produced by photochemical reactions of nitrogen oxides(NOx) and volatile organic compounds(VOCs) emitted from vehicles and industrial sites, adversely affecting vegetation and the human body. In South Korea, ozone is monitored in real-time at stations(i.e., point measurements), but it is difficult to monitor and analyze its continuous spatial distribution. In this study, surface ozone concentrations were interpolated to have a spatial resolution of 1.5km every hour using the stacking ensemble technique, followed by a 5-fold cross-validation. Base models for the stacking ensemble were cokriging, multi-linear regression(MLR), random forest(RF), and support vector regression(SVR), while MLR was used as the meta model, having all base model results as additional input variables. The results showed that the stacking ensemble model yielded the better performance than the individual base models, resulting in an averaged R of 0.76 and RMSE of 0.0065ppm during the study period of 2020. The surface ozone concentration distribution generated by the stacking ensemble model had a wider range with a spatial pattern similar with terrain and urbanization variables, compared to those by the base models. Not only should the proposed model be capable of producing the hourly spatial distribution of ozone, but it should also be highly applicable for calculating the daily maximum 8-hour ozone concentrations.

Cloning and Transcription Analysis of Sporulation Gene (spo5) in Schizosaccharomyces pombe (Schizosaccharomyces bombe 포자형성 유전자(spo5)의 Cloning 및 전사조절)

  • 김동주
    • The Korean Journal of Food And Nutrition
    • /
    • v.15 no.2
    • /
    • pp.112-118
    • /
    • 2002
  • Sporulation in the fission yeast Schizosaccharomyces pombe has been regarded as an important model of cellular development and differentiation. S. pombe cells proliferate by mitosis and binary fission on growth medium. Deprivation of nutrients especially nitrogen sources, causes the cessation of mitosis and initiates sexual reproduction by matting between two sexually compatible cell types. Meiosis is then followed in a diploid cell in the absence of nitrogen source. DNA fragment complemented with the mutations of sporulation gene was isolated from the S. pombe gene library constructed in the vector, pDB 248' and designated as pDB(spo5)1. We futher analyzed six recombinant plasmids, pDB(spo5)2, pDB(spo5)3, pDB(spo5)4, pDB(spo5)5, pDB (spo5)6, pDB(spo5)7 and found each of these plasmids is able to rescue the spo5-2, spo5-3, spo5-4, spo5-5, spo5-6, spo5-7 mutations, respectively. Mapping of the integrated plasmid into the homologous site of the S. pombe chromosomes demonstrated that pDB(spo5)1, and pDB(spu5)Rl contained the spo5 gene. Transcripts of spo5 gene were analyzed by Northern hybridization. Two transcripts of 3.2 kb and 2.5kb were detected with 5kb Hind Ⅲ fragment containing a part of the spo5 gene as a probe. The small mRNA(2.5kb) appeared only when a wild-type strain was cultured in the absence of nitrogen source in which condition the large mRNA (3.2kb) was produced constitutively. Appearance of a 2.5kb spo5-mRNA depends upon the function of the meil, mei2 and mei3 genes.

A Study on Basic Plan for Upscaling Environmental Conservation Value Assessment Map(ECVAM) of National Land in South Korea (대축척 국토환경성평가지도 작성방안 연구)

  • Lee, Moung-Jin;Jeon, Seong-Woo;Lee, Chong-Soo;Kang, Byung-Jin;Song, Won-Kyong
    • Journal of Environmental Policy
    • /
    • v.6 no.3
    • /
    • pp.115-145
    • /
    • 2007
  • This study was performed for developing upscaling Environmental Conservation Value Assessment Map(ECVAM) of National Land in South Korea and presenting the application method of ECVAM. This ECVAM adopted the least indicator method and uses a Geographic Information System(GIS). This map is made through evaluation of 67 items. As a result, the construction of ECVAM was defined as a process of identifying land use to scientifically assess the physical and environmental value of land and classify conservation value into several grades for the sustainable management of environmental resources. After applying ECVAM criteria of five degrees to the whole of study area, Grade I, showing the highest conservation value, accounted for 29.3% by land area of ECVAM. Grades II, III, IV and V likewise accounted for, respectively, 21.7%, 17.2%, 7.1% and the lowest conservation value of 24.7%. other result, ECVAM and land suitability assessment agreement rate is Grade I 33.05%, Grades II, III, IV and V likewise accounted for 12.92%, 15.05%, 36.93% and last value of 53.28% This study set up "the realization of the improvement ECVAM" as the vision of the advancing strategy. In order to accomplish the vision, this study established the purpose as follow; constructing strategic assessment value relation to ECVAM based on knowledge, arranging the foundation to upscaling assessment value And this study devised preparatory plans to achieve the vision and the purpose as next; construction on base theme map by 1:5,000 scalie, base on land register theme map and precision land cover map. Therefore, for applying the result of this study to the upscaling Environmental Conservation Value Assessment Map(ECVAM), it considers regularly the systematic categorization of preceding item, consideration issue of national environmental geographic information using the ECVAM.

  • PDF

Incorporating Social Relationship discovered from User's Behavior into Collaborative Filtering (사용자 행동 기반의 사회적 관계를 결합한 사용자 협업적 여과 방법)

  • Thay, Setha;Ha, Inay;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.1-20
    • /
    • 2013
  • Nowadays, social network is a huge communication platform for providing people to connect with one another and to bring users together to share common interests, experiences, and their daily activities. Users spend hours per day in maintaining personal information and interacting with other people via posting, commenting, messaging, games, social events, and applications. Due to the growth of user's distributed information in social network, there is a great potential to utilize the social data to enhance the quality of recommender system. There are some researches focusing on social network analysis that investigate how social network can be used in recommendation domain. Among these researches, we are interested in taking advantages of the interaction between a user and others in social network that can be determined and known as social relationship. Furthermore, mostly user's decisions before purchasing some products depend on suggestion of people who have either the same preferences or closer relationship. For this reason, we believe that user's relationship in social network can provide an effective way to increase the quality in prediction user's interests of recommender system. Therefore, social relationship between users encountered from social network is a common factor to improve the way of predicting user's preferences in the conventional approach. Recommender system is dramatically increasing in popularity and currently being used by many e-commerce sites such as Amazon.com, Last.fm, eBay.com, etc. Collaborative filtering (CF) method is one of the essential and powerful techniques in recommender system for suggesting the appropriate items to user by learning user's preferences. CF method focuses on user data and generates automatic prediction about user's interests by gathering information from users who share similar background and preferences. Specifically, the intension of CF method is to find users who have similar preferences and to suggest target user items that were mostly preferred by those nearest neighbor users. There are two basic units that need to be considered by CF method, the user and the item. Each user needs to provide his rating value on items i.e. movies, products, books, etc to indicate their interests on those items. In addition, CF uses the user-rating matrix to find a group of users who have similar rating with target user. Then, it predicts unknown rating value for items that target user has not rated. Currently, CF has been successfully implemented in both information filtering and e-commerce applications. However, it remains some important challenges such as cold start, data sparsity, and scalability reflected on quality and accuracy of prediction. In order to overcome these challenges, many researchers have proposed various kinds of CF method such as hybrid CF, trust-based CF, social network-based CF, etc. In the purpose of improving the recommendation performance and prediction accuracy of standard CF, in this paper we propose a method which integrates traditional CF technique with social relationship between users discovered from user's behavior in social network i.e. Facebook. We identify user's relationship from behavior of user such as posts and comments interacted with friends in Facebook. We believe that social relationship implicitly inferred from user's behavior can be likely applied to compensate the limitation of conventional approach. Therefore, we extract posts and comments of each user by using Facebook Graph API and calculate feature score among each term to obtain feature vector for computing similarity of user. Then, we combine the result with similarity value computed using traditional CF technique. Finally, our system provides a list of recommended items according to neighbor users who have the biggest total similarity value to the target user. In order to verify and evaluate our proposed method we have performed an experiment on data collected from our Movies Rating System. Prediction accuracy evaluation is conducted to demonstrate how much our algorithm gives the correctness of recommendation to user in terms of MAE. Then, the evaluation of performance is made to show the effectiveness of our method in terms of precision, recall, and F1-measure. Evaluation on coverage is also included in our experiment to see the ability of generating recommendation. The experimental results show that our proposed method outperform and more accurate in suggesting items to users with better performance. The effectiveness of user's behavior in social network particularly shows the significant improvement by up to 6% on recommendation accuracy. Moreover, experiment of recommendation performance shows that incorporating social relationship observed from user's behavior into CF is beneficial and useful to generate recommendation with 7% improvement of performance compared with benchmark methods. Finally, we confirm that interaction between users in social network is able to enhance the accuracy and give better recommendation in conventional approach.

Index-based Searching on Timestamped Event Sequences (타임스탬프를 갖는 이벤트 시퀀스의 인덱스 기반 검색)

  • 박상현;원정임;윤지희;김상욱
    • Journal of KIISE:Databases
    • /
    • v.31 no.5
    • /
    • pp.468-478
    • /
    • 2004
  • It is essential in various application areas of data mining and bioinformatics to effectively retrieve the occurrences of interesting patterns from sequence databases. For example, let's consider a network event management system that records the types and timestamp values of events occurred in a specific network component(ex. router). The typical query to find out the temporal casual relationships among the network events is as fellows: 'Find all occurrences of CiscoDCDLinkUp that are fellowed by MLMStatusUP that are subsequently followed by TCPConnectionClose, under the constraint that the interval between the first two events is not larger than 20 seconds, and the interval between the first and third events is not larger than 40 secondsTCPConnectionClose. This paper proposes an indexing method that enables to efficiently answer such a query. Unlike the previous methods that rely on inefficient sequential scan methods or data structures not easily supported by DBMSs, the proposed method uses a multi-dimensional spatial index, which is proven to be efficient both in storage and search, to find the answers quickly without false dismissals. Given a sliding window W, the input to a multi-dimensional spatial index is a n-dimensional vector whose i-th element is the interval between the first event of W and the first occurrence of the event type Ei in W. Here, n is the number of event types that can be occurred in the system of interest. The problem of‘dimensionality curse’may happen when n is large. Therefore, we use the dimension selection or event type grouping to avoid this problem. The experimental results reveal that our proposed technique can be a few orders of magnitude faster than the sequential scan and ISO-Depth index methods.hods.

The Analysis of Predictive Factors for the Identification of Patients Who Could Benefit from Respiratory-Gated Radiotherapy in Non-Small Cell Lung Cancer (비소세포성 폐암에서 호흡동기방사선치료 적용 환자군의 선택을 위한 예측인자들의 분석)

  • Jang, Seong-Soon;Park, Ji-Chan
    • Radiation Oncology Journal
    • /
    • v.27 no.4
    • /
    • pp.228-239
    • /
    • 2009
  • Purpose: 4DCT scans performed for radiotherapy were retrospectively analyzed to assess the possible benefits of respiratory gating in non-small cell lung cancer (NSCLC) and established the predictive factors for identifying patients who could benefit from this approach. Materials and Methods: Three treatment planning was performed for 15 patients with stage I~III NSCLC using different planning target volumes (PTVs) as follows: 1) PTVroutine, derived from the addition of conventional uniform margins to gross tumor volume (GTV) of a single bin, 2) PTVall phases (patient-specific PTV), derived from the composite GTV of all 6 bins of the 4DCT, and 3) PTVgating, derived from the composite GTV of 3 consecutive bins at end-exhalation. Results: The reductions in PTV were 43.2% and 9.5%, respectively, for the PTVall phases vs. PTVroutine and PTVgating vs. PTVall phases. Compared to PTVroutine, the use of PTVall phases and PTVgating reduced the mean lung dose (MLD) by 18.1% and 21.6%, and $V_{20}$ by 18.2% and 22.0%, respectively. Significant correlations were seen between certain predictive factors selected from the tumor mobility and volume analysis, such as the 3D mobility vector, the reduction in 3D mobility and PTV with gating, and the ratio of GTV overlap between 2 extreme bins and additional reductions in both MLD and $V_{20}$ with gating. Conclusion: The additional benefits with gating compared to the use of patient-specific PTV were modest; however, there were distinct correlations and differences according to the predictive factors. Therefore, these predictive factors might be useful for identifying patients who could benefit from respiratory-gated radiotherapy.

Identifying sources of heavy metal contamination in stream sediments using machine learning classifiers (기계학습 분류모델을 이용한 하천퇴적물의 중금속 오염원 식별)

  • Min Jeong Ban;Sangwook Shin;Dong Hoon Lee;Jeong-Gyu Kim;Hosik Lee;Young Kim;Jeong-Hun Park;ShunHwa Lee;Seon-Young Kim;Joo-Hyon Kang
    • Journal of Wetlands Research
    • /
    • v.25 no.4
    • /
    • pp.306-314
    • /
    • 2023
  • Stream sediments are an important component of water quality management because they are receptors of various pollutants such as heavy metals and organic matters emitted from upland sources and can be secondary pollution sources, adversely affecting water environment. To effectively manage the stream sediments, identification of primary sources of sediment contamination and source-associated control strategies will be required. We evaluated the performance of machine learning models in identifying primary sources of sediment contamination based on the physico-chemical properties of stream sediments. A total of 356 stream sediment data sets of 18 quality parameters including 10 heavy metal species(Cd, Cu, Pb, Ni, As, Zn, Cr, Hg, Li, and Al), 3 soil parameters(clay, silt, and sand fractions), and 5 water quality parameters(water content, loss on ignition, total organic carbon, total nitrogen, and total phosphorous) were collected near abandoned metal mines and industrial complexes across the four major river basins in Korea. Two machine learning algorithms, linear discriminant analysis (LDA) and support vector machine (SVM) classifiers were used to classify the sediments into four cases of different combinations of the sampling period and locations (i.e., mine in dry season, mine in wet season, industrial complex in dry season, and industrial complex in wet season). Both models showed good performance in the classification, with SVM outperformed LDA; the accuracy values of LDA and SVM were 79.5% and 88.1%, respectively. An SVM ensemble model was used for multi-label classification of the multiple contamination sources inlcuding landuses in the upland areas within 1 km radius from the sampling sites. The results showed that the multi-label classifier was comparable performance with sinlgle-label SVM in classifying mines and industrial complexes, but was less accurate in classifying dominant land uses (50~60%). The poor performance of the multi-label SVM is likely due to the overfitting caused by small data sets compared to the complexity of the model. A larger data set might increase the performance of the machine learning models in identifying contamination sources.

An Intelligence Support System Research on KTX Rolling Stock Failure Using Case-based Reasoning and Text Mining (사례기반추론과 텍스트마이닝 기법을 활용한 KTX 차량고장 지능형 조치지원시스템 연구)

  • Lee, Hyung Il;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.1
    • /
    • pp.47-73
    • /
    • 2020
  • KTX rolling stocks are a system consisting of several machines, electrical devices, and components. The maintenance of the rolling stocks requires considerable expertise and experience of maintenance workers. In the event of a rolling stock failure, the knowledge and experience of the maintainer will result in a difference in the quality of the time and work to solve the problem. So, the resulting availability of the vehicle will vary. Although problem solving is generally based on fault manuals, experienced and skilled professionals can quickly diagnose and take actions by applying personal know-how. Since this knowledge exists in a tacit form, it is difficult to pass it on completely to a successor, and there have been studies that have developed a case-based rolling stock expert system to turn it into a data-driven one. Nonetheless, research on the most commonly used KTX rolling stock on the main-line or the development of a system that extracts text meanings and searches for similar cases is still lacking. Therefore, this study proposes an intelligence supporting system that provides an action guide for emerging failures by using the know-how of these rolling stocks maintenance experts as an example of problem solving. For this purpose, the case base was constructed by collecting the rolling stocks failure data generated from 2015 to 2017, and the integrated dictionary was constructed separately through the case base to include the essential terminology and failure codes in consideration of the specialty of the railway rolling stock sector. Based on a deployed case base, a new failure was retrieved from past cases and the top three most similar failure cases were extracted to propose the actual actions of these cases as a diagnostic guide. In this study, various dimensionality reduction measures were applied to calculate similarity by taking into account the meaningful relationship of failure details in order to compensate for the limitations of the method of searching cases by keyword matching in rolling stock failure expert system studies using case-based reasoning in the precedent case-based expert system studies, and their usefulness was verified through experiments. Among the various dimensionality reduction techniques, similar cases were retrieved by applying three algorithms: Non-negative Matrix Factorization(NMF), Latent Semantic Analysis(LSA), and Doc2Vec to extract the characteristics of the failure and measure the cosine distance between the vectors. The precision, recall, and F-measure methods were used to assess the performance of the proposed actions. To compare the performance of dimensionality reduction techniques, the analysis of variance confirmed that the performance differences of the five algorithms were statistically significant, with a comparison between the algorithm that randomly extracts failure cases with identical failure codes and the algorithm that applies cosine similarity directly based on words. In addition, optimal techniques were derived for practical application by verifying differences in performance depending on the number of dimensions for dimensionality reduction. The analysis showed that the performance of the cosine similarity was higher than that of the dimension using Non-negative Matrix Factorization(NMF) and Latent Semantic Analysis(LSA) and the performance of algorithm using Doc2Vec was the highest. Furthermore, in terms of dimensionality reduction techniques, the larger the number of dimensions at the appropriate level, the better the performance was found. Through this study, we confirmed the usefulness of effective methods of extracting characteristics of data and converting unstructured data when applying case-based reasoning based on which most of the attributes are texted in the special field of KTX rolling stock. Text mining is a trend where studies are being conducted for use in many areas, but studies using such text data are still lacking in an environment where there are a number of specialized terms and limited access to data, such as the one we want to use in this study. In this regard, it is significant that the study first presented an intelligent diagnostic system that suggested action by searching for a case by applying text mining techniques to extract the characteristics of the failure to complement keyword-based case searches. It is expected that this will provide implications as basic study for developing diagnostic systems that can be used immediately on the site.