• Title/Summary/Keyword: Vector field clustering

Search Result 18, Processing Time 0.107 seconds

On 5-Axis Freeform Surface Machining Optimization: Vector Field Clustering Approach

  • My Chu A;Bohez Erik L J;Makhanov Stanlislav S;Munlin M;Phien Huynh N;Tabucanon Mario T
    • International Journal of CAD/CAM
    • /
    • v.5 no.1
    • /
    • pp.1-10
    • /
    • 2005
  • A new approach based on vector field clustering for tool path optimization of 5-axis CNC machining is presented in this paper. The strategy of the approach is to produce an efficient tool path with respect to the optimal cutting direction vector field. The optimal cutting direction maximizes the machining strip width. We use the normalized cut clustering technique to partition the vector field into clusters. The spiral and the zigzag patterns are then applied to generate tool path on the clusters. The iso-scallop method is used for calculating the tool path. Finally, our numerical examples and real cutting experiment show that the tool path generated by the proposed method is more efficient than the tool path generated by the traditional iso-parametric method.

Performance Comparison of Clustering Validity Indices with Business Applications (경영사례를 이용한 군집화 유효성 지수의 성능비교)

  • Lee, Soo-Hyun;Jeong, Youngseon;Kim, Jae-Yun
    • Journal of the Korean Operations Research and Management Science Society
    • /
    • v.41 no.2
    • /
    • pp.17-33
    • /
    • 2016
  • Clustering is one of the leading methods to analyze big data and is used in many different fields. This study deals with Clustering Validity Index (CVI) to verify the effectiveness of clustering results. We compare the performance of CVIs with business applications of various field. In this study, the used CVIs for comparing performance are DU, CH, DB, SVDU, SVCH, and SVDB. The first three CVIs are well-known ones in the existing research and the last three CVIs are based on support vector data description. It has been verified with outstanding performance and qualified as the application ability of CVIs based on support vector data description.

A Massively Parallel Algorithm for Fuzzy Vector Quantization (퍼지 벡터 양자화를 위한 대규모 병렬 알고리즘)

  • Huynh, Luong Van;Kim, Cheol-Hong;Kim, Jong-Myon
    • The KIPS Transactions:PartA
    • /
    • v.16A no.6
    • /
    • pp.411-418
    • /
    • 2009
  • Vector quantization algorithm based on fuzzy clustering has been widely used in the field of data compression since the use of fuzzy clustering analysis in the early stages of a vector quantization process can make this process less sensitive to its initialization. However, the process of fuzzy clustering is computationally very intensive because of its complex framework for the quantitative formulation of the uncertainty involved in the training vector space. To overcome the computational burden of the process, this paper introduces an array architecture for the implementation of fuzzy vector quantization (FVQ). The arrayarchitecture, which consists of 4,096 processing elements (PEs), provides a computationally efficient solution by employing an effective vector assignment strategy during the clustering process. Experimental results indicatethat the proposed parallel implementation providessignificantly greater performance and efficiency than appropriately scaled alternative array systems. In addition, the proposed parallel implementation provides 1000x greater performance and 100x higher energy efficiency than other implementations using today's ARMand TI DSP processors in the same 130nm technology. These results demonstrate that the proposed parallel implementation shows the potential for improved performance and energy efficiency.

Similarity Analysis of Hospitalization using Crowding Distance

  • Jung, Yong Gyu;Choi, Young Jin;Cha, Byeong Heon
    • International journal of advanced smart convergence
    • /
    • v.5 no.2
    • /
    • pp.53-58
    • /
    • 2016
  • With the growing use of big data and data mining, it serves to understand how such techniques can be used to understand various relationships in the healthcare field. This study uses hierarchical methods of data analysis to explore similarities in hospitalization across several New York state counties. The study utilized methods of measuring crowding distance of data for age-specific hospitalization period. Crowding distance is defined as the longest distance, or least similarity, between urban cities. It is expected that the city of Clinton have the greatest distance, while Albany the other cities are closer because they are connected by the shortest distance to each step. Similarities were stronger across hospital stays categorized by age. Hierarchical clustering can be applied to predict the similarity of data across the 10 cities of hospitalization with the measurement of crowding distance. In order to enhance the performance of hierarchical clustering, comparison can be made across congestion distance when crowding distance is applied first through the application of converting text to an attribute vector. Measurements of similarity between two objects are dependent on the measurement method used in clustering but is distinguished from the similarity of the distance; where the smaller the distance value the more similar two things are to one other. By applying this specific technique, it is found that the distance between crowding is reduced consistently in relationship to similarity between the data increases to enhance the performance of the experiments through the application of special techniques. Furthermore, through the similarity by city hospitalization period, when the construction of hospital wards in cities, by referring to results of experiments, or predict possible will land to the extent of the size of the hospital facilities hospital stay is expected to be useful in efficiently managing the patient in a similar area.

Empirical Comparison of Word Similarity Measures Based on Co-Occurrence, Context, and a Vector Space Model

  • Kadowaki, Natsuki;Kishida, Kazuaki
    • Journal of Information Science Theory and Practice
    • /
    • v.8 no.2
    • /
    • pp.6-17
    • /
    • 2020
  • Word similarity is often measured to enhance system performance in the information retrieval field and other related areas. This paper reports on an experimental comparison of values for word similarity measures that were computed based on 50 intentionally selected words from a Reuters corpus. There were three targets, including (1) co-occurrence-based similarity measures (for which a co-occurrence frequency is counted as the number of documents or sentences), (2) context-based distributional similarity measures obtained from a latent Dirichlet allocation (LDA), nonnegative matrix factorization (NMF), and Word2Vec algorithm, and (3) similarity measures computed from the tf-idf weights of each word according to a vector space model (VSM). Here, a Pearson correlation coefficient for a pair of VSM-based similarity measures and co-occurrence-based similarity measures according to the number of documents was highest. Group-average agglomerative hierarchical clustering was also applied to similarity matrices computed by individual measures. An evaluation of the cluster sets according to an answer set revealed that VSM- and LDA-based similarity measures performed best.

Seabed Sediment Classification Algorithm using Continuous Wavelet Transform

  • Lee, Kibae;Bae, Jinho;Lee, Chong Hyun;Kim, Juho;Lee, Jaeil;Cho, Jung Hong
    • Journal of Advanced Research in Ocean Engineering
    • /
    • v.2 no.4
    • /
    • pp.202-208
    • /
    • 2016
  • In this paper, we propose novel seabed sediment classification algorithm using feature obtained by continuous wavelet transform (CWT). Contrast to previous researches using direct reflection coefficient of seabed which is function of frequency and is highly influenced by sediment types, we develop an algorithm using both direct reflection signal and backscattering signal. In order to obtain feature vector, we employ CWT of the signal and obtain histograms extracted from local binary patterns of the scalogram. The proposed algorithm also adopts principal component analysis (PCA) to reduce dimension of the feature vector so that it requires low computational cost to classify seabed sediment. For training and classification, we adopts K-means clustering algorithm which can be done with low computational cost and does not require prior information of the sediment. To verify the proposed algorithm, we obtain field data measured at near Jeju island and show that the proposed classification algorithm has reliable discrimination performance by comparing the classification results with actual physical properties of the sediments.

Genetic Algorithm for Image Feature Selection (영상 특징 선택을 위한 유전 알고리즘)

  • Shin Youns-Geun;Park Sang-Sung;Jang Dong-Sik
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2006.06b
    • /
    • pp.193-195
    • /
    • 2006
  • As multimedia information increases sharply, In image retrieval field the method that can analyze image data quickly and exactly is required. In the case of image data, because each data includes a lot of informations, between accuracy and speed of retrieval become trade-off. To solve these problem, feature vector extracting process that use Genetic Algorithm for implementing prompt and correct image clustering system in case of retrieval of mass image data is proposed. After extracting color and texture features, the representative feature vector among these features is extracted by using Genetic Algorithm.

  • PDF

Modified multi-sense skip-gram using weighted context and x-means (가중 문맥벡터와 X-means 방법을 이용한 변형 다의어스킵그램)

  • Jeong, Hyunwoo;Lee, Eun Ryung
    • The Korean Journal of Applied Statistics
    • /
    • v.34 no.3
    • /
    • pp.389-399
    • /
    • 2021
  • In recent years, word embedding has been a popular field of natural language processing research and a skip-gram has become one successful word embedding method. It assigns a word embedding vector to each word using contexts, which provides an effective way to analyze text data. However, due to the limitation of vector space model, primary word embedding methods assume that every word only have a single meaning. As one faces multi-sense words, that is, words with more than one meaning, in reality, Neelakantan (2014) proposed a multi-sense skip-gram (MSSG) to find embedding vectors corresponding to the each senses of a multi-sense word using a clustering method. In this paper, we propose a modified method of the MSSG to improve statistical accuracy. Moreover, we propose a data-adaptive choice of the number of clusters, that is, the number of meanings for a multi-sense word. Some numerical evidence is given by conducting real data-based simulations.

User modeling based on fuzzy category and interest for web usage mining

  • Lee, Si-Hun;Lee, Jee-Hyong
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.5 no.1
    • /
    • pp.88-93
    • /
    • 2005
  • Web usage mining is a research field for searching potentially useful and valuable information from web log file. Web log file is a simple list of pages that users refer. Therefore, it is not easy to analyze user's current interest field from web log file. This paper presents web usage mining method for finding users' current interest based on fuzzy categories. We consider not only how many times a user visits pages but also when he visits. We describe a user's current interest with a fuzzy interest degree to categories. Based on fuzzy categories and fuzzy interest degrees, we also propose a method to cluster users according to their interests for user modeling. For user clustering, we define a category vector space. Experiments show that our method properly reflects the time factor of users' web visiting as well as the users' visit number.

Boundary Detection using Adaptive Bayesian Approach to Image Segmentation (적응적 베이즈 영상분할을 이용한 경계추출)

  • Kim Kee Tae;Choi Yoon Su;Kim Gi Hong
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.22 no.3
    • /
    • pp.303-309
    • /
    • 2004
  • In this paper, an adaptive Bayesian approach to image segmentation was developed for boundary detection. Both image intensities and texture information were used for obtaining better quality of the image segmentation by using the C programming language. Fuzzy c-mean clustering was applied fer the conditional probability density function, and Gibbs random field model was used for the prior probability density function. To simply test the algorithm, a synthetic image (256$\times$256) with a set of low gray values (50, 100, 150 and 200) was created and normalized between 0 and 1 n double precision. Results have been presented that demonstrate the effectiveness of the algorithm in segmenting the synthetic image, resulting in more than 99% accuracy when noise characteristics are correctly modeled. The algorithm was applied to the Antarctic mosaic that was generated using 1963 Declassified Intelligence Satellite Photographs. The accuracy of the resulting vector map was estimated about 300-m.