• Title/Summary/Keyword: O-Algorithm

Search Result 1,529, Processing Time 0.029 seconds

A Bottleneck Search Algorithm for Digraph Using Maximum Adjacency Merging Method (최대 인접 병합 방법을 적용한 방향 그래프의 병목지점 탐색 알고리즘)

  • Lee, Sang-Un
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.12 no.5
    • /
    • pp.129-139
    • /
    • 2012
  • Given digraph network $D=(N,A),n{\in}N,a=c(u,v){\in}A$ with source s and sink t, the maximum flow from s to t is determined by cut (S, T) that splits N to $s{\in}S$ and $t{\in}T$ disjoint sets with minimum cut value. The Ford-Fulkerson (F-F) algorithm with time complexity $O(NA^2)$ has been well known to this problem. The F-F algorithm finds all possible augmenting paths from s to t with residual capacity arcs and determines bottleneck arc that has a minimum residual capacity among the paths. After completion of algorithm, you should be determine the minimum cut by combination of bottleneck arcs. This paper suggests maximum adjacency merging and compute cut value method is called by MA-merging algorithm. We start the initial value to S={s}, T={t}, Then we select the maximum capacity $_{max}c(u,v)$ in the graph and merge to adjacent set S or T. Finally, we compute cut value of S or T. This algorithm runs n-1 times. We experiment Ford-Fulkerson and MA-merging algorithm for various 8 digraph. As a results, MA-merging algorithm can be finds minimum cut during the n-1 running times with time complexity O(N).

An Algorithm of SBIBD based Load Balancing Applicable to a Random Network (랜덤 네트워크에 적용 가능한 SBIBD기반의 부하 균형 알고리즘)

  • Lee, Ok-Bin;Lee, Yeo-Jin;Choi, Dong-Min;Chung, Il-Yong
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.3
    • /
    • pp.331-340
    • /
    • 2008
  • In order to make load balancing, workload information of nodes should be informed to the network. In a load balancing algorithm[13] based on the SBIBD(Symmetric Balanced Incomplete Block Design), each node receives global workload information by only two round message exchange with O (${\upsilon}{\sqrt{\upsilon}}$) traffic overhead, where ${\upsilon}$ is the number of nodes. It is very efficient but works well only when ${\upsilon}=p^2+p+1$ for a prime number p. In this paper, we generate a special incidence structure in order for the algorithm works well for an arbitrary number of nodes. In the experiment with w,($5{\leq}w{\leq}5,000$), nodes and more than 80% of receiving workload information, traffic overhead was less than O ($w{\sqrt{w}})$ and the result for standard deviation of traffic overhead showed that each node has largely balanced amount of traffic overhead.

  • PDF

Sintering process optimization of ZnO varistor materials by machine learning based metamodel (기계학습 기반의 메타모델을 활용한 ZnO 바리스터 소결 공정 최적화 연구)

  • Kim, Boyeol;Seo, Ga Won;Ha, Manjin;Hong, Youn-Woo;Chung, Chan-Yeup
    • Journal of the Korean Crystal Growth and Crystal Technology
    • /
    • v.31 no.6
    • /
    • pp.258-263
    • /
    • 2021
  • ZnO varistor is a semiconductor device which can serve to protect the circuit from surge voltage because its non-linear I-V characteristics by controlling the microstructure of grain and grain boundaries. In order to obtain desired electrical properties, it is important to control microstructure evolution during the sintering process. In this research, we defined a dataset composed of process conditions of sintering and relative permittivity of sintered body, and collected experimental dataset with DOE. Meta-models can predict permittivity were developed by learning the collected experimental dataset on various machine learning algorithms. By utilizing the meta-model, we can derive optimized sintering conditions that could show the maximum permittivity from the numerical-based HMA (Hybrid Metaheuristic Algorithm) optimization algorithm. It is possible to search the optimal process conditions with minimum number of experiments if meta-model-based optimization is applied to ceramic processing.

(Methods for Improving ENHHO of CDMA Mobile Communication) (CDMA 이동통신 ENHHO의 성능개선 방안)

  • 최동유;김재섭;이강현;박창균
    • Journal of the Institute of Electronics Engineers of Korea TE
    • /
    • v.39 no.1
    • /
    • pp.55-64
    • /
    • 2002
  • One of the most important things is Hand-0ff(H/O) in mobile phone service. Meantime soft H/O and softer H/O have been used between the same generations(2G$\longleftrightarrow$2G) and there is little problem. With user's needs and the development of communication technology, the system with different generation coexists and pilot beacon and time-periodic beacon are used for H/O between generations(2.5G$\longleftrightarrow$G) to start service, but it is not economic and efficient. To improve such problems, ENHHO developed in April 2001 has recently used. Unfortunately, this method also has the defect of momentary cutoff of communication information during the procedure of H/O in case of mote than 5 pilot signals. Accordingly, we suggest algorithm to improve the problem of ENHHO and examines its propriety by analyzing the results of field survey using the algorithm.

Impossible Differential Cryptanalysis on DVB-CSA

  • Zhang, Kai;Guan, Jie;Hu, Bin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.4
    • /
    • pp.1944-1956
    • /
    • 2016
  • The Digital Video Broadcasting-Common Scrambling Algorithm is an ETSI-designated algorithm designed for protecting MPEG-2 signal streams, and it is universally used. Its structure is a typical hybrid symmetric cipher which contains stream part and block part within a symmetric cipher, although the entropy is 64 bits, there haven't any effective cryptanalytic results up to now. This paper studies the security level of CSA against impossible differential cryptanalysis, a 20-round impossible differential for the block cipher part is proposed and a flaw in the cipher structure is revealed. When we attack the block cipher part alone, to recover 16 bits of the initial key, the data complexity of the attack is O(244.5), computational complexity is O(222.7) and memory complexity is O(210.5) when we attack CSA-BC reduced to 21 rounds. According to the structure flaw, an attack on CSA with block cipher part reduced to 21 rounds is proposed, the computational complexity is O(221.7), data complexity is O(243.5) and memory complexity is O(210.5), we can recover 8 bits of the key accordingly. Taking both the block cipher part and stream cipher part of CSA into consideration, it is currently the best result on CSA which is accessible as far as we know.

Design of an Expandable VLSI Rebound Sorter (확장형 VLSI 리바운드 정렬기의 설계)

  • Yun, Ji-Heon;Ahn, Byoung-Chul
    • The Transactions of the Korea Information Processing Society
    • /
    • v.2 no.3
    • /
    • pp.433-442
    • /
    • 1995
  • This paper presents an improved VLSI implementation of a parallel sorter to achieve O(Ν) time complexity. Many fast VLSI sort algorithms have been proposed for sorting N elements in O(log Ν) time. However, most such algorithms proposed have complex network structure without considering data input and output time. They are also very difficult to expand or to use in real applications. After analyzing the chip area and time complexity of several parallel sort algorithms with overlapping data input and output time, the most effective algorithm, the rebound sort algorithm, is implemented in VLSI with some improvements. To achieve O(Ν) time complexity, an improved rebound sorter is able to sort 8 16-bits records on a chip. And it is possible to sort more than 8 records by connecting chips in a chain vertically.

  • PDF

Text Summarization on Large-scale Vietnamese Datasets

  • Ti-Hon, Nguyen;Thanh-Nghi, Do
    • Journal of information and communication convergence engineering
    • /
    • v.20 no.4
    • /
    • pp.309-316
    • /
    • 2022
  • This investigation is aimed at automatic text summarization on large-scale Vietnamese datasets. Vietnamese articles were collected from newspaper websites and plain text was extracted to build the dataset, that included 1,101,101 documents. Next, a new single-document extractive text summarization model was proposed to evaluate this dataset. In this summary model, the k-means algorithm is used to cluster the sentences of the input document using different text representations, such as BoW (bag-of-words), TF-IDF (term frequency - inverse document frequency), Word2Vec (Word-to-vector), Glove, and FastText. The summary algorithm then uses the trained k-means model to rank the candidate sentences and create a summary with the highest-ranked sentences. The empirical results of the F1-score achieved 51.91% ROUGE-1, 18.77% ROUGE-2 and 29.72% ROUGE-L, compared to 52.33% ROUGE-1, 16.17% ROUGE-2, and 33.09% ROUGE-L performed using a competitive abstractive model. The advantage of the proposed model is that it can perform well with O(n,k,p) = O(n(k+2/p)) + O(nlog2n) + O(np) + O(nk2) + O(k) time complexity.

Recursive SPIHT(Set Partitioning in Hierarchy Trees) Algorithm for Embedded Image Coding (내장형 영상코딩을 위한 재귀적 SPIHT 알고리즘)

  • 박영석
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.4 no.4
    • /
    • pp.7-14
    • /
    • 2003
  • A number of embedded wavelet image coding methods have been proposed since the introduction of EZW(Embedded Zerotree Wavelet) algorithm. A common characteristic of these methods is that they use fundamental ideas found in the EZW algorithm. Especially, one of these methods is the SPIHT(Set Partitioning in Hierarchy Trees) algorithm, which became very popular since it was able to achieve equal or better performance than EZW without having to use an arithmetic encoder. In this paper We propose a recursive set partitioning in hierarchy trees(RSPIHT) algorithm for embedded image coding and evaluate it's effectiveness experimentally. The proposed RSPIHT algorithm takes the simple and regular form and the worst case time complexity of O(n). From the viewpoint of processing time, the RSPIHT algorithm takes about 16.4% improvement in average than the SPIHT algorithm at T-layer over 4 of experimental images. Also from the viewpoint of coding rate, the RSPIHT algorithm takes similar results at T-layer under 7 but the improved results at other T-layer of experimental images.

  • PDF

An Efficient Computing Method of the Orthogonal Projection Matrix for the Balanced Factorial Design

  • Kim, Byung-Chun;Park, Jong-Tae
    • Journal of the Korean Statistical Society
    • /
    • v.22 no.2
    • /
    • pp.249-258
    • /
    • 1993
  • It is well known that design matrix X for any factorial design can be represented by a product $X = TX_o$ where T is replication matrix and $X_o$ is the corresponding balanced design matrix. Since $X_o$ consists of regular arrangement of 0's and 1's, we can easily find the spectral decomposition of $X_o',X_o$. Also using this we propose an efficient algorithm for computing the orthogonal projection matrix for a balanced factorial design.

  • PDF