• Title/Summary/Keyword: Ratio algorithm

Search Result 3,109, Processing Time 0.033 seconds

Whale Optimization Algorithm and Blockchain Technology for Intelligent Networks

  • Sulthana, Shazia;Reddy, BN Manjunatha
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.7
    • /
    • pp.157-164
    • /
    • 2022
  • The proposed privacy preserving scheme has identified the drawbacks of existing schemes in Vehicular Networks. This prototype enhances the number of nodes by decreasing the cluster size. This algorithm is integrated with the whale optimization algorithm and Block Chain Technology. A set of results are done through the NS-2 simulator in the direction to check the effectiveness of proposed algorithm. The proposed method shows better results than with the existing techniques in terms of Delay, Drop, Delivery ratio, Overhead, throughout under the denial of attack.

A Study on the Improvement of Wavelet-Based Best-Basis Algorithm for Image Compression (영상압축을 위한 웨이브릿 기반 Best-Basis 알고리즘의 개선에 관한 연구)

  • 안종구;추형석;박제선
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.52 no.10
    • /
    • pp.591-597
    • /
    • 2003
  • In this paper, a best-basis selection algorithm that improves the performance of the coding gains and the computational complexity is proposed. The proposed algorithm limits the computational complexity according to the resolved threshold value and decomposes the parent subbands by using the top-down tree search and the relative energy between the parent subbands and the child subbands. For the experiments of the proposed algorithm, the bit-rates, the peak signal-to-noise ratio (PSNR), and the reconstructed images are presented by using the Quad-tree coder. The result of the proposed algorithm is compared to that of DWT algorithm using the Quad-tree coder for a set of standard test images. In addition, the result of the proposed algorithm is compared to that of JPEG-2000 algorithm and that of S+P algorithm.

Segmented Douglas-Peucker Algorithm Based on the Node Importance

  • Wang, Xiaofei;Yang, Wei;Liu, Yan;Sun, Rui;Hu, Jun;Yang, Longcheng;Hou, Boyang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.4
    • /
    • pp.1562-1578
    • /
    • 2020
  • Vector data compression algorithm can meet requirements of different levels and scales by reducing the data amount of vector graphics, so as to reduce the transmission, processing time and storage overhead of data. In view of the fact that large threshold leading to comparatively large error in Douglas-Peucker vector data compression algorithm, which has difficulty in maintaining the uncertainty of shape features and threshold selection, a segmented Douglas-Peucker algorithm based on node importance is proposed. Firstly, the algorithm uses the vertical chord ratio as the main feature to detect and extract the critical points with large contribution to the shape of the curve, so as to ensure its basic shape. Then, combined with the radial distance constraint, it selects the maximum point as the critical point, and introduces the threshold related to the scale to merge and adjust the critical points, so as to realize local feature extraction between two critical points to meet the requirements in accuracy. Finally, through a large number of different vector data sets, the improved algorithm is analyzed and evaluated from qualitative and quantitative aspects. Experimental results indicate that the improved vector data compression algorithm is better than Douglas-Peucker algorithm in shape retention, compression error, results simplification and time efficiency.

A New Fast Training Algorithm for Vector Quantizer Design (벡터양자화기의 코드북을 구하는 새로운 고속 학습 알고리듬)

  • Lee, Dae-Ryong;Baek, Seong-Joon;Sung, Koeng-Mo
    • The Journal of the Acoustical Society of Korea
    • /
    • v.15 no.5
    • /
    • pp.107-112
    • /
    • 1996
  • In this paper we propose a new fast codebook training algorithm for reducing the searching time of LBG algorithm. For each training data, the proposed algorithm stores the indexes of codewords that are close to that training data in the first iteration. It reduces computation time by searching only those codewords, the indexes of which are stored for each training data. Compared to one of the previous fast training algorithm, FSLBG, it obtains a better codebook with less exccution time. In our experiment, the performance of the codebook generated by the proposed algorithm in terms of peak signal-to-noise ratio(TSNR) is very close to that of LBG algorithm. However, the codewords to be searched for each training data of the proposed algorithm is only about 6%, for a codebook size of 256 and 1.6%, for a codebook size of 1.24, of LBG algorithm.

  • PDF

Efficient Cooperative Caching Algorithm for Distributed File Systems (분산 파일시스템을 위한 효율적인 협력캐쉬 알고리즘)

  • 박새미;이석재;유재수
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2003.11a
    • /
    • pp.234-244
    • /
    • 2003
  • In distributed file-systems, cooperative caching algorithm which owns the data cached at each node jointly is used to reduce an expense of disk access. Cooperative caching algorithm is the method that increases a cache hit-ratio and decrease a disk access as it holds the cache information of distributed systems in common and makes cache larger virtually. Recently, several cooperative caching algorithms decrease the message costs by using approximate information of the cache and increase the cache hit-ratio by using local and global cache fields dynamically. And they have an advantage that increases the whole field hit-ratio by sending a replaced block to the idel node on cache replacement in order to maintain the replaced block in the cache field. However the wrong approximate information deteriorates the performance, the concistency maintenance goes to great expense to exchange messeges and the cost that manages Age-information of each node to choose the idle node increases. In this thesis, we propose a cooperative cache algorithm that maintains correct cache information, minimizes the maintance cost for consistency and the management cost for cache Age-information. Also, we show the superiority of our algorithm through the performance evaluation.

  • PDF

Efficient Flooding Algorithm for Mobile P2P Systems using Super Peer (슈퍼피어를 이용한 모바일 P2P시스템을 위한 효율적인 플러딩 알고리즘)

  • Kang, So-Young;Lee, Kwang-Jo;Yang, Sung-Bong
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.2
    • /
    • pp.217-221
    • /
    • 2010
  • As the appearances of various new mobile devices and the explosive growth of mobile device users, many researches related to mobile P2P systems have been proceeded actively. In this paper, we propose a new search algorithm for the double-layered super peer system in the mobile environment. For the proposed search algorithm, we divide the entire experiment region into a grid of cells, each of which has the same size. The grid is configured properly by considering the communication range of a mobile device and the number of peers in the system. The proposed search algorithm is a partial flooding search method based on the directions of cells involved with the search. It reduces successfully the network traffic, but shows a low search hit ratio. To enhance the search hit ratio, we introduce a bridge-peer table for a super peer and utilize an n-way search. The experimental results show that the proposed algorithm made an average of 20~30% reduction in the number of message packets over the double-layered system. The success ratio was also improved about 2~5% over the double-layered system.

Document Classification of Small Size Documents Using Extended Relief-F Algorithm (확장된 Relief-F 알고리즘을 이용한 소규모 크기 문서의 자동분류)

  • Park, Heum
    • The KIPS Transactions:PartB
    • /
    • v.16B no.3
    • /
    • pp.233-238
    • /
    • 2009
  • This paper presents an approach to the classifications of small size document using the instance-based feature filtering Relief-F algorithm. In the document classifications, we have not always good classification performances of small size document included a few features. Because total number of feature in the document set is large, but feature count of each document is very small relatively, so the similarities between documents are very low when we use general assessment of similarity and classifiers. Specially, in the cases of the classification of web document in the directory service and the classification of the sectors that cannot connect with the original file after recovery hard-disk, we have not good classification performances. Thus, we propose the Extended Relief-F(ERelief-F) algorithm using instance-based feature filtering algorithm Relief-F to solve problems of Relief-F as preprocess of classification. For the performance comparison, we tested information gain, odds ratio and Relief-F for feature filtering and getting those feature values, and used kNN and SVM classifiers. In the experimental results, the Extended Relief-F(ERelief-F) algorithm, compared with the others, performed best for all of the datasets and reduced many irrelevant features from document sets.

IT-based Technology An Efficient Global Buffer Management ,algorithm for SAN Environments (SAN 환경을 위한 효율적인 전역버퍼 관리 알고리즘)

  • 이석재;박새미;송석일;유재수;이장선
    • The Journal of the Korea Contents Association
    • /
    • v.4 no.3
    • /
    • pp.71-80
    • /
    • 2004
  • In distributed file-systems, cooperative caching algorithm which owns the data cached at each node jointly is used to reduce an expense of disk access. Cooperative caching algorithm is the method that increases a cache hit-ratio and decrease a disk access as it holds the cache information of distributed systems in common and makes cache larger virtually. Recently, several cooperative caching algorithms decrease the message costs by using approximate information of the cache and increase the cache hit-ratio by using local and global cache fields dynamically. And they have an advantage that increases the whole field hit-ratio by sending a replaced buffer to the idle node on buffers replacement in order to maintain the replaced cache in the cache field. However the wrong approximate information deteriorates the performance, the consistency maintenance goes to great expense to exchange messages and the cost that manages Age-information of each node to choose the idle node increases. In this thesis, we propose a cooperative cache algorithm that maintains correct cache information, minimizes the maintenance cost for consistency and the management cost for buffer Age-information. Also, we show the superiority of our algorithm through the performance evaluation.

  • PDF

High-Performance and Low-Complexity Decoding of High-Weight LDPC Codes (높은 무게 LDPC 부호의 저복잡도 고성능 복호 알고리즘)

  • Cho, Jun-Ho;Sung, Won-Yong
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.5C
    • /
    • pp.498-504
    • /
    • 2009
  • A high-performance low-complexity decoding algorithm for LDPC codes is proposed in this paper, which has the advantages of both bit-flipping (BF) algorithm and sum-product algorithm (SPA). The proposed soft bit-flipping algorithm requires only simple comparison and addition operations for computing the messages between bit and check nodes, and the amount of those operations is also small. By increasing the utilization ratio of the computed messages and by adopting nonuniform quantization, the signal-to-noise ratio (SNR) gap to the SPA is reduced to 0.4dB at the frame error rate of 10-4 with only 5-bit assignment for quantization. LDPC codes with high column or row weights, which are not suitable for the SPA decoding due to the complexity, can be practically implemented without much worsening the error performance.

WRR Cell Scheduling Algorithm of BSW structure (BSW구조의 셀 스케쥴링 알고리즘)

  • 조해성;임청규;전병실
    • Journal of the Korea Society of Computer and Information
    • /
    • v.5 no.3
    • /
    • pp.119-125
    • /
    • 2000
  • A network of Asynchronous Transfer Mode (ATM) will be required to carry the traffics(CVR, VBR, UBR, ABR) generated by a wide range of services. The algorithm of WRR cell multiplexing is designed to serve no only CBR, VBR traffic but also ABR, UBR traffic in ATM. BSW algorithm was Proposed to carry on manage buffer efficiently at implementing of WRR scheduler. But, BSW a1gorithm cause serious degradation to the weight of each VC and the ratio of scheduler throughput because it allocates more weight than the weight allocated actually in VC and because it could not serve cell if the VC queue is empty. In this paper, we propose the WRR scheduling algorithm of BSW structure which improves the cell service ratio and cell delay. The proposed algorithm is capable of maintaining an allocated VC's weight correctly and decrease of average cell delay by serving other VC cell when empty in each VC queue and increase of cell service ratio as a whole.

  • PDF