• 제목/요약/키워드: Division Algorithm

검색결과 3,058건 처리시간 0.036초

Spatio- Temporal Join for Trajectory of Moving Objects in the Moving Object Database

  • Lee Jai-Ho;Nam Kwang-Woo;Kim Kwang-Soo
    • 대한원격탐사학회:학술대회논문집
    • /
    • 대한원격탐사학회 2004년도 Proceedings of ISRS 2004
    • /
    • pp.287-290
    • /
    • 2004
  • In the moving object database system, spatiotemporal join is very import operation when we process join moving objects. Processing time of spatio-temporal join operation increases by geometric progression with numbers of moving objects. Therefore efficient methods of spatio-temporal join is essential to moving object database system. In this paper, we propose spatio-temporal join algorithm with TB-Tree that preserves trajectories of moving objects, and show result of test. We first present basic algorithm, and propose cpu-time tunning algorithm and IO-time tunning algorithm. We show result of test with data set created by moving object generator tool.

  • PDF

TWDM PON의 동적대역할당 알고리즘 (Dynamic Bandwidth Allocation Algorithm of TWDM PON)

  • 홍성학;한만수
    • 한국정보통신학회:학술대회논문집
    • /
    • 한국정보통신학회 2013년도 춘계학술대회
    • /
    • pp.743-744
    • /
    • 2013
  • 본 논문에서는 XGPON (10-Gigabit-capable passive optical network)을 기반으로 하는 TWDM (time and wave-division multiplexing) PON의 새로운 동적대역할당 알고리즘을 제안한다. 대부분의 TWDM PON은 EPON (Ethernet PON)을 기반으로 하여 동작하며 동적대역할당도 EPON의 동적대역할당 알고리즘을 변형시켜 사용한다. 본 논문에서는 XGPON의 동적대역할당 알고리즘을 개선하여 TWDM PON에 적합한 새로운 동적대역할당 알고리즘을 제안한다.

  • PDF

Joint Subcarrier Matching and Power Allocation in OFDM Two-Way Relay Systems

  • Vu, Ha Nguyen;Kong, Hyung-Yun
    • Journal of Communications and Networks
    • /
    • 제14권3호
    • /
    • pp.257-266
    • /
    • 2012
  • A decode-and-forward two-way relay system benefits from orthogonal frequency division multiplexing (OFDM) and relay transmission. In this paper, we consider a decode-and-forward two-way relay system over OFDMwith two strategies: A joint subcarrier matching algorithm and a power allocation algorithm operating with a total power constraint for all subcarriers. The two strategies are studied based on average capacity using numerical analysis by uniformly allocating power constraints for each subcarrier matching group. An optimal subcarrier matching algorithm is proposed to match subcarriers in order of channel power gain for both transmission sides. Power allocation is defined based on equally distributing the capacity of each hop in each matching group. Afterward, a modified water-filling algorithm is also considered to allocate the power among all matching groups in order to increase the overall capacity of the network. Finally, Monte Carlo simulations are completed to confirm the numerical results and show the advantages of the joint subcarrier matching, power allocation and water filling algorithms, respectively.

Light Source Target Detection Algorithm for Vision-based UAV Recovery

  • Won, Dae-Yeon;Tahk, Min-Jea;Roh, Eun-Jung;Shin, Sung-Sik
    • International Journal of Aeronautical and Space Sciences
    • /
    • 제9권2호
    • /
    • pp.114-120
    • /
    • 2008
  • In the vision-based recovery phase, a terminal guidance for the blended-wing UAV requires visual information of high accuracy. This paper presents the light source target design and detection algorithm for vision-based UAV recovery. We propose a recovery target design with red and green LEDs. This frame provides the relative position between the target and the UAV. The target detection algorithm includes HSV-based segmentation, morphology, and blob processing. These techniques are employed to give efficient detection results in day and night net recovery operations. The performance of the proposed target design and detection algorithm are evaluated through ground-based experiments.

Vertical Edge Based Algorithm for Korean License Plate Extraction and Recognition

  • Yu, Mei;Kim, Yong Deak
    • 한국통신학회논문지
    • /
    • 제25권7A호
    • /
    • pp.1076-1083
    • /
    • 2000
  • Vehicle license plate recognition identifies vehicle as a unique, and have many applications in traffic monitoring field. In this paper, a vertical edge based algorithm to extract license plate within input gray-scale image is proposed. A size-and-shape filter based on seed-filling algorithm is applied to remove the edges that are impossible to be the vertical edges of license plate. Then the remaining edges are matched with each other according to some restricted conditions so as to locate license plate in input image. After license plate is extracted. normalized and segmented, the characters on it are recognized by template matching method. Experimental results show that the proposed algorithm can deal with license plates in normal shape effectively, as well as the license plates that are out of shape due to the angle of view.

  • PDF

Recognition of Car License Plate using Kohonen Algorithm

  • Lim, Eun-Kyoung;Yang, Hwang-Kyu;Kwang Baek kim
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2000년도 ITC-CSCC -2
    • /
    • pp.785-788
    • /
    • 2000
  • The recognition system of a car plate is largely classified as the extraction and recognition of number plate. In this paper, we extract the number plate domain by using a thresholding method as a preprocess step. The computation of the density in a given mask provides a clue of a candidate domain whose density ratio corresponds to the properties of the number plate obtained in the best condition. The contour of the number plate for the recognition of the texts of number plate is extracted by operating Kohonen Algorithm in a localized region. The algorithm reduces noises around the contour. The recognition system with the density computation and Kohonen Algorithm shows a high performance in the real system in connection with a car number plate.

  • PDF

Performance Optimization of Big Data Center Processing System - Big Data Analysis Algorithm Based on Location Awareness

  • Zhao, Wen-Xuan;Min, Byung-Won
    • International Journal of Contents
    • /
    • 제17권3호
    • /
    • pp.74-83
    • /
    • 2021
  • A location-aware algorithm is proposed in this study to optimize the system performance of distributed systems for processing big data with low data reliability and application performance. Compared with previous algorithms, the location-aware data block placement algorithm uses data block placement and node data recovery strategies to improve data application performance and reliability. Simulation and actual cluster tests showed that the location-aware placement algorithm proposed in this study could greatly improve data reliability and shorten the application processing time of I/O interfaces in real-time.

Community Discovery in Weighted Networks Based on the Similarity of Common Neighbors

  • Liu, Miaomiao;Guo, Jingfeng;Chen, Jing
    • Journal of Information Processing Systems
    • /
    • 제15권5호
    • /
    • pp.1055-1067
    • /
    • 2019
  • In view of the deficiencies of existing weighted similarity indexes, a hierarchical clustering method initialize-expand-merge (IEM) is proposed based on the similarity of common neighbors for community discovery in weighted networks. Firstly, the similarity of the node pair is defined based on the attributes of their common neighbors. Secondly, the most closely related nodes are fast clustered according to their similarity to form initial communities and expand the communities. Finally, communities are merged through maximizing the modularity so as to optimize division results. Experiments are carried out on many weighted networks, which have verified the effectiveness of the proposed algorithm. And results show that IEM is superior to weighted common neighbor (CN), weighted Adamic-Adar (AA) and weighted resources allocation (RA) when using the weighted modularity as evaluation index. Moreover, the proposed algorithm can achieve more reasonable community division for weighted networks compared with cluster-recluster-merge-algorithm (CRMA) algorithm.

COMPARISON OF SUB-SAMPLING ALGORITHM FOR LRIT IMAGE GENERATION

  • Bae, Hee-Jin;Ahn, Sang-Il
    • 대한원격탐사학회:학술대회논문집
    • /
    • 대한원격탐사학회 2007년도 Proceedings of ISRS 2007
    • /
    • pp.109-113
    • /
    • 2007
  • The COMS provides the LRIT/HRIT services to users. The COMS LRIT/HRIT broadcast service should satisfy the 15 minutes timeliness requirement. The requirement is important and critical enough to impact overall performance of the LHGS. HRIT image data is acquired from INRSM output receiving but LRIT image data is generated by sub-sampling HRIT image data in the LHGS. Specially, since LRIT is acquired from sub-sampled HRIT image data, LRIT processing spent more time. Besides, some of data loss for LRIT occurs since LRIT is compressed by lossy JPEG. Therefore, algorithm with the fastest processing speed and simplicity to be implemented should be selected to satisfy the requirement. Investigated sub-sampling algorithm for the LHGS were nearest neighbour algorithm, bilinear algorithm and bicubic algorithm. Nearest neighbour algorithm is selected for COMS LHGS considering the speed, simplicity and anti-aliasing corresponding to the guideline of user (KMA: Korea Meteorological Administration) to maintain the most cloud itself information in a view of meteorology. But the nearest neighbour algorithm is known as the worst performance. Therefore, it is studied in this paper that the selection of nearest neighbour algorithm for the LHGS is reasonable. First of all, characteristic of 3 sub-sampling algorithms is studied and compared. Then, several sub-sampling algorithm were applied to MTSAT-1R image data corresponding to COMS HRIT. Also, resized image was acquired from sub-sampled image with the identical sub-sampling algorithms applied to sub-sampling from HRIT to LRIT. And the difference between original image and resized image is compared. Besides, PSNR and MSE are calculated for each algorithm. This paper shows that it is appropriate to select nearest neighbour algorithm for COMS LHGS since sub-sampled image by nearest neighbour algorithm is little difference with that of other algorithms in quality performance from PSNR.

  • PDF

방사성폐기물 핵종분석 검증용 이상 탐지를 위한 인공지능 기반 알고리즘 개발 (Development of an Anomaly Detection Algorithm for Verification of Radionuclide Analysis Based on Artificial Intelligence in Radioactive Wastes)

  • 장승수;이장희;김영수;김지석;권진형;김송현
    • 방사선산업학회지
    • /
    • 제17권1호
    • /
    • pp.19-32
    • /
    • 2023
  • The amount of radioactive waste is expected to dramatically increase with decommissioning of nuclear power plants such as Kori-1, the first nuclear power plant in South Korea. Accurate nuclide analysis is necessary to manage the radioactive wastes safely, but research on verification of radionuclide analysis has yet to be well established. This study aimed to develop the technology that can verify the results of radionuclide analysis based on artificial intelligence. In this study, we propose an anomaly detection algorithm for inspecting the analysis error of radionuclide. We used the data from 'Updated Scaling Factors in Low-Level Radwaste' (NP-5077) published by EPRI (Electric Power Research Institute), and resampling was performed using SMOTE (Synthetic Minority Oversampling Technique) algorithm to augment data. 149,676 augmented data with SMOTE algorithm was used to train the artificial neural networks (classification and anomaly detection networks). 324 NP-5077 report data verified the performance of networks. The anomaly detection algorithm of radionuclide analysis was divided into two modules that detect a case where radioactive waste was incorrectly classified or discriminate an abnormal data such as loss of data or incorrectly written data. The classification network was constructed using the fully connected layer, and the anomaly detection network was composed of the encoder and decoder. The latter was operated by loading the latent vector from the end layer of the classification network. This study conducted exploratory data analysis (i.e., statistics, histogram, correlation, covariance, PCA, k-mean clustering, DBSCAN). As a result of analyzing the data, it is complicated to distinguish the type of radioactive waste because data distribution overlapped each other. In spite of these complexities, our algorithm based on deep learning can distinguish abnormal data from normal data. Radionuclide analysis was verified using our anomaly detection algorithm, and meaningful results were obtained.