• Title/Summary/Keyword: $A^*$ search algorithm

Search Result 3,553, Processing Time 0.036 seconds

Cycle Detection of Discrete Logarithm using an Array (배열을 이용한 이산대수의 사이클 검출)

  • Sang-Un Lee
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.23 no.5
    • /
    • pp.15-20
    • /
    • 2023
  • Until now, Pollard's Rho algorithm has been known as the most efficient way for discrete algebraic problems to decrypt symmetric keys. However, the algorithm is being studied on how to further reduce the complexity of O(${\sqrt{p}}$) performance, along with the disadvantage of having to store the giant stride m=⌈${\sqrt{p}}$⌉ data. This paper proposes an array method for cycle detection in discrete logarithms. The proposed method reduces the number of updates of stack memory by at least 73%. This is done by only updating the array when (xi<0.5xi-1)∩(xi<0.5(p-1)). The proposed array method undergoes the same number of modular calculation as stack method, but significantly reduces the number of updates and the execution time for array through the use of a binary search method.

An Adaptive Priority-based Sequenced Route Query Processing Method in Road Networks (도로 네트워크 환경에서 적응적 우선순위 기반의 순차적 경로 처리 기법)

  • Ryu, Hyeongcheol;Jung, Sungwon
    • KIISE Transactions on Computing Practices
    • /
    • v.20 no.12
    • /
    • pp.652-657
    • /
    • 2014
  • Given a starting point, destination point and various Points Of Interest (POIs), which contain a full or partial order, for a user to visit we wish to create, a sequenced route from the starting point to the destination point that includes one member of each POI type in a particular order. This paper proposes a method for finding the approximate shortest route between the start point, destination point and one member of each POI type. There are currently two algorithms that perform this task but they both have weaknesses. One of the algorithms only considers the distance between the visited POI (or starting point) and POI to visit next. The other algorithm chooses candidate points near the straight-line distance between the start point and destination but does not consider the order of visits on the corresponding network path. This paper outlines an algorithm that chooses the candidate points that are nearer to the network path between the start point and destination using network search. The algorithm looks for routes using the candidate points and finds the approximate shortest route by assigning an adaptive priority to the route that visits more POIs in a short amount of time.

A Searching Technique of the Weak Connectivity Boundary using Small Unmanned Aerial Vehicle in Wireless Tactical Data Networks (무선 전술 데이터 네트워크에서 소형 무안항공기를 이용한 연결성 약화 지역 탐색 기법)

  • Li, Jin;Song, Ju-Bin
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37 no.1C
    • /
    • pp.89-96
    • /
    • 2012
  • Since tactical robots are going to be grown and tactical data communications will be more network-centric, the reliability of wireless tactical data networks is going to be very important in the future. However, the connectivity of such wireless tactical data networks can be extremely uncertain in practical circumstances. In this paper, we propose a searching technique to find out the weak boundary area of the network connectivity using a small UAV(unmanned aerial vehicle) which has a simple polling access function to wireless nodes on the ground in wireless tactical data networks. The UA V calculates the network topology of the wireless tactical data networks and coverts it to the Lapalcian matrix. In the proposed algorithm, we iteratively search the eigenvalues and find a minimum cut in the network resulting in finding the weak boundary of the connectivity for the wireless tactical data networks. If a UAV works as a relay nodes for the weak area, we evaluate that the throughput performance of the proposed algorithm outperforms star connection method and MST(minimum Spanning Tree) connection method. The proposed algorithm can be applied for recovering the connectivity of wireless tactical data networks.

Efficient Collaboration Method Between CPU and GPU for Generating All Possible Cases in Combination (조합에서 모든 경우의 수를 만들기 위한 CPU와 GPU의 효율적 협업 방법)

  • Son, Ki-Bong;Son, Min-Young;Kim, Young-Hak
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.7 no.9
    • /
    • pp.219-226
    • /
    • 2018
  • One of the systematic ways to generate the number of all cases is a combination to construct a combination tree, and its time complexity is O($2^n$). A combination tree is used for various purposes such as the graph homogeneity problem, the initial model for calculating frequent item sets, and so on. However, algorithms that must search the number of all cases of a combination are difficult to use realistically due to high time complexity. Nevertheless, as the amount of data becomes large and various studies are being carried out to utilize the data, the number of cases of searching all cases is increasing. Recently, as the GPU environment becomes popular and can be easily accessed, various attempts have been made to reduce time by parallelizing algorithms having high time complexity in a serial environment. Because the method of generating the number of all cases in combination is sequential and the size of sub-task is biased, it is not suitable for parallel implementation. The efficiency of parallel algorithms can be maximized when all threads have tasks with similar size. In this paper, we propose a method to efficiently collaborate between CPU and GPU to parallelize the problem of finding the number of all cases. In order to evaluate the performance of the proposed algorithm, we analyze the time complexity in the theoretical aspect, and compare the experimental time of the proposed algorithm with other algorithms in CPU and GPU environment. Experimental results show that the proposed CPU and GPU collaboration algorithm maintains a balance between the execution time of the CPU and GPU compared to the previous algorithms, and the execution time is improved remarkable as the number of elements increases.

Top-down Hierarchical Clustering using Multidimensional Indexes (다차원 색인을 이용한 하향식 계층 클러스터링)

  • Hwang, Jae-Jun;Mun, Yang-Se;Hwang, Gyu-Yeong
    • Journal of KIISE:Databases
    • /
    • v.29 no.5
    • /
    • pp.367-380
    • /
    • 2002
  • Due to recent increase in applications requiring huge amount of data such as spatial data analysis and image analysis, clustering on large databases has been actively studied. In a hierarchical clustering method, a tree representing hierarchical decomposition of the database is first created, and then, used for efficient clustering. Existing hierarchical clustering methods mainly adopted the bottom-up approach, which creates a tree from the bottom to the topmost level of the hierarchy. These bottom-up methods require at least one scan over the entire database in order to build the tree and need to search most nodes of the tree since the clustering algorithm starts from the leaf level. In this paper, we propose a novel top-down hierarchical clustering method that uses multidimensional indexes that are already maintained in most database applications. Generally, multidimensional indexes have the clustering property storing similar objects in the same (or adjacent) data pares. Using this property we can find adjacent objects without calculating distances among them. We first formally define the cluster based on the density of objects. For the definition, we propose the concept of the region contrast partition based on the density of the region. To speed up the clustering algorithm, we use the branch-and-bound algorithm. We propose the bounds and formally prove their correctness. Experimental results show that the proposed method is at least as effective in quality of clustering as BIRCH, a bottom-up hierarchical clustering method, while reducing the number of page accesses by up to 26~187 times depending on the size of the database. As a result, we believe that the proposed method significantly improves the clustering performance in large databases and is practically usable in various database applications.

Cost-Based Directed Scheduling : Part II, An Inter-Job Cost Propagation Algorithm (비용기반 스케줄링 : Part II, 작업간 비용 전파 알고리즘)

  • Suh, Min-Soo;Kim, Jae-Kyeong
    • Journal of Intelligence and Information Systems
    • /
    • v.14 no.1
    • /
    • pp.117-129
    • /
    • 2008
  • The cost-based scheduling work has been done in both the Operations Research (OR) and Artificial Intelligence (AI) literature. To deal with more realistic problems, AI-based heuristic scheduling approach with non-regular performance measures has been studied. However, there has been little research effort to develop a full inter-job cost propagation algorithm (CPA) for different jobs having multiple downstream and upstream activities. Without such a CPA, decision-making in scheduling heuristics relies upon local, incomplete cost information, resulting in poor schedule performance from the overall cost minimizing objective. For such a purpose, we need two types of CPAs : intra-job CPA and inter-job CPA. Whenever there is a change in cost information of an activity in a job in the process of scheduling, the intra-job CPA updates cost curves of other activities connected through temporal constraints within the same job. The inter-job CPA extends cost propagation into other jobs connected through precedence relationships. By utilizing the cost information provided by CPAs, we propose cost-based scheduling heuristics that attempt to minimize the total schedule cost. This paper develops inter-job CPAs that create and update cost curves of each activity in each search state, and propagate cost information throughout a whole network of temporal constraints. Also we propose various cost-based scheduling heuristics that attempt to minimize the total schedule cost by utilizing the cost propagation algorithm.

  • PDF

A Hybrid Genetic Algorithm for the Identical Parallel Machine Total Tardiness Problem (동종 병렬기계에서 납기지연 최소화를 위한 혼합형 유전 알고리즘의 개발)

  • Choe, Hong-Jin;Lee, Jong-Yeong;Park, Mun-Won
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 2004.05a
    • /
    • pp.624-627
    • /
    • 2004
  • 본 연구는 동일한 병렬기계에서의 총 납기지연의 합을 최소화하는 일정계획 문제에 대해 다룬다. 이 문제는 Lenstra et al. (1977)에 의해 NP-hard로 알려져 있으며, 작업의 수와 기계의 수가 큰 현실적 문제에 대해 적절한 시간 내에 최적해를 찾는다는 것은 사실상 불가능하다. 따라서 본 연구에서는 이 문제를 해결하기 위하여 혼합형 유전 알고리즘(hybrid genetic algorithm)을 제안한다. 혼합형 유전 알고리즘에서는 임의로 발생시킨 모집단에 대해 먼저 유전 알고리즘(genetic algorithm)이 세대를 진행하며 해를 개선한다. 유전 알고리즘이 일정기간동안 더 이상 해를 개선하지 못하면, 부분탐색 알고리즘(local-search algorithm))이 유전 알고리즘의 모집단의 개체들에 대해 해의 개선을 시도한다. 즉, 부분 탐색 알고리즘은 모집단 속의 각각의 개체를 초기해로 하여 모집단 내의 개체 수만큼의 부분 최적해(local optimum)들을 구한다. 이렇게 구한 부분 최적해들로 새로운 모집단을 구성하면 다시 유전 알고리즘이 진행된다. 이 과정을 종료조건에 이를 때까지 번갈아가며 반복 수행한다. 본 연구에서 제안한 유전 알고리즘에서는 Bean(1994)이 제안한 Random key 방법으로 개체를 표현하였으며 Park(2000)이 제안한 3가지 교차 연산자들을 채용하였다. 부분탐색 알고리즘을 위해서는 쌍대교환(pair-wise interchange) 방법을 통해 이웃해를 생성하였다. 선행실험을 통하여 제안한 혼합형 유전알고리즘에서 사용하는 다양한 모수(parameter)값들을 최적화하였으며 알고리즘의 성능을 비교하기 위하여 기존의 알고리즘과도 비교실험을 수행하였다.복적인 지표가 채택되는 경우를 포함하고 있다. 셋째는 추상적이며 측정이 어려운 지표를 채택하고 있는 경우이다. 여기에는 지표에 대한 정확한 정의가 이루어져 있지 않아 피 평가자가 불필요하거나 과다한 평가 자료를 준비해야 하거나 평가자로 하여금 평가 시 혼돈을 유발할 가능성이 있거나, 또는 상위개념의 평가항목과 하위개념의 평가항목이 혼재되어 구분이 모호한 경우를 포함하고 있다. 바탕으로 '생태적 합리성'이라는 체계적인 지식교육을 거쳐서, '환경정의' 의식의 제고로 이어가고, 굵직한 '환경갈등'의 상황에서 뚜렷한 정치적 태도와 실천을 할 수 있는 '생태적 인간상'의 육성으로 나아갈 수 있어야 한다는 것이 필자의 생각이다. 이를 위해서는 어찌되었건 체험학습 영역에서는 환경현안에 대한 사회적 실천을 '교육 소재'로 삼을 수 있어야 하며, 교과학습 영역에서는 한국사회의 환경현안에 대한 정치경제적 접근을 외면하지 말고 교과서 저작의 소재로 삼을 수 있어야 하며, 이는 '환경관리주의'와 '녹색소비'에 머물러 있는 '환경 지식교육'과 실천을 한단계 진전시키는 작업으로 이어질 것이다. 이후 10년의 환경교육은 바로 '생태적 합리성'과 '환경정의'라는 두 '화두'에 터하여 세워져야 한다.배액에서 약해를 보였으나, 25% 야자지방산의 경우 50 ${\sim}$ 100배액 어디에서도 액해를 보이지 않았다. 별도로 적용한 시험에서, 토마토의 경우에도 25% 야자지방산 비누 50 ${\sim}$ 100배액 모두 약해를 발생하지 않았으나, 오이에서는 25% 야자지방산 비누 100배액에도 약해를 나타내었다. 12. 이상의 결과, 천연지방산을 이용하여 유기농업에 허용되는 각종의 살충비누를 제조할 수 있었으

  • PDF

Analyzing the Main Paths and Intellectual Structure of the Data Literacy Research Domain (데이터 리터러시 연구 분야의 주경로와 지적구조 분석)

  • Jae Yun Lee
    • Journal of the Korean Society for information Management
    • /
    • v.40 no.4
    • /
    • pp.403-428
    • /
    • 2023
  • This study investigates the development path and intellectual structure of data literacy research, aiming to identify emerging topics in the field. A comprehensive search for data literacy-related articles on the Web of Science reveals that the field is primarily concentrated in Education & Educational Research and Information Science & Library Science, accounting for nearly 60% of the total. Citation network analysis, employing the PageRank algorithm, identifies key papers with high citation impact across various topics. To accurately trace the development path of data literacy research, an enhanced PageRank main path algorithm is developed, which overcomes the limitations of existing methods confined to the Education & Educational Research field. Keyword bibliographic coupling analysis is employed to unravel the intellectual structure of data literacy research. Utilizing the PNNC algorithm, the detailed structure and clusters of the derived keyword bibliographic coupling network are revealed, including two large clusters, one with two smaller clusters and the other with five smaller clusters. The growth index and mean publishing year of each keyword and cluster are measured to pinpoint emerging topics. The analysis highlights the emergence of critical data literacy for social justice in higher education amidst the ongoing pandemic and the rise of AI chatbots. The enhanced PageRank main path algorithm, developed in this study, demonstrates its effectiveness in identifying parallel research streams developing across different fields.

Fast RSST Algorithm Using Link Classification and Elimination Technique (가지 분류 및 제거기법을 이용한 고속 RSST 알고리듬)

  • Hong, Won-Hak
    • 전자공학회논문지 IE
    • /
    • v.43 no.4
    • /
    • pp.43-51
    • /
    • 2006
  • Segmentation method using RSST has many advantages in extracting of accurate region boundaries and controlling the resolution of segmented result and so on. In this paper, we propose three fast RSST algorithms for image segmentation. In first method, we classify links according to weight size for fast link search. In the second method, very similar links before RSST construction are eliminated. In third method, the links of very small regions which are not important for human eye are eliminated. As a result, the total times elapsed for segmentation are reduced by about 10 $\sim$ 40 times, and reconstructed images based on the segmentation results show little degradation of PSNR and visual quality.

Camera Extrinsic Parameter Estimation using 2D Homography and LM Method based on PPIV Recognition (PPIV 인식기반 2D 호모그래피와 LM방법을 이용한 카메라 외부인수 산출)

  • Cha Jeong-Hee;Jeon Young-Min
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.43 no.2 s.308
    • /
    • pp.11-19
    • /
    • 2006
  • In this paper, we propose a method to estimate camera extrinsic parameter based on projective and permutation invariance point features. Because feature informations in previous research is variant to c.:men viewpoint, extraction of correspondent point is difficult. Therefore, in this paper, we propose the extracting method of invariant point features, and new matching method using similarity evaluation function and Graham search method for reducing time complexity and finding correspondent points accurately. In the calculation of camera extrinsic parameter stage, we also propose two-stage motion parameter estimation method for enhancing convergent degree of LM algorithm. In the experiment, we compare and analyse the proposed method with existing method by using various indoor images to demonstrate the superiority of the proposed algorithms.