• Title/Summary/Keyword: Search Ranking Model

Search Result 27, Processing Time 0.021 seconds

Implementation of GA Processor with Multiple Operators, Based on Subpopulation Architecture (분할구조 기반의 다기능 연산 유전자 알고리즘 프로세서의 구현)

  • Cho Min-Sok;Chung Duck-Jin
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.52 no.5
    • /
    • pp.295-304
    • /
    • 2003
  • In this paper, we proposed a hardware-oriented Genetic Algorithm Processor(GAP) based on subpopulation architecture for high-performance convergence and reducing computation time. The proposed architecture was applied to enhancing population diversity for correspondence to premature convergence. In addition, the crossover operator selection and linear ranking subpop selection were newly employed for efficient exploration. As stochastic search space selection through linear ranking and suitable genetic operator selection with respect to the convergence state of each subpopulation was used, the elapsed time of searching optimal solution was shortened. In the experiments, the computation speed was increased by over $10\%$ compared to survival-based GA and Modified-tournament GA. Especially, increased by over $20\%$ in the multi-modal function. The proposed Subpop GA processor was implemented on FPGA device APEX EP20K600EBC652-3 of AGENT 2000 design kit.

A Model for Ranking Semantic Associations in a Social Network (소셜 네트워크에서 관계 랭킹 모델)

  • Oh, Sunju
    • The Journal of Society for e-Business Studies
    • /
    • v.18 no.3
    • /
    • pp.93-105
    • /
    • 2013
  • Much Interest has focused on social network services such as Facebook and Twitter. Previous research conducted on social network often emphasized the architecture of the social network that is the existence of path between any objects on network and the centrality of the object in the network. However, studies on the semantic association in the network are rare. Studies on searching semantic associations between entities are necessary for future business enhancements. In this research, the ontology based social network analysis is performed. A new method to search and rank relation sequences that consist of several relations between entities is proposed. In addition, several heuristics to measure the strength of the relation sequences are proposed. To evaluate the proposed method, an experiment was performed. A group of social relationships among the university and organizations are constructed. Some social connections are searched using the proposed ranking method. The proposed method is expected to be used to search the association among entities in ontology based knowledge base.

Implementation of a Parallel Web Crawler for the Odysseus Large-Scale Search Engine (오디세우스 대용량 검색 엔진을 위한 병렬 웹 크롤러의 구현)

  • Shin, Eun-Jeong;Kim, Yi-Reun;Heo, Jun-Seok;Whang, Kyu-Young
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.14 no.6
    • /
    • pp.567-581
    • /
    • 2008
  • As the size of the web is growing explosively, search engines are becoming increasingly important as the primary means to retrieve information from the Internet. A search engine periodically downloads web pages and stores them in the database to provide readers with up-to-date search results. The web crawler is a program that downloads and stores web pages for this purpose. A large-scale search engines uses a parallel web crawler to retrieve the collection of web pages maximizing the download rate. However, the service architecture or experimental analysis of parallel web crawlers has not been fully discussed in the literature. In this paper, we propose an architecture of the parallel web crawler and discuss implementation issues in detail. The proposed parallel web crawler is based on the coordinator/agent model using multiple machines to download web pages in parallel. The coordinator/agent model consists of multiple agent machines to collect web pages and a single coordinator machine to manage them. The parallel web crawler consists of three components: a crawling module for collecting web pages, a converting module for transforming the web pages into a database-friendly format, a ranking module for rating web pages based on their relative importance. We explain each component of the parallel web crawler and implementation methods in detail. Finally, we conduct extensive experiments to analyze the effectiveness of the parallel web crawler. The experimental results clarify the merit of our architecture in that the proposed parallel web crawler is scalable to the number of web pages to crawl and the number of machines used.

Security Constrained Optimal Power Flow by Hybrid Algorithms (하이브리드 알고리즘을 응용하여 안전도제약을 만족시키는 최적전력조류)

  • Kim, Gyu-Ho;Lee, Sang-Bong;Lee, Jae-Gyu;Yu, Seok-Gu
    • The Transactions of the Korean Institute of Electrical Engineers A
    • /
    • v.49 no.6
    • /
    • pp.305-311
    • /
    • 2000
  • This paper presents a hybrid algorithm for solving optimal power flow(OPF) in order to enhance a systems capability to cope with outages, which is based on combined application of evolutionary computation and local search method. The efficient algorithm combining main advantages of two methods is as follows : Firstly, evolutionary computation is used to perform global exploitation among a population. This gives a good initial point of conventional method. Then, local methods are used to perform local exploitation. The hybrid approach often outperforms either method operating alone and reduces the total computation time. The objective function of the security constrained OPF is the minimization of generation fuel costs and real power losses. The resulting optimal operating point has to be feasible after outages such as any single line outage(respect of voltage magnitude, reactive power generation and power flow limits). In OPF considering security, the outages are selected by contingency ranking method(contingency screening model). The OPF considering security, the outages are selected by contingency ranking method(contingency screening model). The method proposed is applied to IEEE 30 buses system to show its effectiveness.

  • PDF

Implementation of Content-based Image Retrieval System using Color Spatial and Shape Information (칼라 공간과 형태 정보를 이용한 내용기반 이미지 검색 시스템 구현)

  • Ban, Hong-Oh;Kang, Mun-Ju;Choi, Heyung-Jin
    • The KIPS Transactions:PartB
    • /
    • v.10B no.6
    • /
    • pp.681-686
    • /
    • 2003
  • In recent years automatic image indexing and retrieval have been increasingly studied. However, content-based retrieval techniques for general images are still inadequate for many purposes. The novelty and originality of this thesis are the definition and use of a spatial information model as a contribution to the accuracy and efficiency of image search. In addition, the model is applied to represent color and shape image contents as a vector using the method of image features extraction, which was inspired by the previous work on the study of human visual perception. The indexing scheme using the color, shape and spatial model shows the potential of being applied with the well-developed algorithms of features extraction and image search, like ranking operations. To conclude, user can retrieved more similar images with high precision and fast speed using the proposed system.

Proposal of a Hypothesis Test Prediction System for Educational Social Precepts using Deep Learning Models

  • Choi, Su-Youn;Park, Dea-Woo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.25 no.9
    • /
    • pp.37-44
    • /
    • 2020
  • AI technology has developed in the form of decision support technology in law, patent, finance and national defense and is applied to disease diagnosis and legal judgment. To search real-time information with Deep Learning, Big data Analysis and Deep Learning Algorithm are required. In this paper, we try to predict the entrance rate to high-ranking universities using a Deep Learning model, RNN(Recurrent Neural Network). First, we analyzed the current status of private academies in administrative districts and the number of students by age in administrative districts, and established a socially accepted hypothesis that students residing in areas with a high educational fever have a high rate of enrollment in high-ranking universities. This is to verify based on the data analyzed using the predicted hypothesis and the government's public data. The predictive model uses data from 2015 to 2017 to learn to predict the top enrollment rate, and the trained model predicts the top enrollment rate in 2018. A prediction experiment was performed using RNN, a Deep Learning model, for the high-ranking enrollment rate in the special education zone. In this paper, we define the correlation between the high-ranking enrollment rate by analyzing the household income and the participation rate of private education about the current status of private institutes in regions with high education fever and the effect on the number of students by age.

Re-ranking the Results from Two Image Retrieval System in Cooperative Manner (두 영상검색 시스템의 협력적 이용을 통한 재순위화)

  • Hwang, Joong-Won;Kim, Hyunwoo;Kim, Junmo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.24 no.1
    • /
    • pp.7-15
    • /
    • 2014
  • Image retrieval has become a huge part of computer vision and data mining. Although commercial image retrieval systems such as Google show great performances, the improvement on the performances are constantly on demand because of the rapid growth of data on web space. To satisfy the demand, many re-ranking methods, which enhance the performances by reordering retrieved results with independent algorithms, has been proposed. Conventional re-ranking algorithms are based on the assumption that visual patterns are not used on initial image retrieval stage. However, image search engines in present have begun to use the visual and the assumption is required to be reconsidered. Also, though it is possible to suspect that integration of multiple retrieval systems can improve the overall performance, the research on the topic has not been done sufficiently. In this paper, we made the condition that other manner than cooperation cannot improve the ranking result. We evaluate the algorithm on toy model and show that propose module can improve the retrieval results.

A Ranking Algorithm for Semantic Web Resources: A Class-oriented Approach (시맨틱 웹 자원의 랭킹을 위한 알고리즘: 클래스중심 접근방법)

  • Rho, Sang-Kyu;Park, Hyun-Jung;Park, Jin-Soo
    • Asia pacific journal of information systems
    • /
    • v.17 no.4
    • /
    • pp.31-59
    • /
    • 2007
  • We frequently use search engines to find relevant information in the Web but still end up with too much information. In order to solve this problem of information overload, ranking algorithms have been applied to various domains. As more information will be available in the future, effectively and efficiently ranking search results will become more critical. In this paper, we propose a ranking algorithm for the Semantic Web resources, specifically RDF resources. Traditionally, the importance of a particular Web page is estimated based on the number of key words found in the page, which is subject to manipulation. In contrast, link analysis methods such as Google's PageRank capitalize on the information which is inherent in the link structure of the Web graph. PageRank considers a certain page highly important if it is referred to by many other pages. The degree of the importance also increases if the importance of the referring pages is high. Kleinberg's algorithm is another link-structure based ranking algorithm for Web pages. Unlike PageRank, Kleinberg's algorithm utilizes two kinds of scores: the authority score and the hub score. If a page has a high authority score, it is an authority on a given topic and many pages refer to it. A page with a high hub score links to many authoritative pages. As mentioned above, the link-structure based ranking method has been playing an essential role in World Wide Web(WWW), and nowadays, many people recognize the effectiveness and efficiency of it. On the other hand, as Resource Description Framework(RDF) data model forms the foundation of the Semantic Web, any information in the Semantic Web can be expressed with RDF graph, making the ranking algorithm for RDF knowledge bases greatly important. The RDF graph consists of nodes and directional links similar to the Web graph. As a result, the link-structure based ranking method seems to be highly applicable to ranking the Semantic Web resources. However, the information space of the Semantic Web is more complex than that of WWW. For instance, WWW can be considered as one huge class, i.e., a collection of Web pages, which has only a recursive property, i.e., a 'refers to' property corresponding to the hyperlinks. However, the Semantic Web encompasses various kinds of classes and properties, and consequently, ranking methods used in WWW should be modified to reflect the complexity of the information space in the Semantic Web. Previous research addressed the ranking problem of query results retrieved from RDF knowledge bases. Mukherjea and Bamba modified Kleinberg's algorithm in order to apply their algorithm to rank the Semantic Web resources. They defined the objectivity score and the subjectivity score of a resource, which correspond to the authority score and the hub score of Kleinberg's, respectively. They concentrated on the diversity of properties and introduced property weights to control the influence of a resource on another resource depending on the characteristic of the property linking the two resources. A node with a high objectivity score becomes the object of many RDF triples, and a node with a high subjectivity score becomes the subject of many RDF triples. They developed several kinds of Semantic Web systems in order to validate their technique and showed some experimental results verifying the applicability of their method to the Semantic Web. Despite their efforts, however, there remained some limitations which they reported in their paper. First, their algorithm is useful only when a Semantic Web system represents most of the knowledge pertaining to a certain domain. In other words, the ratio of links to nodes should be high, or overall resources should be described in detail, to a certain degree for their algorithm to properly work. Second, a Tightly-Knit Community(TKC) effect, the phenomenon that pages which are less important but yet densely connected have higher scores than the ones that are more important but sparsely connected, remains as problematic. Third, a resource may have a high score, not because it is actually important, but simply because it is very common and as a consequence it has many links pointing to it. In this paper, we examine such ranking problems from a novel perspective and propose a new algorithm which can solve the problems under the previous studies. Our proposed method is based on a class-oriented approach. In contrast to the predicate-oriented approach entertained by the previous research, a user, under our approach, determines the weights of a property by comparing its relative significance to the other properties when evaluating the importance of resources in a specific class. This approach stems from the idea that most queries are supposed to find resources belonging to the same class in the Semantic Web, which consists of many heterogeneous classes in RDF Schema. This approach closely reflects the way that people, in the real world, evaluate something, and will turn out to be superior to the predicate-oriented approach for the Semantic Web. Our proposed algorithm can resolve the TKC(Tightly Knit Community) effect, and further can shed lights on other limitations posed by the previous research. In addition, we propose two ways to incorporate data-type properties which have not been employed even in the case when they have some significance on the resource importance. We designed an experiment to show the effectiveness of our proposed algorithm and the validity of ranking results, which was not tried ever in previous research. We also conducted a comprehensive mathematical analysis, which was overlooked in previous research. The mathematical analysis enabled us to simplify the calculation procedure. Finally, we summarize our experimental results and discuss further research issues.

Cross-architecture Binary Function Similarity Detection based on Composite Feature Model

  • Xiaonan Li;Guimin Zhang;Qingbao Li;Ping Zhang;Zhifeng Chen;Jinjin Liu;Shudan Yue
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.8
    • /
    • pp.2101-2123
    • /
    • 2023
  • Recent studies have shown that the neural network-based binary code similarity detection technology performs well in vulnerability mining, plagiarism detection, and malicious code analysis. However, existing cross-architecture methods still suffer from insufficient feature characterization and low discrimination accuracy. To address these issues, this paper proposes a cross-architecture binary function similarity detection method based on composite feature model (SDCFM). Firstly, the binary function is converted into vector representation according to the proposed composite feature model, which is composed of instruction statistical features, control flow graph structural features, and application program interface calling behavioral features. Then, the composite features are embedded by the proposed hierarchical embedding network based on a graph neural network. In which, the block-level features and the function-level features are processed separately and finally fused into the embedding. In addition, to make the trained model more accurate and stable, our method utilizes the embeddings of predecessor nodes to modify the node embedding in the iterative updating process of the graph neural network. To assess the effectiveness of composite feature model, we contrast SDCFM with the state of art method on benchmark datasets. The experimental results show that SDCFM has good performance both on the area under the curve in the binary function similarity detection task and the vulnerable candidate function ranking in vulnerability search task.

Extracting Alternative Word Candidates for Patent Information Search (특허 정보 검색을 위한 대체어 후보 추출 방법)

  • Baik, Jong-Bum;Kim, Seong-Min;Lee, Soo-Won
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.15 no.4
    • /
    • pp.299-303
    • /
    • 2009
  • Patent information search is used for checking existence of earlier works. In patent information search, there are many reasons that fails to get appropriate information. This research proposes a method extracting alternative word candidates in order to minimize search failure due to keyword mismatch. Assuming that two words have similar meaning if they have similar co-occurrence words, the proposed method uses the concept of concentration, association word set, cosine similarity between association word sets and a ranking modification technique. Performance of the proposed method is evaluated using a manually extracted alternative word candidate list. Evaluation results show that the proposed method outperforms the document vector space model in recall.