• 제목/요약/키워드: 서비스 검색 알고리즘

Search Result 199, Processing Time 0.029 seconds

Mapping algorithm for Error Compensation of Indoor Localization System (실내 측위 시스템의 오차 보정을 위한 매핑 알고리즘)

  • Kim, Tae-Kyum;Cho, We-Duke
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.47 no.4
    • /
    • pp.109-117
    • /
    • 2010
  • With the advent of new technologies such as HSDPA, WiBro(Wireless Broadband) and personal devices, we can access various contents and services anytime and anywhere. A location based service(LBS) is essential for providing personalized services with individual location information in ubiquitous computing environment. In this paper, we propose mapping algorithm for error compensation of indoor localization system. Also we explain filter and indoor localization system. we have developed mapping algorithms composed of a map recognition method and a position compensation method. The map recognition method achieves physical space recognition and map element relation extraction. We improved the accuracy of position searching. In addition, we reduced position errors using a dynamic scale factor.

An Algorithm to Detect P2P Heavy Traffic based on Flow Transport Characteristics (플로우 전달 특성 기반의 P2P 헤비 트래픽 검출 알고리즘)

  • Choi, Byeong-Geol;Lee, Si-Young;Seo, Yeong-Il;Yu, Zhibin;Jun, Jae-Hyun;Kim, Sung-Ho
    • Journal of KIISE:Information Networking
    • /
    • v.37 no.5
    • /
    • pp.317-326
    • /
    • 2010
  • Nowadays, transmission bandwidth for network traffic is increasing and the type is varied such as peer-to-peer (PZP), real-time video, and so on, because distributed computing environment is spread and various network-based applications are developed. However, as PZP traffic occupies much volume among Internet backbone traffics, transmission bandwidth and quality of service(QoS) of other network applications such as web, ftp, and real-time video cannot be guaranteed. In previous research, the port-based technique which checks well-known port number and the Deep Packet Inspection(DPI) technique which checks the payload of packets were suggested for solving the problem of the P2P traffics, however there were difficulties to apply those methods to detection of P2P traffics because P2P applications are not used well-known port number and payload of packets may be encrypted. A proposed algorithm for identifying P2P heavy traffics based on flow transport parameters and behavioral characteristics can solve the problem of the port-based technique and the DPI technique. The focus of this paper is to identify P2P heavy traffic flows rather than all P2P traffics. P2P traffics are consist of two steps i)searching the opposite peer which have some contents ii) downloading the contents from one or more peers. We define P2P flow patterns on these P2P applications' features and then implement the system to classify P2P heavy traffics.

Searching a Navigation Path to Avoid Danger Area for Safe Driving (안전운전을 위해 위험지역을 회피하는 내비게이션 경로탐색)

  • Lee, Yong-Hu;Kim, Sang-Woon
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.8
    • /
    • pp.171-179
    • /
    • 2013
  • The primary function of navigation system is to provide route search and road guidance for safe driving for drivers. However, the existing route search system provides a simple service that looks up the shortest route using a safe driving DB without considering different road characteristics for the safety of the drivers. In order to maintain the safe driving, rather than searching the shortest path, a navigation system, in which the danger areas and/or the dangerous time zones have been considered, is required. Therefore, in this paper we propose a strategy of searching a navigation path to avoid danger areas for safe driving by using the A* algorithm. In the strategy, when evaluating the path-specific fitness of the navigation nodes, different heuristic weights were assigned to different types of risk areas. In particular, we considered three kinds of danger areas, such as accident-prone sections where accidents occur frequently, school zones, and intersection regions, as well as the time slots when the probability of danger is high. From computer simulation, the results demonstrate that the proposed scheme can provide the way to avoid danger areas on the route searching and confirm the possibility of providing the actual service.

An Efficient Split Algorithm to Minimize the Overlap between Node Index Spaces in a Multi-dimensional Indexing Scheme M-tree (다차원 색인구조 M-트리에서 노드 색인 공간의 중첩을 최소화하기 위한 효율적인 분할 알고리즘)

  • Im Sang-hyuk;Ku Kyong-I;Kim Ki-chang;Kim Yoo-Sung
    • The KIPS Transactions:PartD
    • /
    • v.12D no.2 s.98
    • /
    • pp.233-246
    • /
    • 2005
  • To enhance the user response time of content-based retrieval service for multimedia information, several multi-dimensional index schemes have been proposed. M-tree, a well-known multidimensional index scheme is of metric space access method, and is based on the distance between objects in the metric space. However, since the overlap between index spaces of nodes might enlarge the number of nodes of M-tree accessed for query processing, the user response time for content-based multimedia information retrieval grows longer. In this paper, we propose a node split algorithm which is able to reduce the sire of overlap between index spaces of nodes in M-tree. In the proposed scheme, we choose a virtual center point as the routing object and entry redistribution as the postprocessing after node split in order to reduce the radius of index space of a node, and finally in order to reduce the overlap between the index spaces of routing nodes. From the experimental results, we can see the proposed split algorithm reduce the overlap between index space of nodes and finally enhance the user response time for similarity-based query processing.

Ontology-based User Customized Search Service Considering User Intention (온톨로지 기반의 사용자 의도를 고려한 맞춤형 검색 서비스)

  • Kim, Sukyoung;Kim, Gunwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.4
    • /
    • pp.129-143
    • /
    • 2012
  • Recently, the rapid progress of a number of standardized web technologies and the proliferation of web users in the world bring an explosive increase of producing and consuming information documents on the web. In addition, most companies have produced, shared, and managed a huge number of information documents that are needed to perform their businesses. They also have discretionally raked, stored and managed a number of web documents published on the web for their business. Along with this increase of information documents that should be managed in the companies, the need of a solution to locate information documents more accurately among a huge number of information sources have increased. In order to satisfy the need of accurate search, the market size of search engine solution market is becoming increasingly expended. The most important functionality among much functionality provided by search engine is to locate accurate information documents from a huge information sources. The major metric to evaluate the accuracy of search engine is relevance that consists of two measures, precision and recall. Precision is thought of as a measure of exactness, that is, what percentage of information considered as true answer are actually such, whereas recall is a measure of completeness, that is, what percentage of true answer are retrieved as such. These two measures can be used differently according to the applied domain. If we need to exhaustively search information such as patent documents and research papers, it is better to increase the recall. On the other hand, when the amount of information is small scale, it is better to increase precision. Most of existing web search engines typically uses a keyword search method that returns web documents including keywords which correspond to search words entered by a user. This method has a virtue of locating all web documents quickly, even though many search words are inputted. However, this method has a fundamental imitation of not considering search intention of a user, thereby retrieving irrelevant results as well as relevant ones. Thus, it takes additional time and effort to set relevant ones out from all results returned by a search engine. That is, keyword search method can increase recall, while it is difficult to locate web documents which a user actually want to find because it does not provide a means of understanding the intention of a user and reflecting it to a progress of searching information. Thus, this research suggests a new method of combining ontology-based search solution with core search functionalities provided by existing search engine solutions. The method enables a search engine to provide optimal search results by inferenceing the search intention of a user. To that end, we build an ontology which contains concepts and relationships among them in a specific domain. The ontology is used to inference synonyms of a set of search keywords inputted by a user, thereby making the search intention of the user reflected into the progress of searching information more actively compared to existing search engines. Based on the proposed method we implement a prototype search system and test the system in the patent domain where we experiment on searching relevant documents associated with a patent. The experiment shows that our system increases the both recall and precision in accuracy and augments the search productivity by using improved user interface that enables a user to interact with our search system effectively. In the future research, we will study a means of validating the better performance of our prototype system by comparing other search engine solution and will extend the applied domain into other domains for searching information such as portal.

Design and Implementation of Geographical Handoff System Using GPS Information (GPS정보를 이용한 위치기반 핸드오프 시스템의 설계 및 구현)

  • Han, Seung-Ho;Yang, Seung-Chur;Kim, Jong-Deok
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.35 no.1A
    • /
    • pp.33-43
    • /
    • 2010
  • Recently, users want to use real-time multimedia services, such as internet, VoIP, etc., using their IEEE 802.11 wireless lan mobile stations. In order to provide such services, a handoff among access points is essential to support the mobility of a node, in such an wide area. However, the legacy handoff methods of IEEE 802.11 technology are easy to lose connections. Also, the recognition of a disconnection and channel re-searching time make the major delay of the next AP to connect. In addition, because IEEE 802.11 decides the selection of an AP depending only on received signal strength, regardless of a node direction, position, etc., it cannot guarantee a stable bandwidth for communication. Therefore, in order to provide a real-time multimedia service, a node must reduce the disconnection time and needs an appropriate algorithm to support a sufficient communication bandwidth. In this paper, we suggest an algorithm which predicts a handoff point of a moving node by using GPS location information, and guarantees a high transmission bandwidth according to the signal strength and the distance. We implemented the suggested algorithm, and confirmed the superiority of our algorithm by reducing around 3.7ms of the layer-2 disconnection time, and guaranteed 24.8% of the communication bandwidth.

An Efficient Top-k Query Processing Algorithm over Encrypted Outsourced-Data in the Cloud (아웃소싱 암호화 데이터에 대한 효율적인 Top-k 질의 처리 알고리즘)

  • Kim, Jong Wook;Suh, Young-Kyoon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.4 no.12
    • /
    • pp.543-548
    • /
    • 2015
  • Recently top-k query processing has been extremely important along with the explosion of data produced by a variety of applications. Top-k queries return the best k results ordered by a user-provided monotone scoring function. As cloud computing service has been getting more popular than ever, a hot attention has been paid to cloud-based data outsourcing in which clients' data are stored and managed by the cloud. The cloud-based data outsourcing, though, exposes a critical secuity concern of sensitive data, resulting in the misuse of unauthorized users. Hence it is essential to encrypt sensitive data before outsourcing the data to the cloud. However, there has been little attention to efficient top-k processing on the encrypted cloud data. In this paper we propose a novel top-k processing algorithm that can efficiently process a large amount of encrypted data in the cloud. The main idea of the algorithm is to prune unpromising intermediate results at the early phase without decrypting the encrypted data by leveraging an order-preserving encrypted technique. Experiment results show that the proposed top-k processing algorithm significantly reduces the overhead of client systems from 10X to 10000X.

Design and Implementation of XML based Global Peer-to-Peer Engine (XML기반 전역 Peer-to-Peer 엔진 설계 및 구현)

  • Kwon Tae-suk;Lee Il-su;Lee Sung-young
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.1B
    • /
    • pp.73-85
    • /
    • 2004
  • SIn this paper, we introduce our experience for designing and implementing new concept of a global XML-based Peer-to-Peer (P2P) engine to support various P2P applications, and interconnection among PC, Web and mobile computing environments. The proposed P2P engine can support to heterogeneous data exchanges and web interconnection by facilitating with the text-base XML while message exchange are necessary. It is also to provide multi-level security functions as well as to apply different types of security algorithms. The system consist of four modules; a message dispatcher to scheduling and filtering the message, a SecureNet to providing security services and data transmission, a Discovery Manager to constructing peer-to-peer networking, and a Repository Manager to processing data management including XML documents. As a feasibility test, we implement various P2P services such as chatting as a communication service, white-board as an authoring tool set during collaborative working, and a file system as a file sharing service. We also compared the proposed system to a Gnutella in order to measure performance of the systems.

Semi-automatic Ontology Modeling for VOD Annotation for IPTV (IPTV의 VOD 어노테이션을 위한 반자동 온톨로지 모델링)

  • Choi, Jung-Hwa;Heo, Gil;Park, Young-Tack
    • Journal of KIISE:Software and Applications
    • /
    • v.37 no.7
    • /
    • pp.548-557
    • /
    • 2010
  • In this paper, we propose a semi-automatic modeling approach of ontology to annotate VOD to realize the IPTV's intelligent searching. The ontology is made by combining partial tree that extracts hypernym, hyponym, and synonym of keywords related to a service domain from WordNet. Further, we add to the partial tree new keywords that are undefined in WordNet, such as foreign words and words written in Chinese characters. The ontology consists of two parts: generic hierarchy and specific hierarchy. The former is the semantic model of vocabularies such as keywords and contents of keywords. They are defined as classes including property restrictions in the ontology. The latter is generated using the reasoning technique by inferring contents of keywords based on the generic hierarchy. An annotation generates metadata (i.e., contents and genre) of VOD based on the specific hierarchy. The generic hierarchy can be applied to other domains, and the specific hierarchy helps modeling the ontology to fit the service domain. This approach is proved as good to generate metadata independent of any specific domain. As a result, the proposed method produced around 82% precision with 2,400 VOD annotation test data.

Sketch Map System using Clustering Method of XML Documents (XML 문서의 클러스터링 기법을 이용한 스케치맵 시스템)

  • Kim, Jung-Sook;Lee, Ya-Ri;Hong, Kyung-Pyo
    • The Journal of the Korea Contents Association
    • /
    • v.9 no.12
    • /
    • pp.19-30
    • /
    • 2009
  • The service that has recently come into the spotlight utilizes the map to first approach the map and then provide various mash-up formed results through the interface. This service can provide precise information to the users but the map is barely reusable. The sketch-map system of this paper, unlike the existing large map system, uses the method of presenting the specific spot and route in XML document and then clustering among sketch-maps. The map service system is designed to show the optimum route to the destination in a simple outline map. It is done by renovating the spot presented by the map into optimum contents. This service system, through the process of analyzing, splitting and clustering of the sketch-map's XML document input, creates a valid form of a sketch-map. It uses the LCS(Longest Common Subsequence) algorithm for splitting and merging sketch-map in the process of query. In addition, the simulation of this system's expected effects is provided. It shows how the maps that share information and knowledge assemble to form a large map and thus presents the system's ability and role as a new research portal.