• Title/Summary/Keyword: Indexing searching

Search Result 147, Processing Time 0.02 seconds

A Study on the Information Searching Behavior of MEDLINE Retrieval in Medical Librarian (의학전문사서의 정보이용행위에 관한 연구)

  • Lee Jin-Young;Jeong Sang-Kyung
    • Journal of Korean Library and Information Science Society
    • /
    • v.30 no.2
    • /
    • pp.123-153
    • /
    • 1999
  • This article aims at finding the ways, on the basis of the studies about the behaviors to search the existing CD-ROM databases, so that the searchers who retrieve the on-line MEDLINE used in the medical libraries can use the data more efficiently than now. We gave the questionnaires to the librarians in 60 medical libraries and searched the literatures and realities on the behaviors of the data uses to examine the search behaviors of the MEDLINE in the medical libraries. The result is as follows: 1) The medical data system rate for single users was $53\%$ and the ons for multi users $43\%$. As for the time which users retrieve for a week, under two hours was $75\%$, between 3 and 8 hours $18.3\%$, and eve. 9 hours $6.7\%$. 2) The increasing factors of the search result are (1) an enough discussion and interview between librarians and users, and (2) the use of the correct indexing terms, Thesaurus, and Keyword. In principle users must search directly. However, the librarians searched instead in case that the retrieval result was under two hours a week$(75\%)$. 3) As for the search fee, $91\%$ was free and $9\%$ was charged. Also search effectiveness was enhanced by the means of Inter-Library Loan Service & Information Network. 4) The medical librarians answered the questionnaire that they need the application education of professional knowledge, medical terms(thesaurus) and electronic medium, and also they need the computer education, interview technique and reeducation to give a satisfactory service. 5) As for the satisfactory degree of MEDLINE application, they answered $44.6\%$ for economy, $38.2\%$ for the conveniency of the time required, and $58.9\%$ for the users' search satisfaction answered respectively. 6) The application of MEDLINE system enhanced the medical libraries' image and had an effect on the users' satisfaction of using the data and search, the data activities and the research achievement. 7) In the past MeSH was used but as the time passes CD-ROM MEDLINE search behavior was preferred to On-line one.

  • PDF

Phase identification and degree of orientation measurements far fine-grained rock forming minerals using micro-area X-ray diffractometer -$Al_{2}SiO_{5}$ Polymorphs- (미소부 X-선 회절분석기를 이용한 미립조암광물의 상동정 및 배향도 측정 -$Al_{2}SiO_{5}$ 3상다형-)

  • 박찬수;김형식
    • The Journal of the Petrological Society of Korea
    • /
    • v.9 no.4
    • /
    • pp.205-210
    • /
    • 2000
  • Measurements of phase identification and degree of orientation for fine-grained (about 0.3 mm in diameter) minerals in rock samples performed by micro-area X-ray diffractometer.$Al_{2}SiO_{5}$ polymorphs (andalusite, kyanite and sillimanite) were chosen for the measurements and target minerals were existed on thin sections. Micro-area X-ray diffractometer is composed of 3(${\omega}\;{\chi}\;{\phi}$)-circle oscillating goniometer and position sensitive proportional counter (PSPC). $CuK_{\alpha}$ radiation was used as X-ray source and a pin hole ($50\;\mu\textrm{m}$$ in diameter) collimator was selected to focus radiation X-ray onto the target minerals. Phase identification and diffracted X-ray peak indexing were carried out by 3(${\omega}\;{\chi}\;{\phi}$)-circle oscillation measurement. Then, 2(${\omega}\;{\phi}$)-circle oscillation measurement was made for the purpose of searching the prevailing lattice plane of the minerals on thin section surface. Finally, for a selected peak by 2-circle oscillation measurement, X-ray pole figure measurement was executed for the purpose of check the degree of orientation of the single lattice direction and examine its pole distribution. As a result of 3-circle oscillation measurement, it was possible that phase identification among $Al_{2}SiO_{5}$ polymorphs. And from the results of 2-circle oscillation measurement and X-ray pole figure measurement, we recognized that poles of andalusite (122), kyanite (200) and sillimanite (310) lattice plances were well developed with direction normal to each mineral surface plane respectively. Therfore, the measurements used with micro-area X-ray diffractometer in this study will be a useful tool of phase identification and degree of orientation measurement for fine-grained rock forming minerals.

  • PDF

Design and Implementation of an Open Object Management System for Spatial Data Mining (공간 데이타 마이닝을 위한 개방형 객체 관리 시스템의 설계 및 구현)

  • Yun, Jae-Kwan;Oh, Byoung-Woo;Han, Ki-Joon
    • Journal of Korea Spatial Information System Society
    • /
    • v.1 no.1 s.1
    • /
    • pp.5-18
    • /
    • 1999
  • Recently, the necessity of automatic knowledge extraction from spatial data stored in spatial databases has been increased. Spatial data mining can be defined as the extraction of implicit knowledge, spatial relationships, or other knowledge not explicitly stored in spatial databases. In order to extract useful knowledge from spatial data, an object management system that can store spatial data efficiently, provide very fast indexing & searching mechanisms, and support a distributed computing environment is needed. In this paper, we designed and implemented an open object management system for spatial data mining, that supports efficient management of spatial, aspatial, and knowledge data. In order to develop this system, we used Open OODB that is a widely used object management system. However, the lark of facilities for spatial data mining in Open OODB, we extended it to support spatial data type, dynamic class generation, object-oriented inheritance, spatial index, spatial operations, etc. In addition, for further increasement of interoperability with other spatial database management systems or data mining systems, we adopted international standards such as ODMG 2.0 for data modeling, SDTS(Spatial Data Transfer Standard) for modeling and exchanging spatial data, and OpenGIS Simple Features Specification for CORBA for connecting clients and servers efficiently.

  • PDF

A Practical Approximate Sub-Sequence Search Method for DNA Sequence Databases (DNA 시퀀스 데이타베이스를 위한 실용적인 유사 서브 시퀀스 검색 기법)

  • Won, Jung-Im;Hong, Sang-Kyoon;Yoon, Jee-Hee;Park, Sang-Hyun;Kim, Sang-Wook
    • Journal of KIISE:Databases
    • /
    • v.34 no.2
    • /
    • pp.119-132
    • /
    • 2007
  • In molecular biology, approximate subsequence search is one of the most important operations. In this paper, we propose an accurate and efficient method for approximate subsequence search in large DNA databases. The proposed method basically adopts a binary trie as its primary structure and stores all the window subsequences extracted from a DNA sequence. For approximate subsequence search, it traverses the binary trie in a breadth-first fashion and retrieves all the matched subsequences from the traversed path within the trie by a dynamic programming technique. However, the proposed method stores only window subsequences of the pre-determined length, and thus suffers from large post-processing time in case of long query sequences. To overcome this problem, we divide a query sequence into shorter pieces, perform searching for those subsequences, and then merge their results. To verify the superiority of the proposed method, we conducted performance evaluation via a series of experiments. The results reveal that the proposed method, which requires smaller storage space, achieves 4 to 17 times improvement in performance over the suffix tree based method. Even when the length of a query sequence is large, our method is more than an order of magnitude faster than the suffix tree based method and the Smith-Waterman algorithm.

Methods for Integration of Documents using Hierarchical Structure based on the Formal Concept Analysis (FCA 기반 계층적 구조를 이용한 문서 통합 기법)

  • Kim, Tae-Hwan;Jeon, Ho-Cheol;Choi, Joong-Min
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.3
    • /
    • pp.63-77
    • /
    • 2011
  • The World Wide Web is a very large distributed digital information space. From its origins in 1991, the web has grown to encompass diverse information resources as personal home pasges, online digital libraries and virtual museums. Some estimates suggest that the web currently includes over 500 billion pages in the deep web. The ability to search and retrieve information from the web efficiently and effectively is an enabling technology for realizing its full potential. With powerful workstations and parallel processing technology, efficiency is not a bottleneck. In fact, some existing search tools sift through gigabyte.syze precompiled web indexes in a fraction of a second. But retrieval effectiveness is a different matter. Current search tools retrieve too many documents, of which only a small fraction are relevant to the user query. Furthermore, the most relevant documents do not nessarily appear at the top of the query output order. Also, current search tools can not retrieve the documents related with retrieved document from gigantic amount of documents. The most important problem for lots of current searching systems is to increase the quality of search. It means to provide related documents or decrease the number of unrelated documents as low as possible in the results of search. For this problem, CiteSeer proposed the ACI (Autonomous Citation Indexing) of the articles on the World Wide Web. A "citation index" indexes the links between articles that researchers make when they cite other articles. Citation indexes are very useful for a number of purposes, including literature search and analysis of the academic literature. For details of this work, references contained in academic articles are used to give credit to previous work in the literature and provide a link between the "citing" and "cited" articles. A citation index indexes the citations that an article makes, linking the articleswith the cited works. Citation indexes were originally designed mainly for information retrieval. The citation links allow navigating the literature in unique ways. Papers can be located independent of language, and words in thetitle, keywords or document. A citation index allows navigation backward in time (the list of cited articles) and forwardin time (which subsequent articles cite the current article?) But CiteSeer can not indexes the links between articles that researchers doesn't make. Because it indexes the links between articles that only researchers make when they cite other articles. Also, CiteSeer is not easy to scalability. Because CiteSeer can not indexes the links between articles that researchers doesn't make. All these problems make us orient for designing more effective search system. This paper shows a method that extracts subject and predicate per each sentence in documents. A document will be changed into the tabular form that extracted predicate checked value of possible subject and object. We make a hierarchical graph of a document using the table and then integrate graphs of documents. The graph of entire documents calculates the area of document as compared with integrated documents. We mark relation among the documents as compared with the area of documents. Also it proposes a method for structural integration of documents that retrieves documents from the graph. It makes that the user can find information easier. We compared the performance of the proposed approaches with lucene search engine using the formulas for ranking. As a result, the F.measure is about 60% and it is better as about 15%.

Index-based Searching on Timestamped Event Sequences (타임스탬프를 갖는 이벤트 시퀀스의 인덱스 기반 검색)

  • 박상현;원정임;윤지희;김상욱
    • Journal of KIISE:Databases
    • /
    • v.31 no.5
    • /
    • pp.468-478
    • /
    • 2004
  • It is essential in various application areas of data mining and bioinformatics to effectively retrieve the occurrences of interesting patterns from sequence databases. For example, let's consider a network event management system that records the types and timestamp values of events occurred in a specific network component(ex. router). The typical query to find out the temporal casual relationships among the network events is as fellows: 'Find all occurrences of CiscoDCDLinkUp that are fellowed by MLMStatusUP that are subsequently followed by TCPConnectionClose, under the constraint that the interval between the first two events is not larger than 20 seconds, and the interval between the first and third events is not larger than 40 secondsTCPConnectionClose. This paper proposes an indexing method that enables to efficiently answer such a query. Unlike the previous methods that rely on inefficient sequential scan methods or data structures not easily supported by DBMSs, the proposed method uses a multi-dimensional spatial index, which is proven to be efficient both in storage and search, to find the answers quickly without false dismissals. Given a sliding window W, the input to a multi-dimensional spatial index is a n-dimensional vector whose i-th element is the interval between the first event of W and the first occurrence of the event type Ei in W. Here, n is the number of event types that can be occurred in the system of interest. The problem of‘dimensionality curse’may happen when n is large. Therefore, we use the dimension selection or event type grouping to avoid this problem. The experimental results reveal that our proposed technique can be a few orders of magnitude faster than the sequential scan and ISO-Depth index methods.hods.

An Analysis of IT Trends Using Tweet Data (트윗 데이터를 활용한 IT 트렌드 분석)

  • Yi, Jin Baek;Lee, Choong Kwon;Cha, Kyung Jin
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.143-159
    • /
    • 2015
  • Predicting IT trends has been a long and important subject for information systems research. IT trend prediction makes it possible to acknowledge emerging eras of innovation and allocate budgets to prepare against rapidly changing technological trends. Towards the end of each year, various domestic and global organizations predict and announce IT trends for the following year. For example, Gartner Predicts 10 top IT trend during the next year, and these predictions affect IT and industry leaders and organization's basic assumptions about technology and the future of IT, but the accuracy of these reports are difficult to verify. Social media data can be useful tool to verify the accuracy. As social media services have gained in popularity, it is used in a variety of ways, from posting about personal daily life to keeping up to date with news and trends. In the recent years, rates of social media activity in Korea have reached unprecedented levels. Hundreds of millions of users now participate in online social networks and communicate with colleague and friends their opinions and thoughts. In particular, Twitter is currently the major micro blog service, it has an important function named 'tweets' which is to report their current thoughts and actions, comments on news and engage in discussions. For an analysis on IT trends, we chose Tweet data because not only it produces massive unstructured textual data in real time but also it serves as an influential channel for opinion leading on technology. Previous studies found that the tweet data provides useful information and detects the trend of society effectively, these studies also identifies that Twitter can track the issue faster than the other media, newspapers. Therefore, this study investigates how frequently the predicted IT trends for the following year announced by public organizations are mentioned on social network services like Twitter. IT trend predictions for 2013, announced near the end of 2012 from two domestic organizations, the National IT Industry Promotion Agency (NIPA) and the National Information Society Agency (NIA), were used as a basis for this research. The present study analyzes the Twitter data generated from Seoul (Korea) compared with the predictions of the two organizations to analyze the differences. Thus, Twitter data analysis requires various natural language processing techniques, including the removal of stop words, and noun extraction for processing various unrefined forms of unstructured data. To overcome these challenges, we used SAS IRS (Information Retrieval Studio) developed by SAS to capture the trend in real-time processing big stream datasets of Twitter. The system offers a framework for crawling, normalizing, analyzing, indexing and searching tweet data. As a result, we have crawled the entire Twitter sphere in Seoul area and obtained 21,589 tweets in 2013 to review how frequently the IT trend topics announced by the two organizations were mentioned by the people in Seoul. The results shows that most IT trend predicted by NIPA and NIA were all frequently mentioned in Twitter except some topics such as 'new types of security threat', 'green IT', 'next generation semiconductor' since these topics non generalized compound words so they can be mentioned in Twitter with other words. To answer whether the IT trend tweets from Korea is related to the following year's IT trends in real world, we compared Twitter's trending topics with those in Nara Market, Korea's online e-Procurement system which is a nationwide web-based procurement system, dealing with whole procurement process of all public organizations in Korea. The correlation analysis show that Tweet frequencies on IT trending topics predicted by NIPA and NIA are significantly correlated with frequencies on IT topics mentioned in project announcements by Nara market in 2012 and 2013. The main contribution of our research can be found in the following aspects: i) the IT topic predictions announced by NIPA and NIA can provide an effective guideline to IT professionals and researchers in Korea who are looking for verified IT topic trends in the following topic, ii) researchers can use Twitter to get some useful ideas to detect and predict dynamic trends of technological and social issues.