• 제목/요약/키워드: Web Search Engines

검색결과 209건 처리시간 0.028초

한국의 초등학생을 대상으로 한 성폭력 예방 교육 프로그램에 관한 통합적 고찰 (An Integrative Literature Review on Sexual Abuse Prevention Education Programs for Elementary School Students in South Korea)

  • 신혜원;이정민;강경아;김신정
    • Child Health Nursing Research
    • /
    • 제25권4호
    • /
    • pp.435-448
    • /
    • 2019
  • Purpose: The purpose of this study was to review sexual abuse prevention education program for Korean elementary school students. Methods: Whittemore and Knafl's integrative review methods were used and Gough's weight of evidence was employed as a quality appraisal tool. Articles published in Korean or English were identified through electronic search engines and scholarly web sites using three keywords: "elementary school student", "sexual abuse", and "prevention education". Peer-reviewed articles published between 2000 and 2018 were included in this review. Results: Twelve articles met the inclusion criteria and were appraised as being high-quality. Among the 12 selected studies, seven were descriptive, while five were intervention studies. Sexual abuse prevention education programs were effective in improving perceptions, knowledge, attitudes, and preventive behaviors among elementary school students. However, deficiencies were found in the variety of educational methods, utilization of experts, and standardization of the content of sexual abuse prevention education. Conclusion: We need to provide various educational methods that are appropriate for specific developmental stages, and the sexual abuse prevention content should draw upon the current sexual education program administered to this population. Furthermore, parents and trained teachers or school health teachers should be included to provide effective education programs for elementary students.

Hazelcast Vs. Ignite: Opportunities for Java Programmers

  • Maxim, Bartkov;Tetiana, Katkova;S., Kruglyk Vladyslav;G., Murtaziev Ernest;V., Kotova Olha
    • International Journal of Computer Science & Network Security
    • /
    • 제22권2호
    • /
    • pp.406-412
    • /
    • 2022
  • Storing large amounts of data has always been a big problem from the beginning of computing history. Big Data has made huge advancements in improving business processes by finding the customers' needs using prediction models based on web and social media search. The main purpose of big data stream processing frameworks is to allow programmers to directly query the continuous stream without dealing with the lower-level mechanisms. In other words, programmers write the code to process streams using these runtime libraries (also called Stream Processing Engines). This is achieved by taking large volumes of data and analyzing them using Big Data frameworks. Streaming platforms are an emerging technology that deals with continuous streams of data. There are several streaming platforms of Big Data freely available on the Internet. However, selecting the most appropriate one is not easy for programmers. In this paper, we present a detailed description of two of the state-of-the-art and most popular streaming frameworks: Apache Ignite and Hazelcast. In addition, the performance of these frameworks is compared using selected attributes. Different types of databases are used in common to store the data. To process the data in real-time continuously, data streaming technologies are developed. With the development of today's large-scale distributed applications handling tons of data, these databases are not viable. Consequently, Big Data is introduced to store, process, and analyze data at a fast speed and also to deal with big users and data growth day by day.

중환자실 환자의 수면에 영향을 미치는 요인: 체계적 고찰 (Influencing factors for Sleep Disturbance in the Intensive Care Unit Patients: A Systematic Review)

  • 조영신;정선애
    • 중환자간호학회지
    • /
    • 제16권2호
    • /
    • pp.1-14
    • /
    • 2023
  • Purpose : Sleep disturbances in patients in the intensive care unit (ICU) are related to health problems after discharge. Therefore, active prevention and management are required. Hence, identification of the factors that affect sleep in patients who are critically ill is necessary. Methods : The PubMed, Cochrane Library, CINAHL, EMBASE, and Web of Science databases were searched. Selection criteria were observational and experimental studies that assessed sleep as an outcome, included adult patients admitted to the ICU, and published between November 2015 and April 2022. Results : A total of 21,136 articles were identified through search engines and manual searches, and 42 articles were selected. From these, 22 influencing factors and 11 interventions were identified. Individual factors included disease severity, age, pain, delirium, comorbidities, alcohol consumption, sex, sleep disturbance before hospitalization, chronic obstructive pulmonary disease (COPD), cardiovascular disease, and high diastolic blood pressure (DBP), low hemoglobin (Hb), and low respiratory rate (RR). Environmental factors included light level, noise level, and temperature. Furthermore, treatment-related factors included use of sedatives, melatonin administration, sleep management guidelines, ventilator application, nursing treatment, and length of ICU stay. Regarding sleep interventions, massage, eye mask and earplugs, quiet time and multicomponent protocols, aromatherapy, acupressure, sounds of the sea, adaptive intervention, circulation lighting, and single occupation in a room were identified. Conclusion : Based on these results, we propose the development and application of various interventions to improve sleep quality in patients who are critically ill.

표면 웹기반 공개정보 수집을 위한 워크플로우 확장 연구 (A Study on the Expansion of Workflow for the Collection of Surface Web-based OSINT(Open Source Intelligence))

  • 이수경;최은정;김지연;이인수;이승훈;김명주
    • 디지털융복합연구
    • /
    • 제20권4호
    • /
    • pp.367-376
    • /
    • 2022
  • 전통적인 형사 사건에서 조사 대상에 관한 정보는 국가의 합법적 조직이 보유하고 있는 개인정보만이 제공되기 때문에 정보 수집에 한계가 있다. 일반 검색엔진으로 검색이 가능한 SNS와 포털사이트를 포함하는 표면 웹 기반 공개정보(OSINT)는 범죄수사에 사용할 수 있는 의미 있는 프로파일링에 활용할 수 있다. 한국형 공개정보 워크플로우를 사용하면 공개정보 기반의 효과적인 프로파일링이 가능하지만 "개인"의 경우에는 "성명"으로 시작되기 때문에 수집할 수 있는 공개정보가 제한적이고 동명이인의 정보가 수집되는 등의 신뢰성의 한계가 있다. 본 논문에서는 이러한 한계를 극복하기 위해 개인과 연관된 정보 즉, 등가정보를 정의하고 이를 기반으로 효율적이고 정확한 정보를 수집할 수 있도록 한다. 따라서, 공개정보에서 특정인과 연관된 정보 즉, 등가정보를 추출할 수 있는 개선된 워크플로우를 제시한다. 이때 인물의 인지도에 따라 서로 다른 워크플로우를 제시한다. 이를 통해 인물(개인)의 효과적인 프로파일링이 가능하여 수사 정보 수집의 신뢰도를 높인다. 본 연구를 통해 향후에는 해당 워크플로우를 인공지능 기술을 이용하여 수집된 정보의 분석과정을 자동화할 수 있는 시스템을 개발함으로써 범죄 수사에 있어서 공개 정보 활용을 위한 기틀을 마련하고 수사 방식 다양화에 기여할 수 있을 것이다.

문서필터링을 위한 질의어 확장과 가중치 부여 기법 (Query Expansion and Term Weighting Method for Document Filtering)

  • 신승은;강유환;오효정;장명길;박상규;이재성;서영훈
    • 정보처리학회논문지B
    • /
    • 제10B권7호
    • /
    • pp.743-750
    • /
    • 2003
  • 본 논문에서는 문서 필터링을 위한 질의어 확장과 가중치 부여 기법을 제안한다. 문서 필터링은 웹 검색 엔진들에 대한 검색 결과의 정확률 향상을 목적으로 한다. 문서 필터링을 위한 질의어 확장은 개념망, 백과사전, 유사도 상위 10% 문서를 이용하며, 각각의 확장 질의어에 가중치를 부쳐하여 질의어와 문서들간의 유사도를 계산한다. 첫 번째 단계에서 개념망과 백과사전을 이용하여 초기 질의어에 대한 1차 확장 질의어를 생성하고, 1차 확장 질의엉 가중치를 부여하여 질의어와 문서들간의 유사도를 계산한다. 다음 단계에서는 높은 유사도를 갖는 상위 10% 문서들을 이용하여 2타 확장 질의어를 생성하고, 2차 확장 질의어에 가중치를 부쳐하여 질의어와 문서들간의 유사도를 계산한다. 다음으로 1차 유사도와 2차 유사도를 결합하여 문서들을 재순위화하고, 임계치보다 낮은 유사도를 갖는 문서들을 필터링함으로써 웹 검색 엔진들의 검색 결과 정확률을 향상시킨다. 실험에서 이러한 문서 필터링을 위한 질의어 확장과 가중치 부여 기법은 정확률-재현율과 F-measure를 이용하여 성능 평가를 할 때 정보 검색 효율성에서 주목할 만한 성능 향상을 보였다.

대상포진후신경통에 관한 인터넷 사이트 평가 (Assessment of the Quality of Postherpetic Neuralgia Related Korean Internet Sites)

  • 이재학;박상규;이두익;정종권;임현경;차영덕
    • The Korean Journal of Pain
    • /
    • 제22권2호
    • /
    • pp.141-145
    • /
    • 2009
  • Background: There is no assessment of internet sites that carry information on chronic pain disease. So we assessed the quality of information about postherpetic neuralgia available on Korean internet sites. Methods: The keywords 'postherpetic neuralgia', 'herpes zoster, neuropathic pain', 'herpes zoster, pain', 'herpes zoster' were searched in Korean on four search engines in Korea between the 1st to the 15th of May, 2009. We evaluated the outcome on two factors; the aspect of the contents which is subdivided into two categories, the content and authorship, and the technical aspect including web design, and efficiency. Results: A total of 26 internet sites were found. Among these sites, 6 (23%) informed by anesthesiologist. The average score of the 26 internet sites was only $37.4{\pm}20.1$ out of a total of 100. A mean score of the contents was $13.3{\pm}8.3$ out of 40 points, the authorship was $10.0{\pm}6.7$ out of 20 points, the design was $9.2{\pm}5.3$ out of 20 points, the efficiency was $6.8{\pm}4.3$ out of 20 points. When comparing the score between anesthesiologist and non-anesthesiologist, the contents was $18.7{\pm}7.4$ vs. $11.7{\pm}7.9$, the authorship was $13.4{\pm}4.7$ vs. $9.0{\pm}6.8$, the design was $12.5{\pm}4.2$ vs. $8.3{\pm}5.2$ and the efficiency was $6.8{\pm}4.5$ vs. $4.3{\pm}4.0$ (P < 0.05). Conclusions: There is a need for more accurate information about postherpetic neuralgia on the Korean internet by anesthesiologists.

두 영상검색 시스템의 협력적 이용을 통한 재순위화 (Re-ranking the Results from Two Image Retrieval System in Cooperative Manner)

  • 황중원;김현우;김준모
    • 한국지능시스템학회논문지
    • /
    • 제24권1호
    • /
    • pp.7-15
    • /
    • 2014
  • 영상검색은 컴퓨터 비전과 데이터 마이닝 분야의 주요한 주제 중 하나이다. 현재 상용화된 영상 검색 시스템이 놀라운 성능을 보여주고 있음에도 불구하고, 폭발적으로 증가하는 웹상의 영상정보를 효율적으로 검색하기 위하여, 영상 검색 기술의 성능향상이 꾸준히 요구되고 있다. 재순위화는, 이런 요구를 만족시키기 위한 방법 중 하나로서, 영상 검색 시스템이 도출한 초기 결과를, 독자적인 알고리즘을 통해 연관도를 한 번 더 계산하고, 그에 따라 검색 결과의 성능을 향상시키는 방법이다. 지금까지 제안된 많은 재순위화 알고리즘이, 영상 검색 시스템이 영상의 시각적 정보를 사용하지 않는다는 가정하에 개발되어 왔지만, 현재 Google 과 Baidu 등 많은 상업적 영상 검색 시스템이 영상의 시각적 패턴에 대한 정보를 사용하기 시작함에 따라 이러한 가정이 어긋나고 있다. 또한, 영상 검색 시스템들이 독자적인 환경에서 개발되었기 때문에, 다수의 영상 검색 시스템의 협력으로 성능을 향상시킬 수 있을 것이라 예측되는 상황에도 불구하고, 이에 대한 연구가 부족했다. 본 논문에서는 이 두 가지 문제를 해결하기 위하여, 두 영상 검색 시스템이 사용하는 시각적 특징을 바탕으로, 두 시스템의 상보적인 정보를 이용하는 방법에 대하여 제안한다. 성능에 대한 평가는 가상 모형에서 이루어졌으며, 제안된 방법을 통하여 성능향상이 가능하다는 것을 보여준다.

A Study on the Meaning and Strategy of Keyword Advertising Marketing

  • Park, Nam Goo
    • 유통과학연구
    • /
    • 제8권3호
    • /
    • pp.49-56
    • /
    • 2010
  • At the initial stage of Internet advertising, banner advertising came into fashion. As the Internet developed into a central part of daily lives and the competition in the on-line advertising market was getting fierce, there was not enough space for banner advertising, which rushed to portal sites only. All these factors was responsible for an upsurge in advertising prices. Consequently, the high-cost and low-efficiency problems with banner advertising were raised, which led to an emergence of keyword advertising as a new type of Internet advertising to replace its predecessor. In the beginning of 2000s, when Internet advertising came to be activated, display advertisement including banner advertising dominated the Net. However, display advertising showed signs of gradual decline, and registered minus growth in the year 2009, whereas keyword advertising showed rapid growth and started to outdo display advertising as of the year 2005. Keyword advertising refers to the advertising technique that exposes relevant advertisements on the top of research sites when one searches for a keyword. Instead of exposing advertisements to unspecified individuals like banner advertising, keyword advertising, or targeted advertising technique, shows advertisements only when customers search for a desired keyword so that only highly prospective customers are given a chance to see them. In this context, it is also referred to as search advertising. It is regarded as more aggressive advertising with a high hit rate than previous advertising in that, instead of the seller discovering customers and running an advertisement for them like TV, radios or banner advertising, it exposes advertisements to visiting customers. Keyword advertising makes it possible for a company to seek publicity on line simply by making use of a single word and to achieve a maximum of efficiency at a minimum cost. The strong point of keyword advertising is that customers are allowed to directly contact the products in question through its more efficient advertising when compared to the advertisements of mass media such as TV and radio, etc. The weak point of keyword advertising is that a company should have its advertisement registered on each and every portal site and finds it hard to exercise substantial supervision over its advertisement, there being a possibility of its advertising expenses exceeding its profits. Keyword advertising severs as the most appropriate methods of advertising for the sales and publicity of small and medium enterprises which are in need of a maximum of advertising effect at a low advertising cost. At present, keyword advertising is divided into CPC advertising and CPM advertising. The former is known as the most efficient technique, which is also referred to as advertising based on the meter rate system; A company is supposed to pay for the number of clicks on a searched keyword which users have searched. This is representatively adopted by Overture, Google's Adwords, Naver's Clickchoice, and Daum's Clicks, etc. CPM advertising is dependent upon the flat rate payment system, making a company pay for its advertisement on the basis of the number of exposure, not on the basis of the number of clicks. This method fixes a price for advertisement on the basis of 1,000-time exposure, and is mainly adopted by Naver's Timechoice, Daum's Speciallink, and Nate's Speedup, etc, At present, the CPC method is most frequently adopted. The weak point of the CPC method is that advertising cost can rise through constant clicks from the same IP. If a company makes good use of strategies for maximizing the strong points of keyword advertising and complementing its weak points, it is highly likely to turn its visitors into prospective customers. Accordingly, an advertiser should make an analysis of customers' behavior and approach them in a variety of ways, trying hard to find out what they want. With this in mind, her or she has to put multiple keywords into use when running for ads. When he or she first runs an ad, he or she should first give priority to which keyword to select. The advertiser should consider how many individuals using a search engine will click the keyword in question and how much money he or she has to pay for the advertisement. As the popular keywords that the users of search engines are frequently using are expensive in terms of a unit cost per click, the advertisers without much money for advertising at the initial phrase should pay attention to detailed keywords suitable to their budget. Detailed keywords are also referred to as peripheral keywords or extension keywords, which can be called a combination of major keywords. Most keywords are in the form of texts. The biggest strong point of text-based advertising is that it looks like search results, causing little antipathy to it. But it fails to attract much attention because of the fact that most keyword advertising is in the form of texts. Image-embedded advertising is easy to notice due to images, but it is exposed on the lower part of a web page and regarded as an advertisement, which leads to a low click through rate. However, its strong point is that its prices are lower than those of text-based advertising. If a company owns a logo or a product that is easy enough for people to recognize, the company is well advised to make good use of image-embedded advertising so as to attract Internet users' attention. Advertisers should make an analysis of their logos and examine customers' responses based on the events of sites in question and the composition of products as a vehicle for monitoring their behavior in detail. Besides, keyword advertising allows them to analyze the advertising effects of exposed keywords through the analysis of logos. The logo analysis refers to a close analysis of the current situation of a site by making an analysis of information about visitors on the basis of the analysis of the number of visitors and page view, and that of cookie values. It is in the log files generated through each Web server that a user's IP, used pages, the time when he or she uses it, and cookie values are stored. The log files contain a huge amount of data. As it is almost impossible to make a direct analysis of these log files, one is supposed to make an analysis of them by using solutions for a log analysis. The generic information that can be extracted from tools for each logo analysis includes the number of viewing the total pages, the number of average page view per day, the number of basic page view, the number of page view per visit, the total number of hits, the number of average hits per day, the number of hits per visit, the number of visits, the number of average visits per day, the net number of visitors, average visitors per day, one-time visitors, visitors who have come more than twice, and average using hours, etc. These sites are deemed to be useful for utilizing data for the analysis of the situation and current status of rival companies as well as benchmarking. As keyword advertising exposes advertisements exclusively on search-result pages, competition among advertisers attempting to preoccupy popular keywords is very fierce. Some portal sites keep on giving priority to the existing advertisers, whereas others provide chances to purchase keywords in question to all the advertisers after the advertising contract is over. If an advertiser tries to rely on keywords sensitive to seasons and timeliness in case of sites providing priority to the established advertisers, he or she may as well make a purchase of a vacant place for advertising lest he or she should miss appropriate timing for advertising. However, Naver doesn't provide priority to the existing advertisers as far as all the keyword advertisements are concerned. In this case, one can preoccupy keywords if he or she enters into a contract after confirming the contract period for advertising. This study is designed to take a look at marketing for keyword advertising and to present effective strategies for keyword advertising marketing. At present, the Korean CPC advertising market is virtually monopolized by Overture. Its strong points are that Overture is based on the CPC charging model and that advertisements are registered on the top of the most representative portal sites in Korea. These advantages serve as the most appropriate medium for small and medium enterprises to use. However, the CPC method of Overture has its weak points, too. That is, the CPC method is not the only perfect advertising model among the search advertisements in the on-line market. So it is absolutely necessary that small and medium enterprises including independent shopping malls should complement the weaknesses of the CPC method and make good use of strategies for maximizing its strengths so as to increase their sales and to create a point of contact with customers.

  • PDF

키워드 자동 생성에 대한 새로운 접근법: 역 벡터공간모델을 이용한 키워드 할당 방법 (A New Approach to Automatic Keyword Generation Using Inverse Vector Space Model)

  • 조원진;노상규;윤지영;박진수
    • Asia pacific journal of information systems
    • /
    • 제21권1호
    • /
    • pp.103-122
    • /
    • 2011
  • Recently, numerous documents have been made available electronically. Internet search engines and digital libraries commonly return query results containing hundreds or even thousands of documents. In this situation, it is virtually impossible for users to examine complete documents to determine whether they might be useful for them. For this reason, some on-line documents are accompanied by a list of keywords specified by the authors in an effort to guide the users by facilitating the filtering process. In this way, a set of keywords is often considered a condensed version of the whole document and therefore plays an important role for document retrieval, Web page retrieval, document clustering, summarization, text mining, and so on. Since many academic journals ask the authors to provide a list of five or six keywords on the first page of an article, keywords are most familiar in the context of journal articles. However, many other types of documents could not benefit from the use of keywords, including Web pages, email messages, news reports, magazine articles, and business papers. Although the potential benefit is large, the implementation itself is the obstacle; manually assigning keywords to all documents is a daunting task, or even impractical in that it is extremely tedious and time-consuming requiring a certain level of domain knowledge. Therefore, it is highly desirable to automate the keyword generation process. There are mainly two approaches to achieving this aim: keyword assignment approach and keyword extraction approach. Both approaches use machine learning methods and require, for training purposes, a set of documents with keywords already attached. In the former approach, there is a given set of vocabulary, and the aim is to match them to the texts. In other words, the keywords assignment approach seeks to select the words from a controlled vocabulary that best describes a document. Although this approach is domain dependent and is not easy to transfer and expand, it can generate implicit keywords that do not appear in a document. On the other hand, in the latter approach, the aim is to extract keywords with respect to their relevance in the text without prior vocabulary. In this approach, automatic keyword generation is treated as a classification task, and keywords are commonly extracted based on supervised learning techniques. Thus, keyword extraction algorithms classify candidate keywords in a document into positive or negative examples. Several systems such as Extractor and Kea were developed using keyword extraction approach. Most indicative words in a document are selected as keywords for that document and as a result, keywords extraction is limited to terms that appear in the document. Therefore, keywords extraction cannot generate implicit keywords that are not included in a document. According to the experiment results of Turney, about 64% to 90% of keywords assigned by the authors can be found in the full text of an article. Inversely, it also means that 10% to 36% of the keywords assigned by the authors do not appear in the article, which cannot be generated through keyword extraction algorithms. Our preliminary experiment result also shows that 37% of keywords assigned by the authors are not included in the full text. This is the reason why we have decided to adopt the keyword assignment approach. In this paper, we propose a new approach for automatic keyword assignment namely IVSM(Inverse Vector Space Model). The model is based on a vector space model. which is a conventional information retrieval model that represents documents and queries by vectors in a multidimensional space. IVSM generates an appropriate keyword set for a specific document by measuring the distance between the document and the keyword sets. The keyword assignment process of IVSM is as follows: (1) calculating the vector length of each keyword set based on each keyword weight; (2) preprocessing and parsing a target document that does not have keywords; (3) calculating the vector length of the target document based on the term frequency; (4) measuring the cosine similarity between each keyword set and the target document; and (5) generating keywords that have high similarity scores. Two keyword generation systems were implemented applying IVSM: IVSM system for Web-based community service and stand-alone IVSM system. Firstly, the IVSM system is implemented in a community service for sharing knowledge and opinions on current trends such as fashion, movies, social problems, and health information. The stand-alone IVSM system is dedicated to generating keywords for academic papers, and, indeed, it has been tested through a number of academic papers including those published by the Korean Association of Shipping and Logistics, the Korea Research Academy of Distribution Information, the Korea Logistics Society, the Korea Logistics Research Association, and the Korea Port Economic Association. We measured the performance of IVSM by the number of matches between the IVSM-generated keywords and the author-assigned keywords. According to our experiment, the precisions of IVSM applied to Web-based community service and academic journals were 0.75 and 0.71, respectively. The performance of both systems is much better than that of baseline systems that generate keywords based on simple probability. Also, IVSM shows comparable performance to Extractor that is a representative system of keyword extraction approach developed by Turney. As electronic documents increase, we expect that IVSM proposed in this paper can be applied to many electronic documents in Web-based community and digital library.