• Title/Summary/Keyword: content retrieval

Search Result 800, Processing Time 0.025 seconds

The Formation Process of Nature-Study in U.S. and Its Implication for Science Education (미국 Nature-Study 형성 과정과 과학교육에의 시사점)

  • Park, Jongseok;Park, Sangmin
    • Journal of the Korean Chemical Society
    • /
    • v.58 no.1
    • /
    • pp.118-125
    • /
    • 2014
  • This study purposes to historical approach the formation process of Nature-Study, and to re-evaluate its definition and direction at present. The idea of Nature-Study originated from Campanella, Ratke and Comenius, who emphasized real things. The idea developed through Object Lessons of Sheldon, the Natural History of Agassiz, and Progressivism of Parker. They acted as the main contributors who evolved the idea of Nature-Study and its core fields that involve: 'studying with real things' in Object Lessons which brought the methodical aspects to the idea, 'studying with nature' from Natural History that enhanced the content characteristics and 'learner-centered education' from Progressivism, which impacted the philosophical aspects. Straight (a fellow student of Agassiz) was a teacher for Sheldon Oswego normal school and Parker's Cook County normal school, who synthesized the fields together and paved the way for the formation of Nature-Study. Jackman of Cook Country normal school established Nature-Study as a school curriculum and Bailey and Comstock of Cornell University formed the American Nature-Study Society and as a result, Nature-Study started to gain popularity. However, many educators increasingly rejected Nature-Study as a unifying topic, and preferred the use of textbooks rather than firsthand experiences. This hindered the nature-study movement and it declined since the 1920s. But today, the Nature-study idea can play a huge role in developing science education, inclusive education centered nature, self-initiated retrieval, sympathy with nature and character building of students.

A Study on the Recognition of Users and Librarians of Obstructive Factors in Online Reference Services (온라인참고서비스의 장애요인에 대한 이용자 및 사서의 인식조사 연구)

  • Noh, Younghee;Park, Hyejin;Shin, Youngji
    • Journal of the Korean Society for Library and Information Science
    • /
    • v.50 no.1
    • /
    • pp.133-159
    • /
    • 2016
  • The purpose of this study is to analyze related studies and domestic/international online reference cases, extract obstructive factors present in online reference services, and reveal whether or not there are differences in perception between the university librarian and the users. The results with respect to the failure of the resources revealed that while the user considers the quantitative / qualitative shortage of content as the greatest obstacle in the online reference service, librarians see the lack of human resources (Specialist Librarian / trained staff) in this light. Users think this is the least of the problems. In addition, other obstacles that are the most highly evaluated by librarians are, in order, the limitation of service because of copyright issues, the difficulty of information retrieval and complexity of methods of use, and a general lack of information in the reference services menu and missing information in the main menu. For the users the other most important obstacles were similar with the limitation of service because of copyright issues being highest, followed by the difficulty of access because of the confusion over service names, and the general lack of information in the reference services menu and missing information in the main menu.

A Study of Intelligent Recommendation System based on Naive Bayes Text Classification and Collaborative Filtering (나이브베이즈 분류모델과 협업필터링 기반 지능형 학술논문 추천시스템 연구)

  • Lee, Sang-Gi;Lee, Byeong-Seop;Bak, Byeong-Yong;Hwang, Hye-Kyong
    • Journal of Information Management
    • /
    • v.41 no.4
    • /
    • pp.227-249
    • /
    • 2010
  • Scholarly information has increased tremendously according to the development of IT, especially the Internet. However, simultaneously, people have to spend more time and exert more effort because of information overload. There have been many research efforts in the field of expert systems, data mining, and information retrieval, concerning a system that recommends user-expected information items through presumption. Recently, the hybrid system combining a content-based recommendation system and collaborative filtering or combining recommendation systems in other domains has been developed. In this paper we resolved the problem of the current recommendation system and suggested a new system combining collaborative filtering and Naive Bayes Classification. In this way, we resolved the over-specialization problem through collaborative filtering and lack of assessment information or recommendation of new contents through Naive Bayes Classification. For verification, we applied the new model in NDSL's paper service of KISTI, especially papers from journals about Sitology and Electronics, and witnessed high satisfaction from 4 experimental participants.

A Study on the National Teacher Recruiting Examination for School Librarian Teacher: Focusing on the School Library Practice Area (사서교사 임용시험 출제경향 고찰 - 학교도서관 실무영역을 중심으로 -)

  • Kyungkuk Noh;Jeonghoon Lim
    • Journal of Korean Library and Information Science Society
    • /
    • v.54 no.4
    • /
    • pp.85-104
    • /
    • 2023
  • The purpose of this study is to analyze the examination questions used in the librarian teacher recruitment exam, including the domains, content, and evaluation factors, and to propose improvements for the recruitment exam. To achieve this, examination questions for librarian teacher recruitment exams since 2002, provided by the Korea Institute for Curriculum and Evaluation, were collected and analyzed the frequency of appearances by section. The analysis revealed that, 106 questions (21.95%) on school library administration, 63 questions (13.04%) on classification and information retrieval 59 questions (12.22%) on library computerization, 58 questions (12.01%) on reading education, 56 questions (11.59%) cataloging and information service, and 18 questions (3.73%) on information media were examined. Next, analyzed the frequency of appearances in the last 10 years (2014-2023) by dividing the examination areas into specialty of librarian and school library practice, and found that there were a total of 149 questions (66.22%) related to specialty of librarian and 76 questions (33.78%) related to school library practice. Based on these findings, recommendations have been made for update assessment areas and factors, expanding the field of information media, and suggested the need for a stable and continuous teacher recruitment policy.

Digital Nudge in an Online Review Environment: How Uploading Pictures First Affects the Quality of Reviews (온라인 리뷰 환경에서의 디지털 넛지: 사진을 먼저 업로드 하는 행동이 리뷰의 품질에 미치는 영향 )

  • Jaemin Lee;Taeyoung Kim;HoGeun Lee
    • Information Systems Review
    • /
    • v.25 no.1
    • /
    • pp.1-26
    • /
    • 2023
  • Consumers tend to trust information provided by other consumers more than information provided by sellers. Therefore, while inducing consumers to write high-quality reviews is a very important task for companies, it is not easy to produce such high-quality reviews. Based on previous research on review writing and memory recall, we decided to develop a way to use digital nudge to help consumers naturally write high-quality reviews. Specifically, we designed an experiment to verify the effect of uploading a photo during the online review process on the quality of review of the review writer. We then recruited subjects and then divided them into groups that upload photos first and groups that do not. A task was assigned to each subject to write positive and negative reviews. As a result, it was confirmed that the behavior of uploading a photo first increases the review length. In addition, it was confirmed that when online users who upload photos first have extremely negative satisfaction with the product, the extent of two-sidedness of the review content increases.

Restoring Omitted Sentence Constituents in Encyclopedia Documents Using Structural SVM (Structural SVM을 이용한 백과사전 문서 내 생략 문장성분 복원)

  • Hwang, Min-Kook;Kim, Youngtae;Ra, Dongyul;Lim, Soojong;Kim, Hyunki
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.2
    • /
    • pp.131-150
    • /
    • 2015
  • Omission of noun phrases for obligatory cases is a common phenomenon in sentences of Korean and Japanese, which is not observed in English. When an argument of a predicate can be filled with a noun phrase co-referential with the title, the argument is more easily omitted in Encyclopedia texts. The omitted noun phrase is called a zero anaphor or zero pronoun. Encyclopedias like Wikipedia are major source for information extraction by intelligent application systems such as information retrieval and question answering systems. However, omission of noun phrases makes the quality of information extraction poor. This paper deals with the problem of developing a system that can restore omitted noun phrases in encyclopedia documents. The problem that our system deals with is almost similar to zero anaphora resolution which is one of the important problems in natural language processing. A noun phrase existing in the text that can be used for restoration is called an antecedent. An antecedent must be co-referential with the zero anaphor. While the candidates for the antecedent are only noun phrases in the same text in case of zero anaphora resolution, the title is also a candidate in our problem. In our system, the first stage is in charge of detecting the zero anaphor. In the second stage, antecedent search is carried out by considering the candidates. If antecedent search fails, an attempt made, in the third stage, to use the title as the antecedent. The main characteristic of our system is to make use of a structural SVM for finding the antecedent. The noun phrases in the text that appear before the position of zero anaphor comprise the search space. The main technique used in the methods proposed in previous research works is to perform binary classification for all the noun phrases in the search space. The noun phrase classified to be an antecedent with highest confidence is selected as the antecedent. However, we propose in this paper that antecedent search is viewed as the problem of assigning the antecedent indicator labels to a sequence of noun phrases. In other words, sequence labeling is employed in antecedent search in the text. We are the first to suggest this idea. To perform sequence labeling, we suggest to use a structural SVM which receives a sequence of noun phrases as input and returns the sequence of labels as output. An output label takes one of two values: one indicating that the corresponding noun phrase is the antecedent and the other indicating that it is not. The structural SVM we used is based on the modified Pegasos algorithm which exploits a subgradient descent methodology used for optimization problems. To train and test our system we selected a set of Wikipedia texts and constructed the annotated corpus in which gold-standard answers are provided such as zero anaphors and their possible antecedents. Training examples are prepared using the annotated corpus and used to train the SVMs and test the system. For zero anaphor detection, sentences are parsed by a syntactic analyzer and subject or object cases omitted are identified. Thus performance of our system is dependent on that of the syntactic analyzer, which is a limitation of our system. When an antecedent is not found in the text, our system tries to use the title to restore the zero anaphor. This is based on binary classification using the regular SVM. The experiment showed that our system's performance is F1 = 68.58%. This means that state-of-the-art system can be developed with our technique. It is expected that future work that enables the system to utilize semantic information can lead to a significant performance improvement.

A survey on the utilization practice and satisfaction of users of food and nutrition information (정보이용자의 식품영양정보 이용 실태와 만족도)

  • Kim, Inhye;Park, Min-Seo;Bae, Hyun-Joo
    • Journal of Nutrition and Health
    • /
    • v.54 no.4
    • /
    • pp.398-411
    • /
    • 2021
  • Purpose: The objective of this study was to investigate food and nutrition information utilization practices of adults aged between 20 and 30 years to provide the basic data for developing customized content. Methods: Statistical analyses were performed using the SPSS program (ver. 24.0) for the 𝛘2-test, t-test, one-way analysis of variance, and Duncan's multiple range test. Results: Of the 570 subjects surveyed, 45.4% were men, 54.6% were women, 66.3% were in their 20s, 33.7% were in their 30s, 41.4% were single-person households, and 58.6% lived with their families. On average, 14.2% of televisions (TVs), 26.0% of personal computers (PCs), and 63.7% of smartphones were used for more than three hours per day. 30.9% of respondents searched for food and nutrition information more than once a week. 70.0% of the respondents had then applied the information in real life and 54.7% of the respondents said they would share information with others. Information retrieval rate was in the order of 'restaurant (64.8%)', 'diet (57.5%)', and 'food recipes (55.7%)'. Overall satisfaction with food and nutrition information averaged 3.33 on a five-point scale. Satisfaction score was in the order of 'enough description and easy to understand (3.43)', 'matching title and content (3.35)', and 'providing new and novel information (3.22)'. Satisfaction scores were significantly higher in the group that searched for information (p < 0.001), the group that used the retrieved information in real life (p < 0.001), and the group that conveyed this information to others (p < 0.001). Conclusion: To improve information user satisfaction, it is necessary to provide customized information that fits the characteristics of information users. For this purpose, it is necessary to continuously conduct surveys and satisfaction evaluations for each target group.

Video Scene Detection using Shot Clustering based on Visual Features (시각적 특징을 기반한 샷 클러스터링을 통한 비디오 씬 탐지 기법)

  • Shin, Dong-Wook;Kim, Tae-Hwan;Choi, Joong-Min
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.47-60
    • /
    • 2012
  • Video data comes in the form of the unstructured and the complex structure. As the importance of efficient management and retrieval for video data increases, studies on the video parsing based on the visual features contained in the video contents are researched to reconstruct video data as the meaningful structure. The early studies on video parsing are focused on splitting video data into shots, but detecting the shot boundary defined with the physical boundary does not cosider the semantic association of video data. Recently, studies on structuralizing video shots having the semantic association to the video scene defined with the semantic boundary by utilizing clustering methods are actively progressed. Previous studies on detecting the video scene try to detect video scenes by utilizing clustering algorithms based on the similarity measure between video shots mainly depended on color features. However, the correct identification of a video shot or scene and the detection of the gradual transitions such as dissolve, fade and wipe are difficult because color features of video data contain a noise and are abruptly changed due to the intervention of an unexpected object. In this paper, to solve these problems, we propose the Scene Detector by using Color histogram, corner Edge and Object color histogram (SDCEO) that clusters similar shots organizing same event based on visual features including the color histogram, the corner edge and the object color histogram to detect video scenes. The SDCEO is worthy of notice in a sense that it uses the edge feature with the color feature, and as a result, it effectively detects the gradual transitions as well as the abrupt transitions. The SDCEO consists of the Shot Bound Identifier and the Video Scene Detector. The Shot Bound Identifier is comprised of the Color Histogram Analysis step and the Corner Edge Analysis step. In the Color Histogram Analysis step, SDCEO uses the color histogram feature to organizing shot boundaries. The color histogram, recording the percentage of each quantized color among all pixels in a frame, are chosen for their good performance, as also reported in other work of content-based image and video analysis. To organize shot boundaries, SDCEO joins associated sequential frames into shot boundaries by measuring the similarity of the color histogram between frames. In the Corner Edge Analysis step, SDCEO identifies the final shot boundaries by using the corner edge feature. SDCEO detect associated shot boundaries comparing the corner edge feature between the last frame of previous shot boundary and the first frame of next shot boundary. In the Key-frame Extraction step, SDCEO compares each frame with all frames and measures the similarity by using histogram euclidean distance, and then select the frame the most similar with all frames contained in same shot boundary as the key-frame. Video Scene Detector clusters associated shots organizing same event by utilizing the hierarchical agglomerative clustering method based on the visual features including the color histogram and the object color histogram. After detecting video scenes, SDCEO organizes final video scene by repetitive clustering until the simiarity distance between shot boundaries less than the threshold h. In this paper, we construct the prototype of SDCEO and experiments are carried out with the baseline data that are manually constructed, and the experimental results that the precision of shot boundary detection is 93.3% and the precision of video scene detection is 83.3% are satisfactory.

Knowledge Extraction Methodology and Framework from Wikipedia Articles for Construction of Knowledge-Base (지식베이스 구축을 위한 한국어 위키피디아의 학습 기반 지식추출 방법론 및 플랫폼 연구)

  • Kim, JaeHun;Lee, Myungjin
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.43-61
    • /
    • 2019
  • Development of technologies in artificial intelligence has been rapidly increasing with the Fourth Industrial Revolution, and researches related to AI have been actively conducted in a variety of fields such as autonomous vehicles, natural language processing, and robotics. These researches have been focused on solving cognitive problems such as learning and problem solving related to human intelligence from the 1950s. The field of artificial intelligence has achieved more technological advance than ever, due to recent interest in technology and research on various algorithms. The knowledge-based system is a sub-domain of artificial intelligence, and it aims to enable artificial intelligence agents to make decisions by using machine-readable and processible knowledge constructed from complex and informal human knowledge and rules in various fields. A knowledge base is used to optimize information collection, organization, and retrieval, and recently it is used with statistical artificial intelligence such as machine learning. Recently, the purpose of the knowledge base is to express, publish, and share knowledge on the web by describing and connecting web resources such as pages and data. These knowledge bases are used for intelligent processing in various fields of artificial intelligence such as question answering system of the smart speaker. However, building a useful knowledge base is a time-consuming task and still requires a lot of effort of the experts. In recent years, many kinds of research and technologies of knowledge based artificial intelligence use DBpedia that is one of the biggest knowledge base aiming to extract structured content from the various information of Wikipedia. DBpedia contains various information extracted from Wikipedia such as a title, categories, and links, but the most useful knowledge is from infobox of Wikipedia that presents a summary of some unifying aspect created by users. These knowledge are created by the mapping rule between infobox structures and DBpedia ontology schema defined in DBpedia Extraction Framework. In this way, DBpedia can expect high reliability in terms of accuracy of knowledge by using the method of generating knowledge from semi-structured infobox data created by users. However, since only about 50% of all wiki pages contain infobox in Korean Wikipedia, DBpedia has limitations in term of knowledge scalability. This paper proposes a method to extract knowledge from text documents according to the ontology schema using machine learning. In order to demonstrate the appropriateness of this method, we explain a knowledge extraction model according to the DBpedia ontology schema by learning Wikipedia infoboxes. Our knowledge extraction model consists of three steps, document classification as ontology classes, proper sentence classification to extract triples, and value selection and transformation into RDF triple structure. The structure of Wikipedia infobox are defined as infobox templates that provide standardized information across related articles, and DBpedia ontology schema can be mapped these infobox templates. Based on these mapping relations, we classify the input document according to infobox categories which means ontology classes. After determining the classification of the input document, we classify the appropriate sentence according to attributes belonging to the classification. Finally, we extract knowledge from sentences that are classified as appropriate, and we convert knowledge into a form of triples. In order to train models, we generated training data set from Wikipedia dump using a method to add BIO tags to sentences, so we trained about 200 classes and about 2,500 relations for extracting knowledge. Furthermore, we evaluated comparative experiments of CRF and Bi-LSTM-CRF for the knowledge extraction process. Through this proposed process, it is possible to utilize structured knowledge by extracting knowledge according to the ontology schema from text documents. In addition, this methodology can significantly reduce the effort of the experts to construct instances according to the ontology schema.

A Study on Improvement of Collaborative Filtering Based on Implicit User Feedback Using RFM Multidimensional Analysis (RFM 다차원 분석 기법을 활용한 암시적 사용자 피드백 기반 협업 필터링 개선 연구)

  • Lee, Jae-Seong;Kim, Jaeyoung;Kang, Byeongwook
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.139-161
    • /
    • 2019
  • The utilization of the e-commerce market has become a common life style in today. It has become important part to know where and how to make reasonable purchases of good quality products for customers. This change in purchase psychology tends to make it difficult for customers to make purchasing decisions in vast amounts of information. In this case, the recommendation system has the effect of reducing the cost of information retrieval and improving the satisfaction by analyzing the purchasing behavior of the customer. Amazon and Netflix are considered to be the well-known examples of sales marketing using the recommendation system. In the case of Amazon, 60% of the recommendation is made by purchasing goods, and 35% of the sales increase was achieved. Netflix, on the other hand, found that 75% of movie recommendations were made using services. This personalization technique is considered to be one of the key strategies for one-to-one marketing that can be useful in online markets where salespeople do not exist. Recommendation techniques that are mainly used in recommendation systems today include collaborative filtering and content-based filtering. Furthermore, hybrid techniques and association rules that use these techniques in combination are also being used in various fields. Of these, collaborative filtering recommendation techniques are the most popular today. Collaborative filtering is a method of recommending products preferred by neighbors who have similar preferences or purchasing behavior, based on the assumption that users who have exhibited similar tendencies in purchasing or evaluating products in the past will have a similar tendency to other products. However, most of the existed systems are recommended only within the same category of products such as books and movies. This is because the recommendation system estimates the purchase satisfaction about new item which have never been bought yet using customer's purchase rating points of a similar commodity based on the transaction data. In addition, there is a problem about the reliability of purchase ratings used in the recommendation system. Reliability of customer purchase ratings is causing serious problems. In particular, 'Compensatory Review' refers to the intentional manipulation of a customer purchase rating by a company intervention. In fact, Amazon has been hard-pressed for these "compassionate reviews" since 2016 and has worked hard to reduce false information and increase credibility. The survey showed that the average rating for products with 'Compensated Review' was higher than those without 'Compensation Review'. And it turns out that 'Compensatory Review' is about 12 times less likely to give the lowest rating, and about 4 times less likely to leave a critical opinion. As such, customer purchase ratings are full of various noises. This problem is directly related to the performance of recommendation systems aimed at maximizing profits by attracting highly satisfied customers in most e-commerce transactions. In this study, we propose the possibility of using new indicators that can objectively substitute existing customer 's purchase ratings by using RFM multi-dimensional analysis technique to solve a series of problems. RFM multi-dimensional analysis technique is the most widely used analytical method in customer relationship management marketing(CRM), and is a data analysis method for selecting customers who are likely to purchase goods. As a result of verifying the actual purchase history data using the relevant index, the accuracy was as high as about 55%. This is a result of recommending a total of 4,386 different types of products that have never been bought before, thus the verification result means relatively high accuracy and utilization value. And this study suggests the possibility of general recommendation system that can be applied to various offline product data. If additional data is acquired in the future, the accuracy of the proposed recommendation system can be improved.