• Title/Summary/Keyword: 콘텐츠 추출

Search Result 1,391, Processing Time 0.025 seconds

Study of Policy on Seowon's Preservation·Support : Focusing on Big Data Analysis on Laws (한국 서원의 보존·지원 정책에 관한 연구 : 법률에 대한 빅데이터 분석을 중심으로)

  • Bang, Mee Young
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.5
    • /
    • pp.875-883
    • /
    • 2023
  • In Korea, the number of preservation and management entities to connect the traditional cultural heritage to next generations is rapidly decreasing. Building an infrastructure to pass on traditional cultural heritage to the next generation and to pay attention to the preservation and management of the next generation is important including the 'Seowon', a World Cultural Heritage listed by UNESCO. This study is based on the laws that regulates the preservation and support of traditional cultural assets and 'Seowon, through Big Data analysis techniques. The main keywords in each law were extracted, schematized, and a mutual Word Network was constructed and policy advice was derived. As policy advice, it is necessary to establish and implement policies to nurture and support businesses specialized in the region for the preservation·utilization, preservation·management and preservation·support of Seowons.

A Study on Word-of-Mouth of an Electric Automobile using YouTube: A Focus on Statistical Network Analysis (유튜브를 활용한 전기 자동차 결함에 대한 구전 확산 연구: 네트워크 통계분석을 중심으로)

  • EuiBeom Jeong;Keontaek Oh
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.29 no.1
    • /
    • pp.15-29
    • /
    • 2024
  • With recent advances in information and communication technology, YouTube has become a powerful online space for users to create and share content about their interests and experiences, creating new cultural phenomena. In particular, there needs to be more research on social media in the manufacturing sector because, unlike distribution and retail, there has been relatively little direct contact with consumers. YouTube can positively affect firms' performance by promoting products and brands. On the other hand, it can also cause risks, such as production disruption due to rumors or misinformation. Thus, it is necessary for firms to examine how information about an electric automobile defects spreads on YouTube according to the number of subscribers and views through statistical network analysis.

A Knowledge Graph-based Chatbot to Prevent the Leakage of LLM User's Sensitive Information (LLM 사용자의 민감정보 유출 방지를 위한 지식그래프 기반 챗봇)

  • Keedong Yoo
    • Knowledge Management Research
    • /
    • v.25 no.2
    • /
    • pp.1-18
    • /
    • 2024
  • With the increasing demand for and utilization of large language models (LLMs), the risk of user sensitive information being inputted and leaked during the use of LLMs also escalates. Typically recognized as a tool for mitigating the hallucination issues of LLMs, knowledge graphs, constructed independently from LLMs, can store and manage sensitive user information separately, thereby minimizing the potential for data breaches. This study, therefore, presents a knowledge graph-based chatbot that transforms user-inputted natural language questions into queries appropriate for the knowledge graph using LLMs, subsequently executing these queries and extracting the results. Furthermore, to evaluate the functional validity of the developed knowledge graph-based chatbot, performance tests are conducted to assess the comprehension and adaptability to existing knowledge graphs, the capability to create new entity classes, and the accessibility of LLMs to the knowledge graph content.

Visualizing the Results of Opinion Mining from Social Media Contents: Case Study of a Noodle Company (소셜미디어 콘텐츠의 오피니언 마이닝결과 시각화: N라면 사례 분석 연구)

  • Kim, Yoosin;Kwon, Do Young;Jeong, Seung Ryul
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.4
    • /
    • pp.89-105
    • /
    • 2014
  • After emergence of Internet, social media with highly interactive Web 2.0 applications has provided very user friendly means for consumers and companies to communicate with each other. Users have routinely published contents involving their opinions and interests in social media such as blogs, forums, chatting rooms, and discussion boards, and the contents are released real-time in the Internet. For that reason, many researchers and marketers regard social media contents as the source of information for business analytics to develop business insights, and many studies have reported results on mining business intelligence from Social media content. In particular, opinion mining and sentiment analysis, as a technique to extract, classify, understand, and assess the opinions implicit in text contents, are frequently applied into social media content analysis because it emphasizes determining sentiment polarity and extracting authors' opinions. A number of frameworks, methods, techniques and tools have been presented by these researchers. However, we have found some weaknesses from their methods which are often technically complicated and are not sufficiently user-friendly for helping business decisions and planning. In this study, we attempted to formulate a more comprehensive and practical approach to conduct opinion mining with visual deliverables. First, we described the entire cycle of practical opinion mining using Social media content from the initial data gathering stage to the final presentation session. Our proposed approach to opinion mining consists of four phases: collecting, qualifying, analyzing, and visualizing. In the first phase, analysts have to choose target social media. Each target media requires different ways for analysts to gain access. There are open-API, searching tools, DB2DB interface, purchasing contents, and so son. Second phase is pre-processing to generate useful materials for meaningful analysis. If we do not remove garbage data, results of social media analysis will not provide meaningful and useful business insights. To clean social media data, natural language processing techniques should be applied. The next step is the opinion mining phase where the cleansed social media content set is to be analyzed. The qualified data set includes not only user-generated contents but also content identification information such as creation date, author name, user id, content id, hit counts, review or reply, favorite, etc. Depending on the purpose of the analysis, researchers or data analysts can select a suitable mining tool. Topic extraction and buzz analysis are usually related to market trends analysis, while sentiment analysis is utilized to conduct reputation analysis. There are also various applications, such as stock prediction, product recommendation, sales forecasting, and so on. The last phase is visualization and presentation of analysis results. The major focus and purpose of this phase are to explain results of analysis and help users to comprehend its meaning. Therefore, to the extent possible, deliverables from this phase should be made simple, clear and easy to understand, rather than complex and flashy. To illustrate our approach, we conducted a case study on a leading Korean instant noodle company. We targeted the leading company, NS Food, with 66.5% of market share; the firm has kept No. 1 position in the Korean "Ramen" business for several decades. We collected a total of 11,869 pieces of contents including blogs, forum contents and news articles. After collecting social media content data, we generated instant noodle business specific language resources for data manipulation and analysis using natural language processing. In addition, we tried to classify contents in more detail categories such as marketing features, environment, reputation, etc. In those phase, we used free ware software programs such as TM, KoNLP, ggplot2 and plyr packages in R project. As the result, we presented several useful visualization outputs like domain specific lexicons, volume and sentiment graphs, topic word cloud, heat maps, valence tree map, and other visualized images to provide vivid, full-colored examples using open library software packages of the R project. Business actors can quickly detect areas by a swift glance that are weak, strong, positive, negative, quiet or loud. Heat map is able to explain movement of sentiment or volume in categories and time matrix which shows density of color on time periods. Valence tree map, one of the most comprehensive and holistic visualization models, should be very helpful for analysts and decision makers to quickly understand the "big picture" business situation with a hierarchical structure since tree-map can present buzz volume and sentiment with a visualized result in a certain period. This case study offers real-world business insights from market sensing which would demonstrate to practical-minded business users how they can use these types of results for timely decision making in response to on-going changes in the market. We believe our approach can provide practical and reliable guide to opinion mining with visualized results that are immediately useful, not just in food industry but in other industries as well.

A Study on Causal Relationship About the Reparations Range (손해배상범위에 관한 인과관계의 연구)

  • Choi Hwan-Seok;Park Jong-Ryeol
    • The Journal of the Korea Contents Association
    • /
    • v.6 no.4
    • /
    • pp.146-157
    • /
    • 2006
  • Causal relationship means what relations the result occurred have with a fact as a reason. In general, a formular that no result exists without reasons is used for the method to confirm existence and inexistence of causal relationship. Problematic causal relationships in Private Law are reparations (Article No. 393 of Private Law) due to debt nonfulfillment and reparation due to tort (Application of Article No. 393 by Article No. 750, and No. 763 of Private Law). The purpose pursued by reparation system in private law is to promote equal burden of damages, and the range of reparation at this time is decided by the range of damage and the range of damage is decided by the principle of causal relationship. That the causal relationship theory fairly causes confusion by treating one problem and the other problem as the same thing, instead of dividing them according to the purpose of protection presented by the law is a reason of the criticism from different views.

  • PDF

A Study on Fast Iris Detection for Iris Recognition in Mobile Phone (휴대폰에서의 홍채인식을 위한 고속 홍채검출에 관한 연구)

  • Park Hyun-Ae;Park Kang-Ryoung
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.2 s.308
    • /
    • pp.19-29
    • /
    • 2006
  • As the security of personal information is becoming more important in mobile phones, we are starting to apply iris recognition technology to these devices. In conventional iris recognition, magnified iris images are required. For that, it has been necessary to use large magnified zoom & focus lens camera to capture images, but due to the requirement about low size and cost of mobile phones, the zoom & focus lens are difficult to be used. However, with rapid developments and multimedia convergence trends in mobile phones, more and more companies have built mega-pixel cameras into their mobile phones. These devices make it possible to capture a magnified iris image without zoom & focus lens. Although facial images are captured far away from the user using a mega-pixel camera, the captured iris region possesses sufficient pixel information for iris recognition. However, in this case, the eye region should be detected for accurate iris recognition in facial images. So, we propose a new fast iris detection method, which is appropriate for mobile phones based on corneal specular reflection. To detect specular reflection robustly, we propose the theoretical background of estimating the size and brightness of specular reflection based on eye, camera and illuminator models. In addition, we use the successive On/Off scheme of the illuminator to detect the optical/motion blurring and sunlight effect on input image. Experimental results show that total processing time(detecting iris region) is on average 65ms on a Samsung SCH-S2300 (with 150MHz ARM 9 CPU) mobile phone. The rate of correct iris detection is 99% (about indoor images) and 98.5% (about outdoor images).

A Study on the Management of Manhwa Contents Records and Archives (만화기록 관리 방안 연구)

  • Kim, Seon Mi;Kim, Ik Han
    • The Korean Journal of Archival Studies
    • /
    • no.28
    • /
    • pp.35-81
    • /
    • 2011
  • Manhwa is a mass media (to expose all faces of an era such as politics, society, cultures, etc with the methodology of irony, parody, etc). Since the Manhwa records is primary culture infrastructure, it can create the high value-added industry by connecting with fancy, character, game, movie, drama, theme park, advertising business. However, due to lack of active and systematic aquisition system, as precious Manhwa manuscript is being lost every year and the contents hard to preserve such as Manhwa content in the form of electronic records are increasing, the countermeasure of Manhwa contents management is needed desperately. In this study, based on these perceptions, the need of Manhwa records management is examined, and the characteristics and the components of Manhwa records were analyzed. And at the same time, the functions of record management process reflecting the characteristics of Manhwa records were extracted by analyzing various cases of overseas Cartoon Archives. And then, the framework of record-keeping regime was segmented into each of acquisition management service areas and the general Manhwa records archiving strategy, which manages the Manhwa contents records, was established and suggested. The acquired Manhwa content records will secure the context among records and warrant the preservation of records and provide diverse access points by reflecting multi classification and multi-level descriptive element. The Manhwa records completed the intellectual arrangement will be preserved after the conservation in an environment equipped with preservation facilities or preserved using digital format in case of electronic records or when there is potential risk of damaging the records. Since the purpose of the Manhwa records is to use them, the information may be provided to diverse classes of users through the exhibition, the distribution, and the development of archival information content. Since the term of "Manhwa records" is unfamiliar yet and almost no study has been conducted in the perspective of records management, it will be the limit of this study only presenting acquisition strategy, management and service strategy of Manhwa contents and suggesting simple examples. However, if Manhwa records management strategy are possibly introduced practically to Manhwa manuscript repositories through archival approach, it will allow systematic acquisition, preservation, arrangement of Manhwa records and will contribute greatly to form a foundation for future Korean culture contents management.

A Study on Knowledge Entity Extraction Method for Individual Stocks Based on Neural Tensor Network (뉴럴 텐서 네트워크 기반 주식 개별종목 지식개체명 추출 방법에 관한 연구)

  • Yang, Yunseok;Lee, Hyun Jun;Oh, Kyong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.25-38
    • /
    • 2019
  • Selecting high-quality information that meets the interests and needs of users among the overflowing contents is becoming more important as the generation continues. In the flood of information, efforts to reflect the intention of the user in the search result better are being tried, rather than recognizing the information request as a simple string. Also, large IT companies such as Google and Microsoft focus on developing knowledge-based technologies including search engines which provide users with satisfaction and convenience. Especially, the finance is one of the fields expected to have the usefulness and potential of text data analysis because it's constantly generating new information, and the earlier the information is, the more valuable it is. Automatic knowledge extraction can be effective in areas where information flow is vast, such as financial sector, and new information continues to emerge. However, there are several practical difficulties faced by automatic knowledge extraction. First, there are difficulties in making corpus from different fields with same algorithm, and it is difficult to extract good quality triple. Second, it becomes more difficult to produce labeled text data by people if the extent and scope of knowledge increases and patterns are constantly updated. Third, performance evaluation is difficult due to the characteristics of unsupervised learning. Finally, problem definition for automatic knowledge extraction is not easy because of ambiguous conceptual characteristics of knowledge. So, in order to overcome limits described above and improve the semantic performance of stock-related information searching, this study attempts to extract the knowledge entity by using neural tensor network and evaluate the performance of them. Different from other references, the purpose of this study is to extract knowledge entity which is related to individual stock items. Various but relatively simple data processing methods are applied in the presented model to solve the problems of previous researches and to enhance the effectiveness of the model. From these processes, this study has the following three significances. First, A practical and simple automatic knowledge extraction method that can be applied. Second, the possibility of performance evaluation is presented through simple problem definition. Finally, the expressiveness of the knowledge increased by generating input data on a sentence basis without complex morphological analysis. The results of the empirical analysis and objective performance evaluation method are also presented. The empirical study to confirm the usefulness of the presented model, experts' reports about individual 30 stocks which are top 30 items based on frequency of publication from May 30, 2017 to May 21, 2018 are used. the total number of reports are 5,600, and 3,074 reports, which accounts about 55% of the total, is designated as a training set, and other 45% of reports are designated as a testing set. Before constructing the model, all reports of a training set are classified by stocks, and their entities are extracted using named entity recognition tool which is the KKMA. for each stocks, top 100 entities based on appearance frequency are selected, and become vectorized using one-hot encoding. After that, by using neural tensor network, the same number of score functions as stocks are trained. Thus, if a new entity from a testing set appears, we can try to calculate the score by putting it into every single score function, and the stock of the function with the highest score is predicted as the related item with the entity. To evaluate presented models, we confirm prediction power and determining whether the score functions are well constructed by calculating hit ratio for all reports of testing set. As a result of the empirical study, the presented model shows 69.3% hit accuracy for testing set which consists of 2,526 reports. this hit ratio is meaningfully high despite of some constraints for conducting research. Looking at the prediction performance of the model for each stocks, only 3 stocks, which are LG ELECTRONICS, KiaMtr, and Mando, show extremely low performance than average. this result maybe due to the interference effect with other similar items and generation of new knowledge. In this paper, we propose a methodology to find out key entities or their combinations which are necessary to search related information in accordance with the user's investment intention. Graph data is generated by using only the named entity recognition tool and applied to the neural tensor network without learning corpus or word vectors for the field. From the empirical test, we confirm the effectiveness of the presented model as described above. However, there also exist some limits and things to complement. Representatively, the phenomenon that the model performance is especially bad for only some stocks shows the need for further researches. Finally, through the empirical study, we confirmed that the learning method presented in this study can be used for the purpose of matching the new text information semantically with the related stocks.

Research for Characteristics of Sound Localization at Monaural System Using Acoustic Energy (청각에너지를 이용한 모노럴 시스템에서의 음상 정위 특성 연구)

  • Koo, Kyo-Sik;Cha, Hyung-Tai
    • The Journal of the Acoustical Society of Korea
    • /
    • v.30 no.4
    • /
    • pp.181-189
    • /
    • 2011
  • According to developments of digital signal processing, 3D sound come into focus on multimedia systems. Many studies on 3d sound have proposed lots of clues to create realistic sounds. But these clues are only focused on binaural systems which two ears are normal. If we make the 3d sound using those clues at monaural systems, the performance goes down dramatically. In order to use the clues for monaural systems, we have studies algorithms such as duplex theory. In duplex theory, the sounds that we listen are affected by human's body, pinna and shoulder. So, we can enhance sound localization performances using its characteristics. In this paper, we propose a new method to use psychoacoustic theory that creates realistic 3D audio at monaural systems. To improve 3d sound, we calculate the excitation energy rates of each symmetric HRTF and extract the weights in each bark range. Finally, they are applied to emphasize the characteristics related to each direction. Informal listening tests show that the proposed method improves sound localization performances much better than the conventional methods.

Robust Semi-auto Calibration Method for Various Cameras and Illumination Changes (다양한 카메라와 조명의 변화에 강건한 반자동 카메라 캘리브레이션 방법)

  • Shin, Dong-Won;Ho, Yo-Sung
    • Journal of Broadcast Engineering
    • /
    • v.21 no.1
    • /
    • pp.36-42
    • /
    • 2016
  • Recently, many 3D contents have been produced through the multiview camera system. In this system, since a difference of the viewpoint between color and depth cameras is inevitable, the camera parameter plays the important role to adjust the viewpoint as a preprocessing step. The conventional camera calibration method is inconvenient to users since we need to choose pattern features manually after capturing a planar chessboard with various poses. Therefore, we propose a semi-auto camera calibration method using a circular sampling and an homography estimation. Firstly, The proposed method extracts the candidates of the pattern features from the images by FAST corner detector. Next, we reduce the amount of the candidates by the circular sampling and obtain the complete point cloud by the homography estimation. Lastly, we compute the accurate position having the sub-pixel accuracy of the pattern features by the approximation of the hyper parabola surface. We investigated which factor affects the result of the pattern feature detection at each step. Compared to the conventional method, we found the proposed method released the inconvenience of the manual operation but maintained the accuracy of the camera parameters.