• Title/Summary/Keyword: database systems

Search Result 2,864, Processing Time 0.029 seconds

Research Direction for Functional Foods Safety (건강기능식품 안전관리 연구방향)

  • Jung, Ki-Hwa
    • Journal of Food Hygiene and Safety
    • /
    • v.25 no.4
    • /
    • pp.410-417
    • /
    • 2010
  • Various functional foods, marketing health and functional effects, have been distributed in the market. These products, being in forms of foods, tablets, and capsules, are likely to be mistaken as drugs. In addition, non-experts may sell these as foods, or use these for therapy. Efforts for creating health food regulations or building regulatory system for improving the current status of functional foods have been made, but these have not been communicated to consumers yet. As a result, problems of circulating functional foods for therapy or adding illegal medical to such products have persisted, which has become worse by internet media. The cause of this problem can be categorized into (1) product itself and (2) its use, but in either case, one possible cause is lack of communications with consumers. Potential problems that can be caused by functional foods include illegal substances, hazardous substances, allergic reactions, considerations when administered to patients, drug interactions, ingredients with purity or concentrations too low to be detected, products with metabolic activations, health risks from over- or under-dose of vitamin and minerals, and products with alkaloids. (Journal of Health Science, 56, Supplement (2010)). The reason why side effects related to functional foods have been increasing is that under-qualified functional food companies are exaggerating the functionality for marketing purposes. KFDA has been informing consumers, through its web pages, to address the above mentioned issues related to functional foods, but there still is room for improvement, to promote proper use of functional foods and avoid drug interactions. Specifically, to address these issues, institutionalizing to collect information on approved products and their side effects, settling reevaluation systems, and standardizing preclinical tests and clinical tests are becoming urgent. Also to provide crucial information, unified database systems, seamlessly aggregating heterogeneous data in different domains, with user interfaces enabling effective one-stop search, are crucial.

Adaptive Lock Escalation in Database Management Systems (데이타베이스 관리 시스템에서의 적응형 로크 상승)

  • Chang, Ji-Woong;Lee, Young-Koo;Whang, Kyu-Young;Yang, Jae-Heon
    • Journal of KIISE:Databases
    • /
    • v.28 no.4
    • /
    • pp.742-757
    • /
    • 2001
  • Since database management systems(DBMSS) have limited lock resources, transactions requesting locks beyond the limit mutt be aborted. In the worst carte, if such transactions are aborted repeatedly, the DBMS can become paralyzed, i.e., transaction execute but cannot commit. Lock escalation is considered a solution to this problem. However, existing lock escalation methods do not provide a complete solution. In this paper, we prognose a new lock escalation method, adaptive lock escalation, that selves most of the problems. First, we propose a general model for lock escalation and present the concept of the unescalatable look, which is the major cause making the transactions to abort. Second, we propose the notions of semi lock escalation, lock blocking, and selective relief as the mechanisms to control the number of unescalatable locks. We then propose the adaptive lock escalation method using these notions. Adaptive lock escalation reduces needless aborts and guarantees that the DBMS is not paralyzed under excessive lock requests. It also allows graceful degradation of performance under those circumstances. Third, through extensive simulation, we show that adaptive lock escalation outperforms existing lock escalation methods. The results show that, compared to the existing methods, adaptive lock escalation reduces the number of aborts and the average response time, and increases the throughput to a great extent. Especially, it is shown that the number of concurrent transactions can be increased more than 16 ~256 fold. The contribution of this paper is significant in that it has formally analysed the role of lock escalation in lock resource management and identified the detailed underlying mechanisms. Existing lock escalation methods rely on users or system administrator to handle the problems of excessive lock requests. In contrast, adaptive lock escalation releases the users of this responsibility by providing graceful degradation and preventing system paralysis through automatic control of unescalatable locks Thus adaptive lock escalation can contribute to developing self-tuning: DBMSS that draw a lot of attention these days.

  • PDF

Twitter Issue Tracking System by Topic Modeling Techniques (토픽 모델링을 이용한 트위터 이슈 트래킹 시스템)

  • Bae, Jung-Hwan;Han, Nam-Gi;Song, Min
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.2
    • /
    • pp.109-122
    • /
    • 2014
  • People are nowadays creating a tremendous amount of data on Social Network Service (SNS). In particular, the incorporation of SNS into mobile devices has resulted in massive amounts of data generation, thereby greatly influencing society. This is an unmatched phenomenon in history, and now we live in the Age of Big Data. SNS Data is defined as a condition of Big Data where the amount of data (volume), data input and output speeds (velocity), and the variety of data types (variety) are satisfied. If someone intends to discover the trend of an issue in SNS Big Data, this information can be used as a new important source for the creation of new values because this information covers the whole of society. In this study, a Twitter Issue Tracking System (TITS) is designed and established to meet the needs of analyzing SNS Big Data. TITS extracts issues from Twitter texts and visualizes them on the web. The proposed system provides the following four functions: (1) Provide the topic keyword set that corresponds to daily ranking; (2) Visualize the daily time series graph of a topic for the duration of a month; (3) Provide the importance of a topic through a treemap based on the score system and frequency; (4) Visualize the daily time-series graph of keywords by searching the keyword; The present study analyzes the Big Data generated by SNS in real time. SNS Big Data analysis requires various natural language processing techniques, including the removal of stop words, and noun extraction for processing various unrefined forms of unstructured data. In addition, such analysis requires the latest big data technology to process rapidly a large amount of real-time data, such as the Hadoop distributed system or NoSQL, which is an alternative to relational database. We built TITS based on Hadoop to optimize the processing of big data because Hadoop is designed to scale up from single node computing to thousands of machines. Furthermore, we use MongoDB, which is classified as a NoSQL database. In addition, MongoDB is an open source platform, document-oriented database that provides high performance, high availability, and automatic scaling. Unlike existing relational database, there are no schema or tables with MongoDB, and its most important goal is that of data accessibility and data processing performance. In the Age of Big Data, the visualization of Big Data is more attractive to the Big Data community because it helps analysts to examine such data easily and clearly. Therefore, TITS uses the d3.js library as a visualization tool. This library is designed for the purpose of creating Data Driven Documents that bind document object model (DOM) and any data; the interaction between data is easy and useful for managing real-time data stream with smooth animation. In addition, TITS uses a bootstrap made of pre-configured plug-in style sheets and JavaScript libraries to build a web system. The TITS Graphical User Interface (GUI) is designed using these libraries, and it is capable of detecting issues on Twitter in an easy and intuitive manner. The proposed work demonstrates the superiority of our issue detection techniques by matching detected issues with corresponding online news articles. The contributions of the present study are threefold. First, we suggest an alternative approach to real-time big data analysis, which has become an extremely important issue. Second, we apply a topic modeling technique that is used in various research areas, including Library and Information Science (LIS). Based on this, we can confirm the utility of storytelling and time series analysis. Third, we develop a web-based system, and make the system available for the real-time discovery of topics. The present study conducted experiments with nearly 150 million tweets in Korea during March 2013.

Life Cylcle Assessment (LCA) on Rice Production Systems: Comparison of Greenhouse Gases (GHGs) Emission on Conventional, Without Agricultural Chemical and Organic Farming (쌀 생산체계에 대한 영농방법별 전과정평가: 관행농, 무농약, 유기농법별 탄소배출량 비교)

  • Ryu, Jong-Hee;Kwon, Young-Rip;Kim, Gun-Yeob;Lee, Jong-Sik;Kim, Kye-Hoon;So, Kyu-Ho
    • Korean Journal of Soil Science and Fertilizer
    • /
    • v.45 no.6
    • /
    • pp.1157-1163
    • /
    • 2012
  • This study was performed a comparative life cycle assessment (LCA) among three rice production systems in order to analyze the difference of greenhouse gases (GHGs) emissions and environment impacts. Its life cycle inventory (LCI) database (DB) was established using data obtained from interview with conventional, without agricultural chemical and organic farming at Gunsan and Iksan, Jeonbuk province in 2011. According to the result of LCI analysis, $CO_2$ was mostly emitted from fertilizer production process and rice cropping phase. $CH_4$ and $N_2O$ were almost emitted from rice cultivation phase. The value of carbon footprint to produce 1 kg rice (unhulled) on conventional rice production system was 1.01E+00 kg $CO_2$-eq. $kg^{-1}$ and it was the highest value among three rice production systems. The value of carbon footprints on without agricultural chemical and organic rice production systems were 5.37E-01 $CO_2$-eq. $kg^{-1}$ and 6.58E-01 $CO_2$-eq. $kg^{-1}$, respectively. Without agricultural chemical rice production system whose input amount was the smallest had the lowest value of carbon footprint. Although the yield of rice from organic farming was the lowest, its value of carbon footprint less than that of conventional farming. Because there is no compound fertilizer inputs in organic farming. Compound fertilizer production and methane emission during rice cultivation were the main factor to GHGs emission in conventional and without agricultural chemical rice production systems. In organic rice production system, the main factors to GHGs emission were using fossil fuel on machine operation and methane emission from rice paddy field.

A Queriable XML Compression using Inferred Data Types (추론한 데이타 타입을 이용한 질의 가능 XML 압축)

  • ;;Chung Chin-Wan
    • Journal of KIISE:Databases
    • /
    • v.32 no.4
    • /
    • pp.441-451
    • /
    • 2005
  • HTML is mostly stored in native file systems instead of specialized repositories such as a database. Like HTML, XML, the standard for the exchange and the representation of data in the Internet, is mostly resident on native file systems. However. since XML data is irregular and verbose, the disk space and the network bandwidth are wasted compared to those of regularly structured data. To overcome this inefficiency of XML data, the research on the compression of XML data has been conducted. Among recently proposed XML compression techniques, some techniques do not support querying compressed data, while other techniques which support querying compressed data blindly encode data values using predefined encoding methods without considering the types of data values which necessitates partial decompression for processing range queries. As a result, the query performance on compressed XML data is degraded. Thus, this research proposes an XML compression technique which supports direct and efficient evaluations of queries on compressed XML data. This XML compression technique adopts an encoding method, called dictionary encoding, to encode each tag of XML data and applies proper encoding methods for encoding data values according to the inferred types of data values. Also, through the implementation and the performance evaluation of the XML compression technique proposed in this research, it is shown that the implemented XML compressor efficiently compresses real-life XML data lets and achieves significant improvements on query performance for compressed XML data.

A Design of Open GIS Compliant Object Web (개방형 GIS 표준에 따른 오브젝트 웹 시스템 설계)

  • Park, Ki-Ho;Jeong, Jae-Gon
    • Journal of Korea Spatial Information System Society
    • /
    • v.1 no.2 s.2
    • /
    • pp.47-62
    • /
    • 1999
  • To meet the interoperability requirements associated with geospatial data access and geoprocessing, much efforts to design prototypical systems conforming to $OpenGIS^{TM}$ specification has been made. With respect to Object Web GIS, however, current internet mapping technology mainly focuses on either developing mapping libraries or client applications regardless of the future needs for interoperability such as an integration of $OpenGIS^{TM}$ standard for CORBA. In this paper, we propose an $OpenGIS^{TM}$ compliant mapping kernel. OpenViews, which is designed to meet those requirements. The kernel of OpenViews encapsulates the process of acquiring geospatial data in the format of $OpenGIS^{TM}$ Geometry through ORB(Object Request Broker). OpenViews, being designed based on well-known design patterns, is a highly extensible in that programmers can easily customize it on the object oriented architecture. The components implemented in OpenViews are CORBA/Java objects, and as such are portable and scalable in a networked environment. A companion package, OpenBroker, is also developed as a portable geoprocessing application server to facilitate the implementation and configuration of server side CORBA objects. It can be used for implementing objects for spatial analysis service which would be independent of legacy spatial database systems in many cases. OpenViews, together with OpenBroker, has been successfully prototyped using the technologies including EJB and servlet as the core components of an Open GIS compliant internet mapping.

  • PDF

Design and Implementation of Query Processor for Moving Objects (이동객체를 위한 질의처리 컴포넌트의 설계 및 구현)

  • Kim, Kyoung-Sook;Kwon, O-Je;Byun, Hee-Young;Jo, Dae-Soo;Kim, Tae-Wan;Li, Ki-Joune
    • Journal of Korea Spatial Information System Society
    • /
    • v.6 no.1 s.11
    • /
    • pp.31-50
    • /
    • 2004
  • With the growth of wireless communication networks and mobile devices taking in GPS, Location-Based Service(LBS) is becoming an integral part of mobile applications. LBS can deal with location-aware features such as persons holding mobile phones or vehicles equipped with GPS, and provide the users with the location information of the features. Thus it is necessary to develop moving object database systems to store, manage, and query moving objects which change their locations continuously as time passes. In this paper, we design and implement a query processing component which deals with moving objects as a key data type. For this component, we define a new SQL-like query language(called MOQL) and as a consequence, design and implement modules that analyze and execute queries. It supports various types of operators that process range queries, infer topological relations, compute trajectories, and find k-nearest neighbors. It can be used as a subsystem if other application systems which deal moving objects and also supports ADO.NET interface that can be used to interact end-users.

  • PDF

A Study on the Aviation Safety Information System in Korea - With the Comparative Analysis of the ICAO, the United States and EU system (한국의 항공안전 데이터베이스 분류체계에 관한 연구)

  • Lee, Kang-Seok
    • The Korean Journal of Air & Space Law and Policy
    • /
    • v.25 no.2
    • /
    • pp.113-137
    • /
    • 2010
  • The international community has done a lot of effort to improve the aviation safety. The new high technologies, systems and development of navigation equipments considerably improved the level of aviation safety. However, the aviation accident rate has been showing unchanged status without the decline since 1980. It means it is difficult to secure enough safety by the existing efforts to improve the aviation safety, eventually needs new methods for the accident prevention. With this in mind, the nations interest new methods that can consider a countermeasure by realizing potential risk factors before fatal accidents. In other words, that is to analyze and share the safety information collected from accidents by establishing aviation safety information system. To establish international aviation safety information system, each country's aviation accident incident reporting systems are need to be established. Also aviation safety data taxonomy is need to be standardized. Then, aviation users will communicate the information efficiently. This study investigates and considers each country's aviation accident incident database taxonomy for the analysis of the aviation data and the necessary prevention activity. This will contribute to a better international aviation safety and strengthen the country's aviation diplomacy.

  • PDF

A Study on Spoken Digits Analysis and Recognition (숫자음 분석과 인식에 관한 연구)

  • 김득수;황철준
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.6 no.3
    • /
    • pp.107-114
    • /
    • 2001
  • This paper describes Connected Digit Recognition with Considering Acoustic Feature in Korea. The recognition rate of connected digit is usually lower than word recognition. Therefore, speech feature parameter and acoustic feature are employed to make robust model for digit, and we could confirm the effect of Considering. Acoustic Feature throughout the experience of recognition. We used KLE 4 connected digit as database and 19 continuous distributed HMM as PLUs(Phoneme Like Units) using phonetical rules. For recognition experience, we have tested two cases. The first case, we used usual method like using Mel-Cepstrum and Regressive Coefficient for constructing phoneme model. The second case, we used expanded feature parameter and acoustic feature for constructing phoneme model. In both case, we employed OPDP(One Pass Dynamic Programming) and FSA(Finite State Automata) for recognition tests. When appling FSN for recognition, we applied various acoustic features. As the result, we could get 55.4% recognition rate for Mel-Cepstrum, and 67.4% for Mel-Cepstrum and Regressive Coefficient. Also, we could get 74.3% recognition rate for expanded feature parameter, and 75.4% for applying acoustic feature. Since, the case of applying acoustic feature got better result than former method, we could make certain that suggested method is effective for connected digit recognition in korean.

  • PDF

Systematizing and Improving of Spatial Environment Data for the Establishment of Spatial Environment Planning (공간환경계획 수립을 위한환경정보의 체계화와 개선방안)

  • Eum, Jeong-Hee;Choi, Hee-Sun;Lee, Gil-Sang
    • Journal of Environmental Policy
    • /
    • v.9 no.2
    • /
    • pp.111-133
    • /
    • 2010
  • Environmental conservation plan, notwithstanding their feasibility and potential utility in construction of environment-friendly spaces, has long been perceived in practice as "declarative" and a "formality." Such perceptions are largely the result of the failure to provide spatial planning that is directly relevant to development of the space in question, and to sufficiently interconnect with urban development plan. This demonstrates the need for ways to link disparate plans, i.e. to enact "spatial environment plan." In response to these issues, this study proposes the systematization of spatial environment data as a necessary prerequisite to the establishment of spatial environment plan, which would provide both linkages with other plans, and ensure the applicability of environmental conservation plan. To this end, this study analyzed existing environmental data, and then proposed systems for links with spatial environment plan. In respect of this, the study examined spatial data systems, and then classified applicable spatial according to each environmental medium. The study also produced spatial information and planning items that can be included in spatial environment plans for each of the nine environmental media, and then constructed a system that could link the existing spatial information system, current spatial environment data, and spatial environment management plan. Furthermore, the study proposed matters for improvement in the construction of spatial environmental data to ensure the activation of spatial environment plan. The construction of a systematic spatial database, by facilitating the smooth establishment of spatial environment plan, can enhance and upgrade environmental conservation plan, while contributing to enhanced linkages with related spatial plans.

  • PDF