• Title/Summary/Keyword: well log data

Search Result 250, Processing Time 0.027 seconds

X-tree Diff: An Efficient Change Detection Algorithm for Tree-structured Data (X-tree Diff: 트리 기반 데이터를 위한 효율적인 변화 탐지 알고리즘)

  • Lee, Suk-Kyoon;Kim, Dong-Ah
    • The KIPS Transactions:PartC
    • /
    • v.10C no.6
    • /
    • pp.683-694
    • /
    • 2003
  • We present X-tree Diff, a change detection algorithm for tree-structured data. Our work is motivated by need to monitor massive volume of web documents and detect suspicious changes, called defacement attack on web sites. From this context, our algorithm should be very efficient in speed and use of memory space. X-tree Diff uses a special ordered labeled tree, X-tree, to represent XML/HTML documents. X-tree nodes have a special field, tMD, which stores a 128-bit hash value representing the structure and data of subtrees, so match identical subtrees form the old and new versions. During this process, X-tree Diff uses the Rule of Delaying Ambiguous Matchings, implying that it perform exact matching where a node in the old version has one-to one corrspondence with the corresponding node in the new, by delaying all the others. It drastically reduces the possibility of wrong matchings. X-tree Diff propagates such exact matchings upwards in Step 2, and obtain more matchings downwsards from roots in Step 3. In step 4, nodes to ve inserted or deleted are decided, We aldo show thst X-tree Diff runs on O(n), woere n is the number of noses in X-trees, in worst case as well as in average case, This result is even better than that of BULD Diff algorithm, which is O(n log(n)) in worst case, We experimented X-tree Diff on reat data, which are about 11,000 home pages from about 20 wev sites, instead of synthetic documets manipulated for experimented for ex[erimentation. Currently, X-treeDiff algorithm is being used in a commeercial hacking detection system, called the WIDS(Web-Document Intrusion Detection System), which is to find changes occured in registered websites, and report suspicious changes to users.

Intelligent Brand Positioning Visualization System Based on Web Search Traffic Information : Focusing on Tablet PC (웹검색 트래픽 정보를 활용한 지능형 브랜드 포지셔닝 시스템 : 태블릿 PC 사례를 중심으로)

  • Jun, Seung-Pyo;Park, Do-Hyung
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.3
    • /
    • pp.93-111
    • /
    • 2013
  • As Internet and information technology (IT) continues to develop and evolve, the issue of big data has emerged at the foreground of scholarly and industrial attention. Big data is generally defined as data that exceed the range that can be collected, stored, managed and analyzed by existing conventional information systems and it also refers to the new technologies designed to effectively extract values from such data. With the widespread dissemination of IT systems, continual efforts have been made in various fields of industry such as R&D, manufacturing, and finance to collect and analyze immense quantities of data in order to extract meaningful information and to use this information to solve various problems. Since IT has converged with various industries in many aspects, digital data are now being generated at a remarkably accelerating rate while developments in state-of-the-art technology have led to continual enhancements in system performance. The types of big data that are currently receiving the most attention include information available within companies, such as information on consumer characteristics, information on purchase records, logistics information and log information indicating the usage of products and services by consumers, as well as information accumulated outside companies, such as information on the web search traffic of online users, social network information, and patent information. Among these various types of big data, web searches performed by online users constitute one of the most effective and important sources of information for marketing purposes because consumers search for information on the internet in order to make efficient and rational choices. Recently, Google has provided public access to its information on the web search traffic of online users through a service named Google Trends. Research that uses this web search traffic information to analyze the information search behavior of online users is now receiving much attention in academia and in fields of industry. Studies using web search traffic information can be broadly classified into two fields. The first field consists of empirical demonstrations that show how web search information can be used to forecast social phenomena, the purchasing power of consumers, the outcomes of political elections, etc. The other field focuses on using web search traffic information to observe consumer behavior, identifying the attributes of a product that consumers regard as important or tracking changes on consumers' expectations, for example, but relatively less research has been completed in this field. In particular, to the extent of our knowledge, hardly any studies related to brands have yet attempted to use web search traffic information to analyze the factors that influence consumers' purchasing activities. This study aims to demonstrate that consumers' web search traffic information can be used to derive the relations among brands and the relations between an individual brand and product attributes. When consumers input their search words on the web, they may use a single keyword for the search, but they also often input multiple keywords to seek related information (this is referred to as simultaneous searching). A consumer performs a simultaneous search either to simultaneously compare two product brands to obtain information on their similarities and differences, or to acquire more in-depth information about a specific attribute in a specific brand. Web search traffic information shows that the quantity of simultaneous searches using certain keywords increases when the relation is closer in the consumer's mind and it will be possible to derive the relations between each of the keywords by collecting this relational data and subjecting it to network analysis. Accordingly, this study proposes a method of analyzing how brands are positioned by consumers and what relationships exist between product attributes and an individual brand, using simultaneous search traffic information. It also presents case studies demonstrating the actual application of this method, with a focus on tablets, belonging to innovative product groups.

Accuracy Analysis of ADCP Stationary Discharge Measurement for Unmeasured Regions (ADCP 정지법 측정 시 미계측 영역의 유량 산정 정확도 분석)

  • Kim, Jongmin;Kim, Seojun;Son, Geunsoo;Kim, Dongsu
    • Journal of Korea Water Resources Association
    • /
    • v.48 no.7
    • /
    • pp.553-566
    • /
    • 2015
  • Acoustic Doppler Current Profilers(ADCPs) have capability to concurrently capitalize three-dimensional velocity vector and bathymetry with highly efficient and rapid manner, and thereby enabling ADCPs to document the hydrodynamic and morphologic data in very high spatial and temporal resolution better than other contemporary instruments. However, ADCPs are also limited in terms of the inevitable unmeasured regions near bottom, surface, and edges of a given cross-section. The velocity in those unmeasured regions are usually extrapolated or assumed for calculating flow discharge, which definitely affects the accuracy in the discharge assessment. This study aimed at scrutinizing a conventional extrapolation method(i.e., the 1/6 power law) for estimating the unmeasured regions to figure out the accuracy in ADCP discharge measurements. For the comparative analysis, we collected spatially dense velocity data using ADV as well as stationary ADCP in a real-scale straight river channel, and applied the 1/6 power law for testing its applicability in conjunction with the logarithmic law which is another representative velocity law. As results, the logarithmic law fitted better with actual velocity measurement than the 1/6 power law. In particular, the 1/6 power law showed a tendency to underestimate the velocity in the near surface region and overestimate in the near bottom region. This finding indicated that the 1/6 power law could be unsatisfactory to follow actual flow regime, thus that resulted discharge estimates in both unmeasured top and bottom region can give rise to discharge bias. Therefore, the logarithmic law should be considered as an alternative especially for the stationary ADCP discharge measurement. In addition, it was found that ADCP should be operated in at least more than 0.6 m of water depth in the left and right edges for better estimate edge discharges. In the future, similar comparative analysis might be required for the moving boat ADCP discharge measurement method, which has been more widely used in the field.

Characterization and Formation Mechanisms of Clogging Materials in Groundwater Wells, Mt. Geumjeong Area, Busan, Korea (부산 금정산 일대 지하수공내 공막힘 물질의 특징과 형성원인)

  • Choo, Chang-Oh;Hamm, Se-Yeong;Lee, Jeong-Hwan;Lee, Chung-Mo;Choo, Youn-Woo;Han, Suk-Jong;Kim, Moo-Jin;Cho, Heuy-Nam
    • The Journal of Engineering Geology
    • /
    • v.22 no.1
    • /
    • pp.67-81
    • /
    • 2012
  • The physical, chemical, and biological properties of clogging materials formed within groundwater wells in the Mt. Geumjeong area, Busan, Korea, were characterized. The particle size distribution (PSD) of clogging materials was measured by a laser analyzer. XRD, SEM, and TEM analyses were performed to obtain mineralogical information on the clogging materials, with an emphasis on identifying and characterizing the mineral species. In most cases, PSD data exhibited an near log-normal distribution; however, variations in frequency distribution were found in some intervals (bi-or trimodal distributions), raising the possibility that particles originated from several sources or were formed at different times. XRD data revealed that the clogging materials were mainly amorphous ironhydroxides such as goethite, ferrihydrite, and lapidocrocite, with lesser amounts of Fe, Mn, and Zn metals and silicates such as quartz, feldspar, micas, and smectite. Reddish brown material was amorphous hydrous ferriciron (HFO), and dark red and dark black materials were Fe, Mn-hydroxides. Greyish white and pale brown materials consisted of silicates. SEM observations indicated that the clogging materials were mainly HFO associated with iron bacteria such as Gallionella and Leptothrix, with small amounts of rock fragments. In TEM analysis, disseminated iron particles were commonly observed in the cell and sheath of iron bacteria, indicating that iron was precipitated in close association with the metabolism of bacterial activity. Rock-forming minerals such as quartz, feldspar, and micas were primarily derived from soils or granite aquifers, which are widely distributed in the study area. The results indicate the importance of elucidating the formation mechanisms of clogging materials to ensure sustainable well capacity.

Performance of Drip Irrigation System in Banana Cultuivation - Data Envelopment Analysis Approach

  • Kumar, K. Nirmal Ravi;Kumar, M. Suresh
    • Agribusiness and Information Management
    • /
    • v.8 no.1
    • /
    • pp.17-26
    • /
    • 2016
  • India is largest producer of banana in the world producing 29.72 million tonnes from an area of 0.803 million ha with a productivity of 35.7 MT ha-1 and accounted for 15.48 and 27.01 per cent of the world's area and production respectively (www.nhb.gov.in). In India, Tamil Nadu leads other states both in terms of area and production followed by Maharashtra, Gujarat and Andhra Pradesh. In Rayalaseema region of Andhra Pradesh, Kurnool district had special reputation in the cultivation of banana in an area of 5765 hectares with an annual production of 2.01 lakh tonnes in the year 2012-13 and hence, it was purposively chosen for the study. On $23^{rd}$ November 2003, the Government of Andhra Pradesh has commenced a comprehensive project called 'Andhra Pradesh Micro Irrigation Project (APMIP)', first of its kind in the world so as to promote water use efficiency. APMIP is offering 100 per cent of subsidy in case of SC, ST and 90 per cent in case of other categories of farmers up to 5.0 acres of land. In case of acreage between 5-10 acres, 70 per cent subsidy and acreage above 10, 50 per cent of subsidy is given to the farmer beneficiaries. The sampling frame consists of Kurnool district, two mandals, four villages and 180 sample farmers comprising of 60 farmers each from Marginal (<1ha), Small (1-2ha) and Other (>2ha) categories. A well structured pre-tested schedule was employed to collect the requisite information pertaining to the performance of drip irrigation among the sample farmers and Data Envelopment Analysis (DEA) model was employed to analyze the performance of drip irrigation in banana farms. The performance of drip irrigation was assessed based on the parameters like: Land Development Works (LDW), Fertigation costs (FC), Volume of water supplied (VWS), Annual maintenance costs of drip irrigation (AMC), Economic Status of the farmer (ES), Crop Productivity (CP) etc. The first four parameters are considered as inputs and last two as outputs for DEA modelling purposes. The findings revealed that, the number of farms operating at CRS are more in number in other farms (46.66%) followed by marginal (45%) and small farms (28.33%). Similarly, regarding the number of farmers operating at VRS, the other farms are again more in number with 61.66 per cent followed by marginal (53.33%) and small farms (35%). With reference to scale efficiency, marginal farms dominate the scenario with 57 per cent followed by others (55%) and small farms (50%). At pooled level, 26.11 per cent of the farms are being operated at CRS with an average technical efficiency score of 0.6138 i.e., 47 out of 180 farms. Nearly 40 per cent of the farmers at pooled level are being operated at VRS with an average technical efficiency score of 0.7241. As regards to scale efficiency, nearly 52 per cent of the farmers (94 out of 180 farmers) at pooled level, either performed at the optimum scale or were close to the optimum scale (farms having scale efficiency values equal to or more than 0.90). Majority of the farms (39.44%) are operating at IRS and only 29 per cent of the farmers are operating at DRS. This signifies that, more resources should be provided to these farms operating at IRS and the same should be decreased towards the farms operating at DRS. Nearly 32 per cent of the farms are operating at CRS indicating efficient utilization of resources. Log linear regression model was used to analyze the major determinants of input use efficiency in banana farms. The input variables considered under DEA model were again considered as influential factors for the CRS obtained for the three categories of farmers. Volume of water supplied ($X_1$) and fertigation cost ($X_2$) are the major determinants of banana farms across all the farmer categories and even at pooled level. In view of their positive influence on the CRS, it is essential to strengthen modern irrigation infrastructure like drip irrigation and offer more fertilizer subsidies to the farmer to enhance the crop production on cost-effective basis in Kurnool district of Andhra Pradesh, India. This study further suggests that, the present era of Information Technology will help the irrigation management in the context of generating new techniques, extension, adoption and information. It will also guide the farmers in irrigation scheduling and quantifying the irrigation water requirements in accordance with the water availability in a particular season. So, it is high time for the Government of India to pay adequate attention towards the applications of 'Information and Communication Technology (ICT) and its applications in irrigation water management' for facilitating the deployment of Decision Supports Systems (DSSs) at various levels of planning and management of water resources in the country.

Development of Evaluation Model for ITS Project using the Probabilistic Risk Analysis (확률적 위험도분석을 이용한 ITS사업의 경제성평가모형)

  • Lee, Yong-Taeck;Nam, Doo-Hee;Lim, Kang-Won
    • Journal of Korean Society of Transportation
    • /
    • v.23 no.3 s.81
    • /
    • pp.95-108
    • /
    • 2005
  • The purpose of this study is to develop the ITS evaluation model using the Probabilistic Risk Analysis (PRA) methodology and to demonstrate the goodness-of-fit of the large ITS projects through the comparative analysis between DEA and PRA model. The results of this study are summarized below. First, the evaluation mode] using PRA with Monte-Carlo Simulation(MCS) and Latin-Hypercube Sampling(LHS) is developed and applied to one of ITS projects initiated by local government. The risk factors are categorized with cost, benefit and social-economic factors. Then, PDF(Probability Density Function) parameters of these factors are estimated. The log-normal distribution, beta distribution and triangular distribution are well fitted with the market and delivered price. The triangular and uniform distributions are valid in benefit data from the simulation analysis based on the several deployment scenarios. Second, the decision making rules for the risk analysis of projects for cost and economic feasibility study are suggested. The developed PRA model is applied for the Daejeon metropolitan ITS model deployment project to validate the model. The results of cost analysis shows that Deterministic Project Cost(DPC), Deterministic Total Project Cost(DTPC) is the biased percentile values of CDF produced by PRA model and this project need Contingency Budget(CB) because these values are turned out to be less than Target Value(TV;85% value), Also, this project has high risk of DTPC and DPC because the coefficient of variation(C.V) of DTPC and DPC are 4 and 15 which are less than that of DTPC(19-28) and DPC(22-107) in construction and transportation projects. The results of economic analysis shows that total system and subsystem of this project is in type II, which means the project is economically feasible with high risk. Third, the goodness-of-fit of PRA model is verified by comparing the differences of the results between PRA and DEA model. The difference of evaluation indices is up to 68% in maximum. Because of this, the deployment priority of ITS subsystems are changed in each mode1. In results. ITS evaluation model using PRA considering the project risk with the probability distribution is superior to DEA. It makes proper decision making and the risk factors estimated by PRA model can be controlled by risk management program suggested in this paper. Further research not only to build the database of deployment data but also to develop the methodologies estimating the ITS effects with PRA model is needed to broaden the usage of PRA model for the evaluation of ITS projects.

An Intelligent Intrusion Detection Model Based on Support Vector Machines and the Classification Threshold Optimization for Considering the Asymmetric Error Cost (비대칭 오류비용을 고려한 분류기준값 최적화와 SVM에 기반한 지능형 침입탐지모형)

  • Lee, Hyeon-Uk;Ahn, Hyun-Chul
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.157-173
    • /
    • 2011
  • As the Internet use explodes recently, the malicious attacks and hacking for a system connected to network occur frequently. This means the fatal damage can be caused by these intrusions in the government agency, public office, and company operating various systems. For such reasons, there are growing interests and demand about the intrusion detection systems (IDS)-the security systems for detecting, identifying and responding to unauthorized or abnormal activities appropriately. The intrusion detection models that have been applied in conventional IDS are generally designed by modeling the experts' implicit knowledge on the network intrusions or the hackers' abnormal behaviors. These kinds of intrusion detection models perform well under the normal situations. However, they show poor performance when they meet a new or unknown pattern of the network attacks. For this reason, several recent studies try to adopt various artificial intelligence techniques, which can proactively respond to the unknown threats. Especially, artificial neural networks (ANNs) have popularly been applied in the prior studies because of its superior prediction accuracy. However, ANNs have some intrinsic limitations such as the risk of overfitting, the requirement of the large sample size, and the lack of understanding the prediction process (i.e. black box theory). As a result, the most recent studies on IDS have started to adopt support vector machine (SVM), the classification technique that is more stable and powerful compared to ANNs. SVM is known as a relatively high predictive power and generalization capability. Under this background, this study proposes a novel intelligent intrusion detection model that uses SVM as the classification model in order to improve the predictive ability of IDS. Also, our model is designed to consider the asymmetric error cost by optimizing the classification threshold. Generally, there are two common forms of errors in intrusion detection. The first error type is the False-Positive Error (FPE). In the case of FPE, the wrong judgment on it may result in the unnecessary fixation. The second error type is the False-Negative Error (FNE) that mainly misjudges the malware of the program as normal. Compared to FPE, FNE is more fatal. Thus, when considering total cost of misclassification in IDS, it is more reasonable to assign heavier weights on FNE rather than FPE. Therefore, we designed our proposed intrusion detection model to optimize the classification threshold in order to minimize the total misclassification cost. In this case, conventional SVM cannot be applied because it is designed to generate discrete output (i.e. a class). To resolve this problem, we used the revised SVM technique proposed by Platt(2000), which is able to generate the probability estimate. To validate the practical applicability of our model, we applied it to the real-world dataset for network intrusion detection. The experimental dataset was collected from the IDS sensor of an official institution in Korea from January to June 2010. We collected 15,000 log data in total, and selected 1,000 samples from them by using random sampling method. In addition, the SVM model was compared with the logistic regression (LOGIT), decision trees (DT), and ANN to confirm the superiority of the proposed model. LOGIT and DT was experimented using PASW Statistics v18.0, and ANN was experimented using Neuroshell 4.0. For SVM, LIBSVM v2.90-a freeware for training SVM classifier-was used. Empirical results showed that our proposed model based on SVM outperformed all the other comparative models in detecting network intrusions from the accuracy perspective. They also showed that our model reduced the total misclassification cost compared to the ANN-based intrusion detection model. As a result, it is expected that the intrusion detection model proposed in this paper would not only enhance the performance of IDS, but also lead to better management of FNE.

Shoulder Uptake in the Bone Scintigraphy in Patients with Hemiplegic Reflex Sympathetic Dystrophy Syndrome (반신마비성 반사성교감신경 이영양증후군 환자의 골스캔상 견관절 섭취)

  • Lee, Jong-Jin;Chung, June-Key;Lee, Dong-Soo;Hong, Joon-Beom;Han, Tai-Ryoon;Lee, Myung-Chul
    • The Korean Journal of Nuclear Medicine
    • /
    • v.38 no.4
    • /
    • pp.288-293
    • /
    • 2004
  • Purpose: increased uptake of wrist and hand joints in three phase bone scintigraphy (TPBS) have been used in the detection of reflex sympathetic dystrophy syndrome (RSDS). TPBS frequently shows increased shoulder uptake in the hemiplegic RSDS patients. We investigated the significance of the shoulder uptake in the detection of these patients. Materials and Methods: Twenty three patients who had hemiplegia due to brain stroke and diagnosed as RSD were enrolled in this study (M:F=16:7, R:L=11:12). The mean age was $63{\pm}10$ yrs. Ter normal volunteer (mean age: $60{\pm}5$, M:F=1:9) data was used as control group. TPBS was performed $59{\pm}32$ days after stoke (acute stage). We obtained the count ratios of bilateral hands by drawing a region of interest (ROI) in three phase images and compared to the count ratios of shoulders in the delayed image. Hand ROI included an ipsilateral wrist. Sensitivity of detecting the affected limb was defined using the right/left count ratio of normal control. Results: Sensitivities using count ratios of hand blood flow, blood pool and delayed image were 45%, 76% and 78%, respectively. Sensitivity of shoulder count ratio was 74%. Log of right/left counts of hand delayed image and that of shoulder delayed image were correlated well with statistical significance (Spearman's R=0.824, p<0.001). Conclusion: Shoulder uptake showed good correlation with hand uptake in the delayed image of TPBS. Shoulder uptake maybe helpful in the diagnosis of reflex sympathetic dystrophy syndrome in patients with hemiplegia.

An Integrated Model based on Genetic Algorithms for Implementing Cost-Effective Intelligent Intrusion Detection Systems (비용효율적 지능형 침입탐지시스템 구현을 위한 유전자 알고리즘 기반 통합 모형)

  • Lee, Hyeon-Uk;Kim, Ji-Hun;Ahn, Hyun-Chul
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.1
    • /
    • pp.125-141
    • /
    • 2012
  • These days, the malicious attacks and hacks on the networked systems are dramatically increasing, and the patterns of them are changing rapidly. Consequently, it becomes more important to appropriately handle these malicious attacks and hacks, and there exist sufficient interests and demand in effective network security systems just like intrusion detection systems. Intrusion detection systems are the network security systems for detecting, identifying and responding to unauthorized or abnormal activities appropriately. Conventional intrusion detection systems have generally been designed using the experts' implicit knowledge on the network intrusions or the hackers' abnormal behaviors. However, they cannot handle new or unknown patterns of the network attacks, although they perform very well under the normal situation. As a result, recent studies on intrusion detection systems use artificial intelligence techniques, which can proactively respond to the unknown threats. For a long time, researchers have adopted and tested various kinds of artificial intelligence techniques such as artificial neural networks, decision trees, and support vector machines to detect intrusions on the network. However, most of them have just applied these techniques singularly, even though combining the techniques may lead to better detection. With this reason, we propose a new integrated model for intrusion detection. Our model is designed to combine prediction results of four different binary classification models-logistic regression (LOGIT), decision trees (DT), artificial neural networks (ANN), and support vector machines (SVM), which may be complementary to each other. As a tool for finding optimal combining weights, genetic algorithms (GA) are used. Our proposed model is designed to be built in two steps. At the first step, the optimal integration model whose prediction error (i.e. erroneous classification rate) is the least is generated. After that, in the second step, it explores the optimal classification threshold for determining intrusions, which minimizes the total misclassification cost. To calculate the total misclassification cost of intrusion detection system, we need to understand its asymmetric error cost scheme. Generally, there are two common forms of errors in intrusion detection. The first error type is the False-Positive Error (FPE). In the case of FPE, the wrong judgment on it may result in the unnecessary fixation. The second error type is the False-Negative Error (FNE) that mainly misjudges the malware of the program as normal. Compared to FPE, FNE is more fatal. Thus, total misclassification cost is more affected by FNE rather than FPE. To validate the practical applicability of our model, we applied it to the real-world dataset for network intrusion detection. The experimental dataset was collected from the IDS sensor of an official institution in Korea from January to June 2010. We collected 15,000 log data in total, and selected 10,000 samples from them by using random sampling method. Also, we compared the results from our model with the results from single techniques to confirm the superiority of the proposed model. LOGIT and DT was experimented using PASW Statistics v18.0, and ANN was experimented using Neuroshell R4.0. For SVM, LIBSVM v2.90-a freeware for training SVM classifier-was used. Empirical results showed that our proposed model based on GA outperformed all the other comparative models in detecting network intrusions from the accuracy perspective. They also showed that the proposed model outperformed all the other comparative models in the total misclassification cost perspective. Consequently, it is expected that our study may contribute to build cost-effective intelligent intrusion detection systems.

Open Digital Textbook for Smart Education (스마트교육을 위한 오픈 디지털교과서)

  • Koo, Young-Il;Park, Choong-Shik
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.177-189
    • /
    • 2013
  • In Smart Education, the roles of digital textbook is very important as face-to-face media to learners. The standardization of digital textbook will promote the industrialization of digital textbook for contents providers and distributers as well as learner and instructors. In this study, the following three objectives-oriented digital textbooks are looking for ways to standardize. (1) digital textbooks should undertake the role of the media for blended learning which supports on-off classes, should be operating on common EPUB viewer without special dedicated viewer, should utilize the existing framework of the e-learning learning contents and learning management. The reason to consider the EPUB as the standard for digital textbooks is that digital textbooks don't need to specify antoher standard for the form of books, and can take advantage od industrial base with EPUB standards-rich content and distribution structure (2) digital textbooks should provide a low-cost open market service that are currently available as the standard open software (3) To provide appropriate learning feedback information to students, digital textbooks should provide a foundation which accumulates and manages all the learning activity information according to standard infrastructure for educational Big Data processing. In this study, the digital textbook in a smart education environment was referred to open digital textbook. The components of open digital textbooks service framework are (1) digital textbook terminals such as smart pad, smart TVs, smart phones, PC, etc., (2) digital textbooks platform to show and perform digital contents on digital textbook terminals, (3) learning contents repository, which exist on the cloud, maintains accredited learning, (4) App Store providing and distributing secondary learning contents and learning tools by learning contents developing companies, and (5) LMS as a learning support/management tool which on-site class teacher use for creating classroom instruction materials. In addition, locating all of the hardware and software implement a smart education service within the cloud must have take advantage of the cloud computing for efficient management and reducing expense. The open digital textbooks of smart education is consdered as providing e-book style interface of LMS to learners. In open digital textbooks, the representation of text, image, audio, video, equations, etc. is basic function. But painting, writing, problem solving, etc are beyond the capabilities of a simple e-book. The Communication of teacher-to-student, learner-to-learnert, tems-to-team is required by using the open digital textbook. To represent student demographics, portfolio information, and class information, the standard used in e-learning is desirable. To process learner tracking information about the activities of the learner for LMS(Learning Management System), open digital textbook must have the recording function and the commnincating function with LMS. DRM is a function for protecting various copyright. Currently DRMs of e-boook are controlled by the corresponding book viewer. If open digital textbook admitt DRM that is used in a variety of different DRM standards of various e-book viewer, the implementation of redundant features can be avoided. Security/privacy functions are required to protect information about the study or instruction from a third party UDL (Universal Design for Learning) is learning support function for those with disabilities have difficulty in learning courses. The open digital textbook, which is based on E-book standard EPUB 3.0, must (1) record the learning activity log information, and (2) communicate with the server to support the learning activity. While the recording function and the communication function, which is not determined on current standards, is implemented as a JavaScript and is utilized in the current EPUB 3.0 viewer, ths strategy of proposing such recording and communication functions as the next generation of e-book standard, or special standard (EPUB 3.0 for education) is needed. Future research in this study will implement open source program with the proposed open digital textbook standard and present a new educational services including Big Data analysis.