• Title/Summary/Keyword: a conditional probability

Search Result 294, Processing Time 0.03 seconds

Context-Based Minimum MSE Prediction and Entropy Coding for Lossless Image Coding

  • Musik-Kwon;Kim, Hyo-Joon;Kim, Jeong-Kwon;Kim, Jong-Hyo;Lee, Choong-Woong
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1999.06a
    • /
    • pp.83-88
    • /
    • 1999
  • In this paper, a novel gray-scale lossless image coder combining context-based minimum mean squared error (MMSE) prediction and entropy coding is proposed. To obtain context of prediction, this paper first defines directional difference according to sharpness of edge and gradients of localities of image data. Classification of 4 directional differences forms“geometry context”model which characterizes two-dimensional general image behaviors such as directional edge region, smooth region or texture. Based on this context model, adaptive DPCM prediction coefficients are calculated in MMSE sense and the prediction is performed. The MMSE method on context-by-context basis is more in accord with minimum entropy condition, which is one of the major objectives of the predictive coding. In entropy coding stage, context modeling method also gives useful performance. To reduce the statistical redundancy of the residual image, many contexts are preset to take full advantage of conditional probability in entropy coding and merged into small number of context in efficient way for complexity reduction. The proposed lossless coding scheme slightly outperforms the CALIC, which is the state-of-the-art, in compression ratio.

Recurrent Neural Network Modeling of Etch Tool Data: a Preliminary for Fault Inference via Bayesian Networks

  • Nawaz, Javeria;Arshad, Muhammad Zeeshan;Park, Jin-Su;Shin, Sung-Won;Hong, Sang-Jeen
    • Proceedings of the Korean Vacuum Society Conference
    • /
    • 2012.02a
    • /
    • pp.239-240
    • /
    • 2012
  • With advancements in semiconductor device technologies, manufacturing processes are getting more complex and it became more difficult to maintain tighter process control. As the number of processing step increased for fabricating complex chip structure, potential fault inducing factors are prevail and their allowable margins are continuously reduced. Therefore, one of the key to success in semiconductor manufacturing is highly accurate and fast fault detection and classification at each stage to reduce any undesired variation and identify the cause of the fault. Sensors in the equipment are used to monitor the state of the process. The idea is that whenever there is a fault in the process, it appears as some variation in the output from any of the sensors monitoring the process. These sensors may refer to information about pressure, RF power or gas flow and etc. in the equipment. By relating the data from these sensors to the process condition, any abnormality in the process can be identified, but it still holds some degree of certainty. Our hypothesis in this research is to capture the features of equipment condition data from healthy process library. We can use the health data as a reference for upcoming processes and this is made possible by mathematically modeling of the acquired data. In this work we demonstrate the use of recurrent neural network (RNN) has been used. RNN is a dynamic neural network that makes the output as a function of previous inputs. In our case we have etch equipment tool set data, consisting of 22 parameters and 9 runs. This data was first synchronized using the Dynamic Time Warping (DTW) algorithm. The synchronized data from the sensors in the form of time series is then provided to RNN which trains and restructures itself according to the input and then predicts a value, one step ahead in time, which depends on the past values of data. Eight runs of process data were used to train the network, while in order to check the performance of the network, one run was used as a test input. Next, a mean squared error based probability generating function was used to assign probability of fault in each parameter by comparing the predicted and actual values of the data. In the future we will make use of the Bayesian Networks to classify the detected faults. Bayesian Networks use directed acyclic graphs that relate different parameters through their conditional dependencies in order to find inference among them. The relationships between parameters from the data will be used to generate the structure of Bayesian Network and then posterior probability of different faults will be calculated using inference algorithms.

  • PDF

Early diagnosis of jaw osteomyelitis by easy digitalized panoramic analysis

  • Park, Moo Soung;Eo, Mi Young;Myoung, Hoon;Kim, Soung Min;Lee, Jong Ho
    • Maxillofacial Plastic and Reconstructive Surgery
    • /
    • v.41
    • /
    • pp.6.1-6.10
    • /
    • 2019
  • Background: Osteomyelitis is an intraosseous inflammatory disease characterized by progressive inflammatory osteoclasia and ossification. The use of quantitative analysis to assist interpretation of osteomyelitis is increasingly being considered. The objective of this study was to perform early diagnosis of osteomyelitis on digital panoramic radiographs using basic functions provided by picture archiving and communication system (PACS), a program used to show radiographic images. Methods: This study targeted a total of 95 patients whose symptoms were confirmed as osteomyelitis under clinical, radiologic, pathological diagnosis over 11 years from 2008 to 2017. Five categorized patients were osteoradionecrosis, bisphosphonate-related osteonecrosis of jaw (BRONJ, suppurative and sclerosing type), and bacterial osteomyelitis (suppurative and sclerosing type), and the control group was 117 randomly sampled. The photographic density in a certain area of the digital panoramic radiograph was determined and compared using the "measure area rectangle," one of the basic PACS functions in INFINITT PACS® (INFINITT Healthcare, Seoul, South Korea). A conditional inference tree, one type of decision making tree, was generated with the program R for statistical analysis with SPSS®. Results: In the conditional inference tree generated from the obtained data, cases where the difference in average value exceeded 54.49 and the difference in minimum value was less than 54.49 and greater than 12.81 and the difference in minimum value exceeded 39 were considered suspicious of osteomyelitis. From these results, the disease could be correctly classified with a probability of 88.1%. There was no difference in photographic density value of BRONJ and bacterial osteomyelitis; therefore, it was not possible to classify BRONJ and bacterial osteomyelitis by quantitative analysis of panoramic radiographs based on existing research. Conclusions: This study demonstrates that it is feasible to measure photographic density using a basic function in PACS and apply the data to assist in the diagnosis of osteomyelitis.

Time-series Mapping and Uncertainty Modeling of Environmental Variables: A Case Study of PM10 Concentration Mapping (시계열 환경변수 분포도 작성 및 불확실성 모델링: 미세먼지(PM10) 농도 분포도 작성 사례연구)

  • Park, No-Wook
    • Journal of the Korean earth science society
    • /
    • v.32 no.3
    • /
    • pp.249-264
    • /
    • 2011
  • A multi-Gaussian kriging approach extended to space-time domain is presented for uncertainty modeling as well as time-series mapping of environmental variables. Within a multi-Gaussian framework, normal score transformed environmental variables are first decomposed into deterministic trend and stochastic residual components. After local temporal trend models are constructed, the parameters of the models are estimated and interpolated in space. Space-time correlation structures of stationary residual components are quantified using a product-sum space-time variogram model. The ccdf is modeled at all grid locations using this space-time variogram model and space-time kriging. Finally, e-type estimates and conditional variances are computed from the ccdf models for spatial mapping and uncertainty analysis, respectively. The proposed approach is illustrated through a case of time-series Particulate Matter 10 ($PM_{10}$) concentration mapping in Incheon Metropolitan city using monthly $PM_{10}$ concentrations at 13 stations for 3 years. It is shown that the proposed approach would generate reliable time-series $PM_{10}$ concentration maps with less mean bias and better prediction capability, compared to conventional spatial-only ordinary kriging. It is also demonstrated that the conditional variances and the probability exceeding a certain thresholding value would be useful information sources for interpretation.

A Selection of Threshold for the Generalized Hough Transform: A Probabilistic Approach (일반화된 허프변환의 임계값 선택을 위한 확률적 접근방식)

  • Chang, Ji Y.
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.51 no.1
    • /
    • pp.161-171
    • /
    • 2014
  • When the Hough transform is applied to identify an instance of a given model, the output is typically a histogram of votes cast by a set of image features into a parameter space. The next step is to threshold the histogram of counts to hypothesize a given match. The question is "What is a reasonable choice of the threshold?" In a standard implementation of the Hough transform, the threshold is selected heuristically, e.g., some fraction of the highest cell count. Setting the threshold too low can give rise to a false alarm of a given shape(Type I error). On the other hand, setting the threshold too high can result in mis-detection of a given shape(Type II error). In this paper, we derive two conditional probability functions of cell counts in the accumulator array of the generalized Hough transform(GHough), that can be used to select a scientific threshold at the peak detection stage of the Ghough.

Fatigue life prediction based on Bayesian approach to incorporate field data into probability model

  • An, Dawn;Choi, Joo-Ho;Kim, Nam H.;Pattabhiraman, Sriram
    • Structural Engineering and Mechanics
    • /
    • v.37 no.4
    • /
    • pp.427-442
    • /
    • 2011
  • In fatigue life design of mechanical components, uncertainties arising from materials and manufacturing processes should be taken into account for ensuring reliability. A common practice is to apply a safety factor in conjunction with a physics model for evaluating the lifecycle, which most likely relies on the designer's experience. Due to conservative design, predictions are often in disagreement with field observations, which makes it difficult to schedule maintenance. In this paper, the Bayesian technique, which incorporates the field failure data into prior knowledge, is used to obtain a more dependable prediction of fatigue life. The effects of prior knowledge, noise in data, and bias in measurements on the distribution of fatigue life are discussed in detail. By assuming a distribution type of fatigue life, its parameters are identified first, followed by estimating the distribution of fatigue life, which represents the degree of belief of the fatigue life conditional to the observed data. As more data are provided, the values will be updated to reduce the credible interval. The results can be used in various needs such as a risk analysis, reliability based design optimization, maintenance scheduling, or validation of reliability analysis codes. In order to obtain the posterior distribution, the Markov Chain Monte Carlo technique is employed, which is a modern statistical computational method which effectively draws the samples of the given distribution. Field data of turbine components are exploited to illustrate our approach, which counts as a regular inspection of the number of failed blades in a turbine disk.

Elliptical Clustering with Incremental Growth and its Application to Skin Color Region Segmentation (점증적으로 증가하는 타원형 군집화 : 피부색 영역 검출에의 적용)

  • Lee Kyoung-Mi
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.9
    • /
    • pp.1161-1170
    • /
    • 2004
  • This paper proposes to segment skin color areas using a clustering algorithm. Most of previously proposed clustering algorithms have some difficulties, since they generally detect hyperspherical clusters, run in a batch mode, and predefine a number of clusters. In this paper, we use a well-known elliptical clustering algorithm, an EM algorithm, and modify it to learn on-line and find automatically the number of clusters, called to an EAM algorithm. The effectiveness of the EAM algorithm is demonstrated on a task of skin color region segmentation. Experimental results present the EAM algorithm automatically finds a right number of clusters in a given image without any information on the number. Comparing with the EM algorithm, we achieved better segmentation results with the EAM algorithm. Successful results were achieved to detect and segment skin color regions using a conditional probability on a region. Also, we applied to classify images with persons and got good classification results.

The Evaluation Model for Natural Resource Conservation Areas - Focused on Site Selection for the National Trust - (자연자원 보전지역의 평가모형 - 내셔널 트러스트 후보지 선정을 중심으로 -)

  • 유주한;정성관
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.30 no.2
    • /
    • pp.39-49
    • /
    • 2002
  • The purpose of this study is to propose an objective and rational methodology for the selection of proposed sites far the National Trust(NT), which is the new alterative proposal far the conservation of natural environments destroyed by injudicious land development and economic growth. That is to enforce many analysis for the effective estimation of rare ecological and landscape resources and to propose a model based on estimation and united indicators. Using the estimative model, we apply it to the selection of the proposed site in micro scale and simultaneously offer the basic methodology of effective and systematic land conservation in macro scale. The results of this study are as follows: 1) The results of analysis for the reliability of estimative items and indicators, presented no problem in that the coefficient of reliability was over 0.7. 2) The correlation measure of the estimative indicator indicated that 'succession'and 'regenerating restorability' were highly correlative in the item of plants. Another three items showed a tendency to be alike. 3) The results of factor analysis on the characteristics of indicators, classified plants into four categories including a stable factor. The item of animals was classified as a stable and rare factor. The item of landscape was classified as a physical and mental factor and the environment as a pollutional and conditional factor. 4) The model of estimation created through factor analysis was valid for the approval of the regression model because significant probability was 0.00. When we consider the NT proposed site as a complex body that is composed of diverse natural and manmade resources, certainly the synthetic methodology of estimation is needed. If these studies are carried out, NT sites will be selected more rationally and effectively than at present. Consequently, they have the potential to play a core role of natural ecosystem conservation in Korea.

Online condition assessment of high-speed trains based on Bayesian forecasting approach and time series analysis

  • Zhang, Lin-Hao;Wang, You-Wu;Ni, Yi-Qing;Lai, Siu-Kai
    • Smart Structures and Systems
    • /
    • v.21 no.5
    • /
    • pp.705-713
    • /
    • 2018
  • High-speed rail (HSR) has been in operation and development in many countries worldwide. The explosive growth of HSR has posed great challenges for operation safety and ride comfort. Among various technological demands on high-speed trains, vibration is an inevitable problem caused by rail/wheel imperfections, vehicle dynamics, and aerodynamic instability. Ride comfort is a key factor in evaluating the operational performance of high-speed trains. In this study, online monitoring data have been acquired from an in-service high-speed train for condition assessment. The measured dynamic response signals at the floor level of a train cabin are processed by the Sperling operator, in which the ride comfort index sequence is used to identify the train's operation condition. In addition, a novel technique that incorporates salient features of Bayesian inference and time series analysis is proposed for outlier detection and change detection. The Bayesian forecasting approach enables the prediction of conditional probabilities. By integrating the Bayesian forecasting approach with time series analysis, one-step forecasting probability density functions (PDFs) can be obtained before proceeding to the next observation. The change detection is conducted by comparing the current model and the alternative model (whose mean value is shifted by a prescribed offset) to determine which one can well fit the actual observation. When the comparison results indicate that the alternative model performs better, then a potential change is detected. If the current observation is a potential outlier or change, Bayes factor and cumulative Bayes factor are derived for further identification. A significant change, if identified, implies that there is a great alteration in the train operation performance due to defects. In this study, two illustrative cases are provided to demonstrate the performance of the proposed method for condition assessment of high-speed trains.

Learning Distribution Graphs Using a Neuro-Fuzzy Network for Naive Bayesian Classifier (퍼지신경망을 사용한 네이브 베이지안 분류기의 분산 그래프 학습)

  • Tian, Xue-Wei;Lim, Joon S.
    • Journal of Digital Convergence
    • /
    • v.11 no.11
    • /
    • pp.409-414
    • /
    • 2013
  • Naive Bayesian classifiers are a powerful and well-known type of classifiers that can be easily induced from a dataset of sample cases. However, the strong conditional independence assumptions can sometimes lead to weak classification performance. Normally, naive Bayesian classifiers use Gaussian distributions to handle continuous attributes and to represent the likelihood of the features conditioned on the classes. The probability density of attributes, however, is not always well fitted by a Gaussian distribution. Another eminent type of classifier is the neuro-fuzzy classifier, which can learn fuzzy rules and fuzzy sets using supervised learning. Since there are specific structural similarities between a neuro-fuzzy classifier and a naive Bayesian classifier, the purpose of this study is to apply learning distribution graphs constructed by a neuro-fuzzy network to naive Bayesian classifiers. We compare the Gaussian distribution graphs with the fuzzy distribution graphs for the naive Bayesian classifier. We applied these two types of distribution graphs to classify leukemia and colon DNA microarray data sets. The results demonstrate that a naive Bayesian classifier with fuzzy distribution graphs is more reliable than that with Gaussian distribution graphs.