• Title/Summary/Keyword: Sensitive data

Search Result 2,488, Processing Time 0.028 seconds

Analysis of Time Domain Active Sensing Data from CX-100 Wind Turbine Blade Fatigue Tests for Damage Assessment

  • Choi, Mijin;Jung, Hwee Kwon;Taylor, Stuart G.;Farinholt, Kevin M.;Lee, Jung-Ryul;Park, Gyuhae
    • Journal of the Korean Society for Nondestructive Testing
    • /
    • v.36 no.2
    • /
    • pp.93-101
    • /
    • 2016
  • This paper presents the results obtained using time-series-based methods for structural damage assessment. The methods are applied to a wind turbine blade structure subjected to fatigue loads. A 9 m CX-100 (carbon experimental 100 kW) blade is harmonically excited at its first natural frequency to introduce a failure mode. Consequently, a through-thickness fatigue crack is visually identified at 8.5 million cycles. The time domain data from the piezoelectric active-sensing techniques are measured during the fatigue loadings and used to detect incipient damage. The damage-sensitive features, such as the first four moments and a normality indicator, are extracted from the time domain data. Time series autoregressive models with exogenous inputs are also implemented. These features could efficiently detect a fatigue crack and are less sensitive to operational variations than the other methods.

A New Measure of Process Capability for Non-Normal Process : $C_{psk}$ (비정규 공정에 대한 공정능력의 새로운 측도: $C_{psk}$)

  • 김홍준;송서일
    • Journal of Korean Society for Quality Management
    • /
    • v.26 no.1
    • /
    • pp.48-60
    • /
    • 1998
  • This paper proposes a fourth generation index $C_{psk}$, constructed from $C_{psk}$, by introducing the factor|$\mu$-T| in the numerator as an extra penalty for the departure of the process mean from the preassigned target value T. The motivation behind the introduction of $C_{psk}$ is that when $T\neqM$ process shifts away from target are evaluated without respect to direction. All indices that are now in use assume normally distributed data, and any use of the indices on non-normal data results in inaccurate capability measurements. In this paper, a new process capability index $C_{psk}$ is introduced for non-normal process. The Pearson curve and the Johnson curve are selected for capability index calculation and data modeling the normal-based index $C_{psk}$ is used as the model for non-normal process. A significant result of this research find that the ranking of the six indices, $C_{p}$, $C_{pk}$, $C_{pm}$, ${C^*}_{psk}$, $C_{pmk}$, $C_{psk}$in terms of sensitivity to departure of the process median from the target value from the most sensitive one up to the least sensitive are $C_{psk}$, $C_{pmk}$, ${C^*}_{psk}$,$C_{pm}$, $C_{pk}$, $C_{p}$.

  • PDF

Motivational Factors of Implementing Corporate Social and Environmental Reporting and Its Impact on Performance

  • INDRASARI, Arum;NUGRAHENI, Peni;HAMZAH, Noradiva;MAELAH, Ruhanita
    • The Journal of Asian Finance, Economics and Business
    • /
    • v.8 no.2
    • /
    • pp.883-892
    • /
    • 2021
  • The issue of environmental crisis encourages companies to develop strategies and programs which incorporate social and environmental considerations into their processes. The objectives of this study are to identify the strategies used in implementing corporate social and environmental reporting (CSER) and to investigate the impact of these strategies on organization performance. This study uses as its sample companies listed on the Indonesia Stock Exchange (IDX) and engaged in environmentally sensitive business activities and applies content analysis to their annual reports. The data used in the study is secondary data in the form of annual and sustainability reports of companies, and primary data in the form of interviews. The results show that companies use both reactive and proactive strategies in reporting their social and environmental activities. The study also identifies the impacts of such reporting on both the financial and non-financial performances of the investigated companies. The study contributes to the social and environmental accounting literature by exploring the motivations and strategies of companies in their CSER. The empirical results will provide important insights into the influence of the strategies employed by companies in their corporate social and environmental reporting and the impacts of such strategies on organizational performance.

Machine Learning-Based Reversible Chaotic Masking Method for User Privacy Protection in CCTV Environment

  • Jimin Ha;Jungho Kang;Jong Hyuk Park
    • Journal of Information Processing Systems
    • /
    • v.19 no.6
    • /
    • pp.767-777
    • /
    • 2023
  • In modern society, user privacy is emerging as an important issue as closed-circuit television (CCTV) systems increase rapidly in various public and private spaces. If CCTV cameras monitor sensitive areas or personal spaces, they can infringe on personal privacy. Someone's behavior patterns, sensitive information, residence, etc. can be exposed, and if the image data collected from CCTV is not properly protected, there can be a risk of data leakage by hackers or illegal accessors. This paper presents an innovative approach to "machine learning based reversible chaotic masking method for user privacy protection in CCTV environment." The proposed method was developed to protect an individual's identity within CCTV images while maintaining the usefulness of the data for surveillance and analysis purposes. This method utilizes a two-step process for user privacy. First, machine learning models are trained to accurately detect and locate human subjects within the CCTV frame. This model is designed to identify individuals accurately and robustly by leveraging state-of-the-art object detection techniques. When an individual is detected, reversible chaos masking technology is applied. This masking technique uses chaos maps to create complex patterns to hide individual facial features and identifiable characteristics. Above all, the generated mask can be reversibly applied and removed, allowing authorized users to access the original unmasking image.

Re-anonymization Technique for Dynamic Data Using Decision Tree Based Machine Learning (결정트리 기반의 기계학습을 이용한 동적 데이터에 대한 재익명화기법)

  • Kim, Young Ki;Hong, Choong Seon
    • Journal of KIISE
    • /
    • v.44 no.1
    • /
    • pp.21-26
    • /
    • 2017
  • In recent years, new technologies such as Internet of Things, Cloud Computing and Big Data are being widely used. And the type and amount of data is dramatically increasing. This makes security an important issue. In terms of leakage of sensitive personal information. In order to protect confidential information, a method called anonymization is used to remove personal identification elements or to substitute the data to some symbols before distributing and sharing the data. However, the existing method performs anonymization by generalizing the level of quasi-identifier hierarchical. It requires a higher level of generalization in case where k-anonymity is not satisfied since records in data table are either added or removed. Loss of information is inevitable from the process, which is one of the factors hindering the utility of data. In this paper, we propose a novel anonymization technique using decision tree based machine learning to improve the utility of data by minimizing the loss of information.

Application of EDA Techniques for Estimating Rainfall Quantiles (확률강우량 산정을 위한 EDA 기법의 적용)

  • Park, Hyunkeun;Oh, Sejeong;Yoo, Chulsang
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.29 no.4B
    • /
    • pp.319-328
    • /
    • 2009
  • This study quantified the data by applying the EDA techniques considering the data structure, and the results were then used for the frequency analysis. Although traditional methods based on the method of moments provide very sensitive statistics to the extreme values, the EDA techniques have an advantage of providing very stable statistics with their small variation. For the application of the EDA techniques to the frequency analysis, it is necessary to normalization transform and inverse-transform to conserve the skewness of the raw data. That is, it is necessary to transform the raw data to make the data follow the normal distribution, to estimate the statistics by applying the EDA techniques, and then finally to inverse-transform the statistics of transformed data. These statistics decided are then applied for the frequency analysis with a given probability density function. This study analyzed the annual maxima one hour rainfall data at Seoul and Pohang stations. As a result, it was found that more stable rainfall quantiles, which were also less sensitive to extreme values, could be estimated by applying the EDA techniques. This methodology may be effectively used for the frequency analysis of rainfall at stations with especially high annual variations of rainfall due to climate change, etc.

Randomized Response Model with Discrete Quantitative Attribute by Three-Stage Cluster Sampling

  • Lee, Gi-Sung;Hong, Ki-Hak
    • Journal of the Korean Data and Information Science Society
    • /
    • v.14 no.4
    • /
    • pp.1067-1082
    • /
    • 2003
  • In this paper, we propose a randomized response model with discrete quantitative attribute by three-stage cluster sampling for obtaining discrete quantitative data by using the Liu & Chow model(1976), when the population was made up of sensitive discrete quantitative clusters. We obtain the minimum variance by calculating the optimum number of fsu, ssu, tsu under the some given constant cost. And we obtain the minimum cost under the some given accuracy.

  • PDF

A Robust Heteroscadastic Test for ARCH Models

  • Kim, Sahm-Yeong
    • Journal of the Korean Data and Information Science Society
    • /
    • v.15 no.2
    • /
    • pp.441-447
    • /
    • 2004
  • Li and Mak (1994) developed a test statistic for detecting the non-linearity and the heteroscedasticity of the time series data. But it is well known that the test statistic may be very sensitive in case of heavy-tailed distributions of the errors. Jiang et al.(2001) suggested the robust method for ARCH models but the calculation procedures for the estimation are very complicated. We suggested the robust method based on Huber's function and our method works quite well rater than the Li and Mak(1994). Also our method is relatively easy to calculate the test statistic.

  • PDF

Privacy Preserving Source Based Deduplication In Cloud Storage (클라우드 스토리지 상에서의 프라이버시 보존형 소스기반 중복데이터 제거기술)

  • Park, Cheolhee;Hong, Dowon;Seo, Changho;Chang, Ku-Young
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.25 no.1
    • /
    • pp.123-132
    • /
    • 2015
  • In cloud storage, processing the duplicated data, namely deduplication, is necessary technology to save storage space. Users who store sensitive data in remote storage want data be encrypted. However Cloud storage server do not detect duplication of conventionally encrypted data. To solve this problem, Convergent Encryption has been proposed. But it inherently have weakness due to brute-force attack. On the other hand, to save storage space as well as save bandwidths, client-side deduplication have been applied. Recently, various client-side deduplication technology has been proposed. However, this propositions still cannot solve the security problem. In this paper, we suggest a secure source-based deduplication technology, which encrypt data to ensure the confidentiality of sensitive data and apply proofs of ownership protocol to control access to the data, from curious cloud server and malicious user.

Analysis and Estimation for Market Share of Biologics based on Google Trends Big Data (구글 트렌드 빅데이터를 통한 바이오의약품의 시장 점유율 분석과 추정)

  • Bong, Ki Tae;Lee, Heesang
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.43 no.2
    • /
    • pp.14-24
    • /
    • 2020
  • Google Trends is a useful tool not only for setting search periods, but also for providing search volume to specific countries, regions, and cities. Extant research showed that the big data from Google Trends could be used for an on-line market analysis of opinion sensitive products instead of an on-site survey. This study investigated the market share of tumor necrosis factor-alpha (TNF-α) inhibitor, which is in a great demand pharmaceutical product, based on big data analysis provided by Google Trends. In this case study, the consumer interest data from Google Trends were compared to the actual product sales of Top 3 TNF-α inhibitors (Enbrel, Remicade, and Humira). A correlation analysis and relative gap were analyzed by statistical analysis between sales-based market share and interest-based market share. Besides, in the country-specific analysis, three major countries (USA, Germany, and France) were selected for market share analysis for Top 3 TNF-α inhibitors. As a result, significant correlation and similarity were identified by data analysis. In the case of Remicade's biosimilars, the consumer interest in two biosimilar products (Inflectra and Renflexis) increased after the FDA approval. The analytical data showed that Google Trends is a powerful tool for market share estimation for biosimilars. This study is the first investigation in market share analysis for pharmaceutical products using Google Trends big data, and it shows that global and regional market share analysis and estimation are applicable for the interest-sensitive products.