• Title/Summary/Keyword: data science department

Search Result 26,696, Processing Time 0.055 seconds

Comparing Results of Classification Techniques Regarding Heart Disease Diagnosing

  • AL badr, Benan Abdullah;AL ghezzi, Raghad Suliman;AL moqhem, ALjohara Suliman;Eljack, Sarah
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.5
    • /
    • pp.135-142
    • /
    • 2022
  • Despite global medical advancements, many patients are misdiagnosed, and more people are dying as a result. We must now develop techniques that provide the most accurate diagnosis of heart disease based on recorded data. To help immediate and accurate diagnose of heart disease, several data mining methods are accustomed to anticipating the disease. A large amount of clinical information offered data mining strategies to uncover the hidden pattern. This paper presents, comparison between different classification techniques, we applied on the same dataset to see what is the best. In the end, we found that the Random Forest algorithm had the best results.

A Study on Quality Checking of National Scholar Content DB

  • Kim, Byung-Kyu;Choi, Seon-Hee;Kim, Jay-Hoon;You, Beom-Jong
    • International Journal of Contents
    • /
    • v.6 no.3
    • /
    • pp.1-4
    • /
    • 2010
  • The national management and retrieval service of the national scholar Content DB are very important. High quality content can improve the user's utilization and satisfaction and be a strong base for both the citation index creation and the calculation of journal impact factors. Therefore, the system is necessary to check data quality effectively. We have closely studied and developed a webbased data quality checking system that will support anything from raw digital data to its automatic validation as well as hands-on validation, all of which will be discussed in this paper.

Test for Independence in Bivariate Weibull Model under Bivariate Random Censorship

  • Cho, Jang-Sik;Cho, Kil-Ho;Lee, Woo-Dong
    • Journal of the Korean Data and Information Science Society
    • /
    • v.14 no.4
    • /
    • pp.789-797
    • /
    • 2003
  • In this paper, we consider two components system which have bivariate weibull model with bivariate random censored data. We proposed large sample test for independence based on maximum likelihood estimator and relative frequency estimator, respectively. Also we derive asymptotic properties for the large sample tests and present a numerical study.

  • PDF

Personal Data Security in Recruitment Platforms

  • Bajoudah, Alya'a;AlSuwat, Hatim
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.6
    • /
    • pp.310-318
    • /
    • 2022
  • Job offers have become more widespread and it has become easier and faster to apply for jobs through electronic recruitment platforms. In order to increase the protection of the data that is attached to the recruitment platforms. In this research, a proposed model was created through the use of hybrid encryption, which is used through the following algorithms: AES,Twofish,. This proposed model proved the effectiveness of using hybrid encryption in protecting personal data.

Improved User Privacy in SocialNetworks Based on Hash Function

  • Alrwuili, Kawthar;Hendaoui, Saloua
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.1
    • /
    • pp.97-104
    • /
    • 2022
  • In recent years, data privacy has become increasingly important. The goal of network cryptography is to protect data while it is being transmitted over the internet or a network. Social media and smartphone apps collect a lot of personal data which if exposed, might be damaging to privacy. As a result, sensitive data is exposed and data is shared without the data owner's consent. Personal Information is one of the concerns in data privacy. Protecting user data and sensitive information is the first step to keeping user data private. Many applications user data can be found on other websites. In this paper, we discuss the issue of privacy and suggest a mechanism for keeping user data hidden in other applications.

Bootstrap Confidence Intervals for Regression Coefficients under Censored Data

  • Cho, Kil-Ho;Jeong, Seong-Hwa
    • Journal of the Korean Data and Information Science Society
    • /
    • v.13 no.2
    • /
    • pp.355-363
    • /
    • 2002
  • Using the Buckley-James method, we construct bootstrap confidence intervals for the regression coefficients under the censored data. And we compare these confidence intervals in terms of the coverage probabilities and the expected confidence interval lengths through Monte Carlo simulation.

  • PDF

Using Machine Learning Algorithms for Housing Price Prediction: The Case of Islamabad Housing Data

  • Imran, Imran;Zaman, Umar;Waqar, Muhammad;Zaman, Atif
    • Soft Computing and Machine Intelligence
    • /
    • v.1 no.1
    • /
    • pp.11-23
    • /
    • 2021
  • House price prediction is a significant financial decision for individuals working in the housing market as well as for potential buyers. From investment to buying a house for residence, a person investing in the housing market is interested in the potential gain. This paper presents machine learning algorithms to develop intelligent regressions models for House price prediction. The proposed research methodology consists of four stages, namely Data Collection, Pre Processing the data collected and transforming it to the best format, developing intelligent models using machine learning algorithms, training, testing, and validating the model on house prices of the housing market in the Capital, Islamabad. The data used for model validation and testing is the asking price from online property stores, which provide a reasonable estimate of the city housing market. The prediction model can significantly assist in the prediction of future housing prices in Pakistan. The regression results are encouraging and give promising directions for future prediction work on the collected dataset.

Parameter Estimation and Comparison for SRGMs and ARIMA Model in Software Failure Data

  • Song, Kwang Yoon;Chang, In Hong;Lee, Dong Su
    • Journal of Integrative Natural Science
    • /
    • v.7 no.3
    • /
    • pp.193-199
    • /
    • 2014
  • As the requirement on the quality of the system has increased, the reliability is very important part in terms of enhance stability and to provide high quality services to customers. Many statistical models have been developed in the past years for the estimation of software reliability. We consider the functions for NHPP software reliability model and time series model in software failure data. We estimate parameters for the proposed models from three data sets. The values of SSE and MSE is presented from three data sets. We compare the predicted number of faults with the actual three data sets using the NHPP software reliability model and time series model.

Forecasting of Stream Qualities in Gumho River by Exponential Smoothing at Gumho2 Measurement Point using Monthly Time Series Data

  • Song, Phil-Jun;Lee, Bo-Ra;Kim, Jin-Yong;Kim, Jong-Tae
    • Journal of the Korean Data and Information Science Society
    • /
    • v.18 no.3
    • /
    • pp.609-617
    • /
    • 2007
  • The goal of this study is to forecast the trend of stream quality and to suggest some policy alternatives in Gumbo river. It used the five different monthly time series data such as BOD, COD, T-N and EC of the nine of Gumbo River measurement points from Jan. 1998 to Dec. 2006. Water pollution is serious at Gumbo2 and Palgeo stream measurement points. BOD, COD, T-N and EC data are analyzed with the exponential smoothing model and the trend is forecasted until Dec. 2009.

  • PDF

Modelling Data Flow in Smart Claim Processing Using Time Invariant Petri Net with Fixed Input Data

  • Amponsah, Anokye Acheampong;Adekoya, Adebayo Felix;Weyori, Benjamin Asubam
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.2
    • /
    • pp.413-423
    • /
    • 2022
  • The NHIS provides free or highly subsidized healthcare to all people by providing financial fortification. However, the financial sustainability of the scheme is threatened by numerous factors. Therefore, this work sought to provide a solution to process claims intelligently. The provided Petri net model demonstrated successful data flow among the various participant. For efficiency, scalability, and performance two main subsystems were modelled and integrated - data input and claims processing subsystems. We provided smart claims processing algorithm that has a simple and efficient error detection method. The complexity of the main algorithm is good but that of the error detection is excellent when compared to literature. Performance indicates that the model output is reachable from input and the token delivery rate is promising.