DOI QR코드

DOI QR Code

Machine Learning Algorithm for Estimating Ink Usage

머신러닝을 통한 잉크 필요량 예측 알고리즘

  • Se Wook Kwon (Department of Digital Healthcare Research Korea Institute of Industrial Technology) ;
  • Young Joo Hyun (Department of Digital Healthcare Research Korea Institute of Industrial Technology) ;
  • Hyun Chul Tae (Department of Digital Healthcare Research Korea Institute of Industrial Technology)
  • 권세욱 (한국생산기술연구원 디지털헬스케어연구부문) ;
  • 현영주 (한국생산기술연구원 디지털헬스케어연구부문) ;
  • 태현철 (한국생산기술연구원 디지털헬스케어연구부문)
  • Received : 2022.11.21
  • Accepted : 2022.12.27
  • Published : 2023.03.31

Abstract

Research and interest in sustainable printing are increasing in the packaging printing industry. Currently, predicting the amount of ink required for each work is based on the experience and intuition of field workers. Suppose the amount of ink produced is more than necessary. In this case, the rest of the ink cannot be reused and is discarded, adversely affecting the company's productivity and environment. Nowadays, machine learning models can be used to figure out this problem. This study compares the ink usage prediction machine learning models. A simple linear regression model, Multiple Regression Analysis, cannot reflect the nonlinear relationship between the variables required for packaging printing, so there is a limit to accurately predicting the amount of ink needed. This study has established various prediction models which are based on CART (Classification and Regression Tree), such as Decision Tree, Random Forest, Gradient Boosting Machine, and XGBoost. The accuracy of the models is determined by the K-fold cross-validation. Error metrics such as root mean squared error, mean absolute error, and R-squared are employed to evaluate estimation models' correctness. Among these models, XGBoost model has the highest prediction accuracy and can reduce 2134 (g) of wasted ink for each work. Thus, this study motivates machine learning's potential to help advance productivity and protect the environment.

Keywords

Acknowledgement

This research was a part of the project titled 'forest science-technology R&D program (2021383A00-2223-0101)', funded by the Korea Forestry Promotion Institute (Korea National Arboretum), Korea. This study has been funded by the following project of KITECH (Korea Institute of Industrial Technology), which is "Development of AI-based packaging manufacturing cost estimation system".

References

  1. Akiba, T., S. Sano, T. Yanase, T. Ohta, and M. Koyama, Optuna: A Next-generation Hyperparameter Optimization Framework. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data mining, 2019.
  2. Ash, A. and Shwartz, M., R2: A Useful Measure of Model Performance when Predicting a Dichotomous Outcome, Statistics in Medicine, 1999, Vol. 18, No. 4, pp. 375-384. https://doi.org/10.1002/(SICI)1097-0258(19990228)18:4<375::AID-SIM20>3.0.CO;2-J
  3. Braga-Neto, U.M. and Dougherty, E.R., Is cross-validation valid for small-sample microarray classification? Bioinformatics, 2004. Vol. 20, No. 3, pp. 374-380. https://doi.org/10.1093/bioinformatics/btg419
  4. Chen, T. and Guestrin, C., Xgboost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM Sigkdd International Conference on Knowledge Discovery and Data Mining, 2016.
  5. Chen, T., He, T., Benesty, M., Khotilovich, V., Tang, Y., Cho, H., and Chen, K., Xgboost: Extreme Gradient Boosting, R package version 0.4-2, 2015, Vol. 1, , No. 4, pp. 1-4.
  6. Dyer, D., Dalzell, F., and Olegario, R., Rising Tide: Lessons from 165 Years of Brand Building at Procter & Gamble, Harvard Business School Press Boston, MA, 2004.
  7. Friede, G., Busch, T., and Bassen, A., ESG and Financial Performance: Aggregated Evidence from More than 2000 Empirical Studies, Journal of Sustainable Finance & Investment, 2015, Vol. 5, No. 4, pp. 210-233. https://doi.org/10.1080/20430795.2015.1118917
  8. Friedman, J.H., Greedy Function Approximation: A Gradient Boosting Machine, Annals of Statistics, 2001, pp. 1189-1232.
  9. Gkerekos, C., Lazakis, I., and Theotokatos, G., Machine Learning Models for Predicting Ship Main Engine Fuel Oil Consumption: A Comparative Study, Ocean Engineering, 2019, Vol. 188, p. 106282.
  10. Gladysz, B., Krystosiak, K., Ejsmont, K., Kluczek, A., and Buczacki, A., Sustainable Printing 4.0-Insights from a Polish Survey, Sustainability, 2021, Vol. 13, No. 19, p. 10916.
  11. Jordan, M.I. and Mitchell, T.M., Machine Learning: Trends, Perspectives, and Prospects, Science, 2015, Vol. 349, No. 6245, pp. 255-260. https://doi.org/10.1126/science.aaa8415
  12. Jung, H. and Kim, J.-W., A Machine Learning Approach for Mechanical Motor Fault Diagnosis, Journal of Korean Society of Industrial and Systems Engineering, 2017, Vol. 40, No. 1, pp. 57-64. https://doi.org/10.11627/jkise.2017.40.1.057
  13. Kaliya, G., Estimation of Ink Consumption Used for Printing Process, 2006.
  14. Lewis, R.J., An Introduction to Classification and Regression Tree (CART) Analysis, in Annual meeting of the society for academic emergency medicine in San Francisco, California, Citeseer, 2000.
  15. Murthy, S.K., Automatic Construction of Decision Trees from Data: A Multi-disciplinary Survey, Data Mining and Knowledge Discovery, 1998, Vol. 2, No. 4, pp. 345-389. https://doi.org/10.1023/A:1009744630224
  16. Natekin, A. and Knoll, A., Gradient Boosting Machines, a Tutorial, Frontiers in Neurorobotics, 2013, Vol. 7, p. 21.
  17. Neter, J., Kutner, M.H., Nachtsheim, C.J., and Wasserman, W., Applied Linear Statistical Models, 1996.
  18. Razi, M.A. and Athappilly, K., A Comparative Predictive Analysis of Neural Networks (NNs), Nonlinear Regression and Classification and Regression Tree (CART) Models, Expert Systems with Applications, 2005, Vol. 29, No. 1, pp. 65-74. https://doi.org/10.1016/j.eswa.2005.01.006
  19. Ruggles, P.K., Printing Estimating: Costing and Pricing Print and Digital Media, PIA/GATFPress, 2008.
  20. Shehadeh, A., Alshboul, O., Al Mamlook, R.E., and Hamedat, O., Machine Learning Models for Predicting the Residual Value of Heavy Construction Equipment: An Evaluation of Modified Decision Tree, LightGBM, and XGBoost Regression, Automation in Construction, 2021, Vol. 129, p. 103827.
  21. Song, Y.-Y. and Ying, L., Decision Tree Methods: Applications for Classification and Prediction, Shanghai Archives of Psychiatry, 2015, Vol. 27, No. 2, p. 130.
  22. Svetnik, V., Liaw, A., Tong, C., Culberson, J.C., Sheridan, R.P., and Feuston, B.P., Random Forest: A Classification and Regression Tool for Compound Classification and QSAR Modeling, Journal of Chemical Information and Computer Sciences, 2003, Vol. 43, No. 6, pp. 1947-1958. https://doi.org/10.1021/ci034160g
  23. Uyanik, G.K. and Guler, N., A Study on Multiple Linear Regression Analysis, Procedia-Social and Behavioral Sciences, 2013, Vol. 106, pp. 234-240. https://doi.org/10.1016/j.sbspro.2013.12.027