Optimal Raw Material Inventory Analysis Using Markov Decision Process with Policy Iteration Method

Authors

  • Ikhsan Maulidi Department of Mathematics, Syiah Kuala University
  • Radhiah Radhiah Department of Mathematics, Syiah Kuala University
  • Chintya Hayati Department of Mathematics, Syiah Kuala University
  • Vina Apriliani UIN Ar-Raniry Banda Aceh

DOI:

https://doi.org/10.31764/jtam.v6i3.8563

Keywords:

Inventory, Markov Decision Process, Policy Iteration Method

Abstract

Inventory of raw materials is a big deal in every production process, both in company production and home business production. In order to meet consumer demand, a business must be able to determine the amount of inventory that should be provided. The purpose of this research is to choose an alternative selection of ordering raw materials that produce the maximum amount of raw materials with minimum costs. The raw material referred to in this study is pandan leaves used to make pandan mats. Analysis of raw material inventory used in this research was the Markov decision process with the policy iteration method by considering the discount factor. From the analysis conducted, it is obtained alternative policies that must be taken by producers to meet raw materials with minimum costs. The results of this study can be a consideration for business actors in the study location in deciding the optimal ordering policy that should be taken to obtain the minimum operational cost.

 

References

Alla, A., Falcone, M., & Kalise, D. (2015). An efficient policy iteration algorithm for dynamic programming equations. SIAM Journal on Scientific Computing, 37(1), A181–A200. https://doi.org/https://doi.org/10.1137/130932284

Bennett, C. C., & Hauser, K. (2013). Artificial intelligence framework for simulating clinical decision-making: A Markov decision process approach. Artificial Intelligence in Medicine, 57(1), 9–19. https://doi.org/10.1016/j.artmed.2012.12.003

Ching, W. K., & Ng, M. K. (2006). Markov Chains : Models, Algoritms and Application. Springer Science.

Feinberg, E. A., & Shwartz, A. (2002). Handbook of Markov Decision Processes: Handbook and Applications. In Kluwer’s International Series (Issue December).

Ferreira, G. O., Arruda, E. F., & Marujo, L. G. (2018). Inventory management of perishable items in long-term humanitarian operations using Markov decision processes. International Journal of Disaster Risk Reduction, 31, 460–469. https://doi.org/https://doi.org/10.1016/j.ijdrr.2018.05.010

Fürnkranz, J., Hüllermeier, E., Cheng, W., & Park, S. H. (2012). Preference-based reinforcement learning: a formal framework and a policy iteration algorithm. Machine Learning, 89(1), 123–156. https://doi.org/https://doi.org/10.1007/s10994-012-5313-8

Hasan, & Iqbal, M. (2004). Pokok-Pokok Teori Pengambilan Keputusan. Ghalia Indonesia.

Iversen, E. B., Morales, J. M., & Madsen, H. (2014). Optimal charging of an electric vehicle using a Markov decision process. Applied Energy, 123, 1–12. https://doi.org/https://doi.org/10.1016/j.apenergy.2014.02.003

Ksentini, A., Taleb, T., & Chen, M. (2014). A Markov decision process-based service migration procedure for follow me cloud. 2014 IEEE International Conference on Communications (ICC), 1350–1354. https://doi.org/https://doi.org/10.1109/ICC.2014.6883509

Layla, M. (2016). Analisis persediaan bahan baku yang optimal menggunakan rantai Markov di PT. PDM Indonesia. Jurnal Mantik Penusa, 20(1), 17–23. https://e-jurnal.pelitanusantara.ac.id/index.php/mantik/article/view/206

Littman, M. L., Dean, T. L., & Kaelbling, L. P. (2013). On the complexity of solving Markov decision problems. The Eleventh Conference on Uncertainty in Artificial Intelligence, 394–402. https://doi.org/https://doi.org/10.48550/arXiv.1302.4971

Liu, D., & Wei, Q. (2013). Policy iteration adaptive dynamic programming algorithm for discrete-time nonlinear systems. IEEE Transactions on Neural Networks and Learning Systems, 25(3), 621–634. https://doi.org/https://doi.org/10.1109/TNNLS.2013.2281663

Liu, Y. J., Cheng, S. M., & Hsueh, Y. L. (2017). eNB selection for machine type communications using reinforcement learning based Markov decision process. IEEE Transactions on Vehicular Technology, 66(12), 11330–11338. https://doi.org/https://doi.org/10.1109/TVT.2017.2730230

Luo, B., Wu, H. N., Huang, T., & Liu, D. (2014). Data-based approximate policy iteration for affine nonlinear continuous-time optimal control design. Automatica, 50(12), 3281–3290. https://doi.org/https://doi.org/10.1016/j.automatica.2014.10.056

Mani, A., Bakar, S. A., Krishnan, P., & Yaacob, S. (2021). Markov decision process approach in the estimation of raw material quality in incoming inspection process. Journal of Physics: Conference Series, 2107(1), 012025. https://doi.org/10.1088/1742-6596/2107/1/012025

Noorida, T. (2003). Analisa persediaan material pada fire tube boiler stock yang optimal dengan menggunakan Markov chain. Jurnal Teknik Industri, 4(2), 40–49. https://doi.org/https://doi.org/10.22219/jtiumm.vol4.no2.40-49

Oktaviyani, O., Dwijanto, D., & Supriyono, S. (2018). Optimization of production scheduling and planning of raw materials supply using Markov chains (a case study of Kinken Cake & Bakery Kutoarjo). Unnes Journal of Mathematics, 7(2), 165–180. https://doi.org/https://doi.org/10.15294/ujm.v7i2.13894

Ong, H. Y., & Kochenderfer, M. J. (2017). Markov decision process-based distributed conflict resolution for drone air traffic management. Journal of Guidance, Control, and Dynamics, 40(1), 69–80. https://doi.org/https://doi.org/10.2514/1.G001822

Pérolat, J., Piot, B., Geist, M., Scherrer, B., & Pietquin, O. (2016). Softened approximate policy iteration for Markov games. International Conference on Machine Learning, 1860–1868. http://proceedings.mlr.press/v48/perolat16.html

Ristono, A. (2009). Manajemen Persediaan. Graha Ilmu.

Rong, H., Zhou, X., Yang, C., Shafiq, Z., & Liu, A. (2016). The rich and the poor: a Markov decision process approach to optimizing taxi driver revenue efficiency. Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, 2329–2334. https://doi.org/https://doi.org/10.1145/2983323.2983689

Ross, S. M. (2010). Introduction to Probability Models (10th ed.). Academic Press.

Sarjono, H., Edwin, E., Sentosa, H., & Bong, F. (2011). Analisis Markov chain terhadap persediaan: studi kasus pada CV Sinar Bahagia Group. Binus Business Review, 2(2), 1071–1076. https://doi.org/https://doi.org/10.21512/bbr.v2i2.1249

Shou, Z., Di, X., Ye, J., Zhu, H., Zhang, H., & Hampshire, R. (2020). Optimal passenger-seeking policies on E-hailing platforms using Markov decision process and imitation learning. Transportation Research Part C: Emerging Technologies, 111, 91–113. https://doi.org/https://doi.org/10.1016/j.trc.2019.12.005

Taylor, H. M., & Samuel, K. (1998). An Introduction to Stochastic Modelling (3rd ed.). Academic Press.

Wu, Y., & Shen, T. (2017). Policy iteration algorithm for optimal control of stochastic logical dynamical systems. IEEE Transactions on Neural Networks and Learning Systems, 29(5), 2031–2036. https://doi.org/https://doi.org/10.1109/TNNLS.2017.2661863

Yang, X., & Wei, Q. (2018). An offâ€policy iteration algorithm for robust stabilization of constrainedâ€input uncertain nonlinear systems. International Journal of Robust and Nonlinear Control, 28(18), 5747–5765. https://doi.org/https://doi.org/10.1002/rnc.4342

Yousefi, S., Derakhshan, F., Karimipour, H., & Aghdasi, H. S. (2020). An efficient route planning model for mobile agents on the internet of things using Markov decision process. Ad Hoc Networks, 98(102053). https://doi.org/https://doi.org/10.1016/j.adhoc.2019.102053

Zheng, J., & Siami Namin, A. (2018). A Markov decision process to determine optimal policies in moving target. Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security, 2321–2323. https://doi.org/https://doi.org/10.1145/3243734.3278489

Published

2022-07-16

Issue

Section

Articles