Optimal Raw Material Inventory Analysis Using Markov Decision Process with Policy Iteration Method
Abstract
Inventory of raw materials is a big deal in every production process, both in company production and home business production. In order to meet consumer demand, a business must be able to determine the amount of inventory that should be provided. The purpose of this research is to choose an alternative selection of ordering raw materials that produce the maximum amount of raw materials with minimum costs. The raw material referred to in this study is pandan leaves used to make pandan mats. Analysis of raw material inventory used in this research was the Markov decision process with the policy iteration method by considering the discount factor. From the analysis conducted, it is obtained alternative policies that must be taken by producers to meet raw materials with minimum costs. The results of this study can be a consideration for business actors in the study location in deciding the optimal ordering policy that should be taken to obtain the minimum operational cost.
Keywords
Full Text:
DOWNLOAD [PDF]References
Alla, A., Falcone, M., & Kalise, D. (2015). An efficient policy iteration algorithm for dynamic programming equations. SIAM Journal on Scientific Computing, 37(1), A181–A200. https://doi.org/https://doi.org/10.1137/130932284
Bennett, C. C., & Hauser, K. (2013). Artificial intelligence framework for simulating clinical decision-making: A Markov decision process approach. Artificial Intelligence in Medicine, 57(1), 9–19. https://doi.org/10.1016/j.artmed.2012.12.003
Ching, W. K., & Ng, M. K. (2006). Markov Chains : Models, Algoritms and Application. Springer Science.
Feinberg, E. A., & Shwartz, A. (2002). Handbook of Markov Decision Processes: Handbook and Applications. In Kluwer’s International Series (Issue December).
Ferreira, G. O., Arruda, E. F., & Marujo, L. G. (2018). Inventory management of perishable items in long-term humanitarian operations using Markov decision processes. International Journal of Disaster Risk Reduction, 31, 460–469. https://doi.org/https://doi.org/10.1016/j.ijdrr.2018.05.010
Fürnkranz, J., Hüllermeier, E., Cheng, W., & Park, S. H. (2012). Preference-based reinforcement learning: a formal framework and a policy iteration algorithm. Machine Learning, 89(1), 123–156. https://doi.org/https://doi.org/10.1007/s10994-012-5313-8
Hasan, & Iqbal, M. (2004). Pokok-Pokok Teori Pengambilan Keputusan. Ghalia Indonesia.
Iversen, E. B., Morales, J. M., & Madsen, H. (2014). Optimal charging of an electric vehicle using a Markov decision process. Applied Energy, 123, 1–12. https://doi.org/https://doi.org/10.1016/j.apenergy.2014.02.003
Ksentini, A., Taleb, T., & Chen, M. (2014). A Markov decision process-based service migration procedure for follow me cloud. 2014 IEEE International Conference on Communications (ICC), 1350–1354. https://doi.org/https://doi.org/10.1109/ICC.2014.6883509
Layla, M. (2016). Analisis persediaan bahan baku yang optimal menggunakan rantai Markov di PT. PDM Indonesia. Jurnal Mantik Penusa, 20(1), 17–23. https://e-jurnal.pelitanusantara.ac.id/index.php/mantik/article/view/206
Littman, M. L., Dean, T. L., & Kaelbling, L. P. (2013). On the complexity of solving Markov decision problems. The Eleventh Conference on Uncertainty in Artificial Intelligence, 394–402. https://doi.org/https://doi.org/10.48550/arXiv.1302.4971
Liu, D., & Wei, Q. (2013). Policy iteration adaptive dynamic programming algorithm for discrete-time nonlinear systems. IEEE Transactions on Neural Networks and Learning Systems, 25(3), 621–634. https://doi.org/https://doi.org/10.1109/TNNLS.2013.2281663
Liu, Y. J., Cheng, S. M., & Hsueh, Y. L. (2017). eNB selection for machine type communications using reinforcement learning based Markov decision process. IEEE Transactions on Vehicular Technology, 66(12), 11330–11338. https://doi.org/https://doi.org/10.1109/TVT.2017.2730230
Luo, B., Wu, H. N., Huang, T., & Liu, D. (2014). Data-based approximate policy iteration for affine nonlinear continuous-time optimal control design. Automatica, 50(12), 3281–3290. https://doi.org/https://doi.org/10.1016/j.automatica.2014.10.056
Mani, A., Bakar, S. A., Krishnan, P., & Yaacob, S. (2021). Markov decision process approach in the estimation of raw material quality in incoming inspection process. Journal of Physics: Conference Series, 2107(1), 012025. https://doi.org/10.1088/1742-6596/2107/1/012025
Noorida, T. (2003). Analisa persediaan material pada fire tube boiler stock yang optimal dengan menggunakan Markov chain. Jurnal Teknik Industri, 4(2), 40–49. https://doi.org/https://doi.org/10.22219/jtiumm.vol4.no2.40-49
Oktaviyani, O., Dwijanto, D., & Supriyono, S. (2018). Optimization of production scheduling and planning of raw materials supply using Markov chains (a case study of Kinken Cake & Bakery Kutoarjo). Unnes Journal of Mathematics, 7(2), 165–180. https://doi.org/https://doi.org/10.15294/ujm.v7i2.13894
Ong, H. Y., & Kochenderfer, M. J. (2017). Markov decision process-based distributed conflict resolution for drone air traffic management. Journal of Guidance, Control, and Dynamics, 40(1), 69–80. https://doi.org/https://doi.org/10.2514/1.G001822
Pérolat, J., Piot, B., Geist, M., Scherrer, B., & Pietquin, O. (2016). Softened approximate policy iteration for Markov games. International Conference on Machine Learning, 1860–1868. http://proceedings.mlr.press/v48/perolat16.html
Ristono, A. (2009). Manajemen Persediaan. Graha Ilmu.
Rong, H., Zhou, X., Yang, C., Shafiq, Z., & Liu, A. (2016). The rich and the poor: a Markov decision process approach to optimizing taxi driver revenue efficiency. Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, 2329–2334. https://doi.org/https://doi.org/10.1145/2983323.2983689
Ross, S. M. (2010). Introduction to Probability Models (10th ed.). Academic Press.
Sarjono, H., Edwin, E., Sentosa, H., & Bong, F. (2011). Analisis Markov chain terhadap persediaan: studi kasus pada CV Sinar Bahagia Group. Binus Business Review, 2(2), 1071–1076. https://doi.org/https://doi.org/10.21512/bbr.v2i2.1249
Shou, Z., Di, X., Ye, J., Zhu, H., Zhang, H., & Hampshire, R. (2020). Optimal passenger-seeking policies on E-hailing platforms using Markov decision process and imitation learning. Transportation Research Part C: Emerging Technologies, 111, 91–113. https://doi.org/https://doi.org/10.1016/j.trc.2019.12.005
Taylor, H. M., & Samuel, K. (1998). An Introduction to Stochastic Modelling (3rd ed.). Academic Press.
Wu, Y., & Shen, T. (2017). Policy iteration algorithm for optimal control of stochastic logical dynamical systems. IEEE Transactions on Neural Networks and Learning Systems, 29(5), 2031–2036. https://doi.org/https://doi.org/10.1109/TNNLS.2017.2661863
Yang, X., & Wei, Q. (2018). An off‐policy iteration algorithm for robust stabilization of constrained‐input uncertain nonlinear systems. International Journal of Robust and Nonlinear Control, 28(18), 5747–5765. https://doi.org/https://doi.org/10.1002/rnc.4342
Yousefi, S., Derakhshan, F., Karimipour, H., & Aghdasi, H. S. (2020). An efficient route planning model for mobile agents on the internet of things using Markov decision process. Ad Hoc Networks, 98(102053). https://doi.org/https://doi.org/10.1016/j.adhoc.2019.102053
Zheng, J., & Siami Namin, A. (2018). A Markov decision process to determine optimal policies in moving target. Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security, 2321–2323. https://doi.org/https://doi.org/10.1145/3243734.3278489
DOI: https://doi.org/10.31764/jtam.v6i3.8563
Refbacks
- There are currently no refbacks.
Copyright (c) 2022 Ikhsan Maulidi, Radhiah, Chintya Hayati, Vina Apriliani
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
_______________________________________________
JTAM already indexing:
_______________________________________________
JTAM (Jurnal Teori dan Aplikasi Matematika) |
_______________________________________________
_______________________________________________
JTAM (Jurnal Teori dan Aplikasi Matematika) Editorial Office: