کد مقاله کد نشریه سال انتشار مقاله انگلیسی نسخه تمام متن
480520 1445973 2016 11 صفحه PDF دانلود رایگان
عنوان انگلیسی مقاله ISI
A two-state partially observable Markov decision process with three actions
ترجمه فارسی عنوان
فرایند تصمیم گیری مارکوف دوحالته تا حدی قابل مشاهده با سه اقدام
کلمات کلیدی
پردازش های تصمیم گیری؛ زنجیره مارکف؛ POMDP؛ محدودیت های کنترل
موضوعات مرتبط
مهندسی و علوم پایه مهندسی کامپیوتر علوم کامپیوتر (عمومی)
چکیده انگلیسی


• Considering a decision process that can shift states without interference.
• Considering an inspection action in addition to continuing or repairing the process.
• Analyzing the optimal strategy under perfect and imperfect inspections.
• The optimal strategy is of a threshold type with regard to the state probability.
• Measuring the value of the inspection action.

A process can be in either a stable or an unstable state interchangeably. The true state is unobservable and can only be inferred from observations. Three actions are available: continue with the process (CON), repair the process for a certain fee – bring the process to the stable state (REP), and obtain the state of the process for a cost (INS). The objective is to maximize the expected discounted value of the total future profits. We formulate the problem as a discrete-time Partially Observable Markov Decision Process (POMDP). We show that the expected profit function is convex and strictly increasing, and that the optimal policy has either one or two control limits. Also, we show that “dominance in expectation” (the expected revenue is larger in the stable state than in the unstable state) suffices for a control limit structure.

ناشر
Database: Elsevier - ScienceDirect (ساینس دایرکت)
Journal: European Journal of Operational Research - Volume 254, Issue 3, 1 November 2016, Pages 957–967
نویسندگان
, , ,