Composition of Web Services Using Markov Decision Processes and Dynamic Programming

We propose a Markov decision process model for solving the Web service composition (WSC) problem. Iterative policy evaluation, value iteration, and policy iteration algorithms are used to experimentally validate our approach, with artificial and real data. The experimental results show the reliabili...

Full description

Bibliographic Details
Main Authors: Víctor Uc-Cetina, Francisco Moo-Mena, Rafael Hernandez-Ucan
Format: Article
Language:English
Published: Hindawi Limited 2015-01-01
Series:The Scientific World Journal
Online Access:http://dx.doi.org/10.1155/2015/545308
id doaj-04c809154d8444208a9cec258bb8a4b5
record_format Article
spelling doaj-04c809154d8444208a9cec258bb8a4b52020-11-25T00:08:39ZengHindawi LimitedThe Scientific World Journal2356-61401537-744X2015-01-01201510.1155/2015/545308545308Composition of Web Services Using Markov Decision Processes and Dynamic ProgrammingVíctor Uc-Cetina0Francisco Moo-Mena1Rafael Hernandez-Ucan2Facultad de Matemáticas, Universidad Autónoma de Yucatán, Anillo Periférico Norte, Tablaje Cat. 13615, Apartado Postal 192, Colonia Chuburná Hidalgo Inn, 97119 Mérida, YUC, MexicoFacultad de Matemáticas, Universidad Autónoma de Yucatán, Anillo Periférico Norte, Tablaje Cat. 13615, Apartado Postal 192, Colonia Chuburná Hidalgo Inn, 97119 Mérida, YUC, MexicoFacultad de Matemáticas, Universidad Autónoma de Yucatán, Anillo Periférico Norte, Tablaje Cat. 13615, Apartado Postal 192, Colonia Chuburná Hidalgo Inn, 97119 Mérida, YUC, MexicoWe propose a Markov decision process model for solving the Web service composition (WSC) problem. Iterative policy evaluation, value iteration, and policy iteration algorithms are used to experimentally validate our approach, with artificial and real data. The experimental results show the reliability of the model and the methods employed, with policy iteration being the best one in terms of the minimum number of iterations needed to estimate an optimal policy, with the highest Quality of Service attributes. Our experimental work shows how the solution of a WSC problem involving a set of 100,000 individual Web services and where a valid composition requiring the selection of 1,000 services from the available set can be computed in the worst case in less than 200 seconds, using an Intel Core i5 computer with 6 GB RAM. Moreover, a real WSC problem involving only 7 individual Web services requires less than 0.08 seconds, using the same computational power. Finally, a comparison with two popular reinforcement learning algorithms, sarsa and Q-learning, shows that these algorithms require one or two orders of magnitude and more time than policy iteration, iterative policy evaluation, and value iteration to handle WSC problems of the same complexity.http://dx.doi.org/10.1155/2015/545308
collection DOAJ
language English
format Article
sources DOAJ
author Víctor Uc-Cetina
Francisco Moo-Mena
Rafael Hernandez-Ucan
spellingShingle Víctor Uc-Cetina
Francisco Moo-Mena
Rafael Hernandez-Ucan
Composition of Web Services Using Markov Decision Processes and Dynamic Programming
The Scientific World Journal
author_facet Víctor Uc-Cetina
Francisco Moo-Mena
Rafael Hernandez-Ucan
author_sort Víctor Uc-Cetina
title Composition of Web Services Using Markov Decision Processes and Dynamic Programming
title_short Composition of Web Services Using Markov Decision Processes and Dynamic Programming
title_full Composition of Web Services Using Markov Decision Processes and Dynamic Programming
title_fullStr Composition of Web Services Using Markov Decision Processes and Dynamic Programming
title_full_unstemmed Composition of Web Services Using Markov Decision Processes and Dynamic Programming
title_sort composition of web services using markov decision processes and dynamic programming
publisher Hindawi Limited
series The Scientific World Journal
issn 2356-6140
1537-744X
publishDate 2015-01-01
description We propose a Markov decision process model for solving the Web service composition (WSC) problem. Iterative policy evaluation, value iteration, and policy iteration algorithms are used to experimentally validate our approach, with artificial and real data. The experimental results show the reliability of the model and the methods employed, with policy iteration being the best one in terms of the minimum number of iterations needed to estimate an optimal policy, with the highest Quality of Service attributes. Our experimental work shows how the solution of a WSC problem involving a set of 100,000 individual Web services and where a valid composition requiring the selection of 1,000 services from the available set can be computed in the worst case in less than 200 seconds, using an Intel Core i5 computer with 6 GB RAM. Moreover, a real WSC problem involving only 7 individual Web services requires less than 0.08 seconds, using the same computational power. Finally, a comparison with two popular reinforcement learning algorithms, sarsa and Q-learning, shows that these algorithms require one or two orders of magnitude and more time than policy iteration, iterative policy evaluation, and value iteration to handle WSC problems of the same complexity.
url http://dx.doi.org/10.1155/2015/545308
work_keys_str_mv AT victoruccetina compositionofwebservicesusingmarkovdecisionprocessesanddynamicprogramming
AT franciscomoomena compositionofwebservicesusingmarkovdecisionprocessesanddynamicprogramming
AT rafaelhernandezucan compositionofwebservicesusingmarkovdecisionprocessesanddynamicprogramming
_version_ 1725415132372140032