Article ID: | iaor201522032 |
Volume: | 45 |
Issue: | 3 |
Start Page Number: | 535 |
End Page Number: | 567 |
Publication Date: | Jun 2014 |
Journal: | Decision Sciences |
Authors: | Gopalakrishnan Mohan, Mohan Srimathy, Printezis Antonios, Alam Ferdous M, Fowler John W |
Keywords: | decision, planning, financial, quality & reliability, e-commerce, queues: applications, markov processes |
Motivated by the technology division of a financial services firm, we study the problem of capacity planning and allocation for Web‐based applications. The steady growth in Web traffic has affected the quality of service (QoS) as measured by response time (RT), for numerous e‐businesses. In addition, the lack of understanding of system interactions and availability of proper planning tools has impeded effective capacity management. Managers typically make decisions to add server capacity on an ad hoc basis when systems reach critical response levels. Very often this turns out to be too late and results in extremely long response times and the system crashes. We present an analytical model to understand system interactions with the goal of making better server capacity decisions based on the results. The model studies the relationships and important interactions between the various components of a Web‐based application using a continuous time Markov chain embedded in a queuing network as the basic framework. We use several structured aggregation schemes to appropriately represent a complex system, and demonstrate how the model can be used to quickly predict system performance, which facilitates effective capacity allocation decision making. Using simulation as a benchmark, we show that our model produces results within 5% accuracy at a fraction of the time of simulation, even at high traffic intensities. This knowledge helps managers quickly analyze the performance of the system and better plan server capacity to maintain desirable levels of QoS. We also demonstrate how to utilize a combination of dedicated and shared resources to achieve QoS using fewer servers.