Results 1 - 10
of
17
Dynamic Resource Allocation in Enterprise Systems
"... It is common that Internet service hosting centres use several logical pools to assign server resources to different applications, and that they try to achieve the highest total revenue by making efficient use of these resources. In this paper, multi-tiered enterprise systems are modelled as multi-c ..."
Abstract
-
Cited by 4 (3 self)
- Add to MetaCart
(Show Context)
It is common that Internet service hosting centres use several logical pools to assign server resources to different applications, and that they try to achieve the highest total revenue by making efficient use of these resources. In this paper, multi-tiered enterprise systems are modelled as multi-class closed queueing networks, with each network station corresponding to each application tier. In such queueing networks, bottlenecks can limit overall system performance, and thus should be avoided. We propose a bottleneck-aware server switching policy, which responds to system bottlenecks and switches servers to alleviate these problems as necessary. The switching engine compares the benefits and penalties of a potential switch, and makes a decision as to whether it is likely to be worthwhile switching. We also propose a simple admission control scheme, in addition to the switching policy, to deal with system overloading and optimise the total revenue of multiple applications in the hosting centre. Performance evaluation has been done via simulation and results are compared with those from a proportional switching policy and also a system that implements no switching policy. The experimental results show that the combination of the bottleneck-aware switching policy and the admission control scheme consistently outperforms the other two policies in terms of revenue contribution. 1
Modeling End-to-End Response Times in Multi-Tier Internet Applications
"... Abstract. Many Internet applications and workflow systems employ multi-tier software architectures. The performance of such multi-tier Internet applications is typically measured by the end-to-end response times. Most of the earlier works in modeling the response times of such systems have limited t ..."
Abstract
-
Cited by 4 (1 self)
- Add to MetaCart
(Show Context)
Abstract. Many Internet applications and workflow systems employ multi-tier software architectures. The performance of such multi-tier Internet applications is typically measured by the end-to-end response times. Most of the earlier works in modeling the response times of such systems have limited their study to modeling the mean. However, since the user-perceived performance is highly influenced by the variability in response times, the variance of the response times is important as well. In this paper, we derive expressions for estimating the mean end-to-end latency of more general applications and provide approximations to its variance. We validate our model by testing on two Web applications. We find that the model has a very good accuracy in estimating the mean endto-end response time and its variance with margins of error less than 10 percent. We believe this model is useful for capacity planning, resource provisioning, and performance prediction of Internet applications and services.
Distributed workload and response time management for web applications
- Proceedings of the 7th International Conference on Network and Services Management
, 2011
"... Abstract-Managing workload for large scale web applications is a fundamental task for satisfactory quality of service, low management and operation cost. In this paper, we present SCOPS, a system of distributed workload management to achieve service differentiation and overload protection in such l ..."
Abstract
-
Cited by 2 (0 self)
- Add to MetaCart
(Show Context)
Abstract-Managing workload for large scale web applications is a fundamental task for satisfactory quality of service, low management and operation cost. In this paper, we present SCOPS, a system of distributed workload management to achieve service differentiation and overload protection in such large scale deployment. Our system splits the workload management logic into distributed components on each back-end server and frontend proxy. The control solution is designed to protect the backend server from overloading and to achieve both efficient usage of system resource and service differentiation by employing a unique optimization target. The control components are automatically organized based on the flow of workloads, such that management overhead is minimized. SCOPS is extremely flexible because it requires no source code changes to host OS, application servers, or web applications. Additionally, the distributed design makes it scalable and robust for cloud scale server deployment. Experiments with our implementation confirm SCOPS's performance with dynamic heavy workload, incurring neglectable runtime overhead. More importantly, SCOPS also ensures fault-tolerance and fast convergence to system failures.
From Autonomic to Self-Self Behaviors: The JADE Experience
"... All in-text references underlined in blue are linked to publications on ResearchGate, letting you access and read them immediately. ..."
Abstract
- Add to MetaCart
All in-text references underlined in blue are linked to publications on ResearchGate, letting you access and read them immediately.
Optimal Resource Allocation in Synchronized Multi-Tier Internet Services
"... Abstract Modern Internet systems have evolved from simple monolithic systems to complex multitiered architectures. For these systems, providing good response time performance is crucial for the commercial success. In practice, the response-time performance of multi-tiered systems is often degraded ..."
Abstract
- Add to MetaCart
Abstract Modern Internet systems have evolved from simple monolithic systems to complex multitiered architectures. For these systems, providing good response time performance is crucial for the commercial success. In practice, the response-time performance of multi-tiered systems is often degraded by severe synchronization problems, causing jobs to wait for responses from different tiers. Synchronization between different tiers is a complicating factor in the optimal control and analysis of the performance. In this paper, we study a generic multi-tier model with synchronization. The system is able to share processing capacity between arriving jobs that need to be sent to other tiers and the responses that have arrived after processing from these tiers. We provide structural results on the optimal resource allocation policy and provide a full characterization of the policy in the framework of Markov decision theory. We also highlight important effects of synchronization in the model and discuss their implications for practice. We validate our expressions through extensive experimentations for a wide range of resource configurations.
Modeling and Predicting End-to-End Response Times in Multi-Tier Internet Applications
"... Many Internet applications employ multi-tier software architectures. The performance of such multi-tier Internet applications is typically measured by the end-toend response times. Most of the earlier works in modeling the response times of such systems have limited their study to modeling the mean. ..."
Abstract
- Add to MetaCart
(Show Context)
Many Internet applications employ multi-tier software architectures. The performance of such multi-tier Internet applications is typically measured by the end-toend response times. Most of the earlier works in modeling the response times of such systems have limited their study to modeling the mean. However, since the user-perceived performance is highly influenced by the variability in response times, the variance of the response times is important as well. We first develop a simple model for the end-to-end response times for multitiered Internet applications. We validate the model by real data from two large-scale applications that are widely deployed on the Internet. Second, we derive exact and approximate expressions for the mean and the variance, respectively, of the end-toend response times. Extensive numerical validation shows that the approximations match very well with simulations. These observations make the results presented highly useful for capacity planning and performance prediction of large-scale multitiered Internet applications.
Request-Aware Scheduling for Busy Internet Services
"... Abstract — Internet traffic is bursty and network servers are often overloaded with surprising events or abnormal client request patterns. This paper studies scheduling algorithms for interactive network services that use multiple threads to handle incoming requests continuously and concurrently. Ou ..."
Abstract
- Add to MetaCart
(Show Context)
Abstract — Internet traffic is bursty and network servers are often overloaded with surprising events or abnormal client request patterns. This paper studies scheduling algorithms for interactive network services that use multiple threads to handle incoming requests continuously and concurrently. Our investigation with applications from Ask Jeeves search shows that during overloaded situations, requests that require excessive computing resource can dramatically affect the overall system throughput and response time. The most effective method is to manage resource usage at a request level instead of a thread or process level. We propose a new size-adaptive request-aware scheduling algorithm called SRQ with dynamic feedbacks to control queue properties and have implemented SRQ in the Linux kernel level. Our experimental results with several application service benchmarks indicate that the proposed scheduler can significantly outperform the standard Linux scheduler. I.