Holistic workload scaling : A new approach to compute acceleration in the cloud

Workload scaling is an approach to accelerating computation and thus improving response times by replicating the exact same request multiple times and processing it in parallel on multiple nodes and accepting the result from the first node to finish. This is not unlike a TV game show, where the same...

Full description

Autores:
Tipo de recurso:
Fecha de publicación:
2018
Institución:
Universidad del Rosario
Repositorio:
Repositorio EdocUR - U. Rosario
Idioma:
eng
OAI Identifier:
oai:repository.urosario.edu.co:10336/19089
Acceso en línea:
http://repository.urosario.edu.co/handle/10336/19089
Palabra clave:
Cloud Computing
Stochastic Systems
Cloud Environments
Inter Processor Communication
Mapreudce
Model And Analysis
Optimal Approaches
Parallelilzation
Performance Modeling And Analysis
Stochastic Variation
Economic And Social Effects
Probabilidades & matemáticas aplicadas
Sistemas estocásticos
Comercio electrónico
Rights
License
Abierto (Texto Completo)
Description
Summary:Workload scaling is an approach to accelerating computation and thus improving response times by replicating the exact same request multiple times and processing it in parallel on multiple nodes and accepting the result from the first node to finish. This is not unlike a TV game show, where the same question is given to multiple contestants and the (correct) answer is accepted from the first to respond. This is different than traditional strategies for parallelization as used in, say, MapReduce workloads, where each node runs a subset of the overall workload. There are a variety of strategies that trade off metrics such as cost, utilization, performance, and interprocessor communication requirements. Performance modeling can help determine optimal approaches for different environments and goals. This is important, because poor performance can lead to application and domain-specific losses, such as e-commerce conversions and sales. Performance modeling and analysis plays an important role in designing and driving the selection of resource scaling mechanisms. Such modeling and analysis is complex due to time-varying workload arrival rates and request sizes, and even more complex in cloud environments due to the additional stochastic variation caused by performance interference due to resource sharing across co-located tenants. Moreover, little is known on how to multi-scale, i.e., dynamically and simultaneously scale resources vertically, horizontally, and through workload scaling. In this article, we first demonstrate the effectiveness of multi-scaling in reducing latency, and then discuss the performance modeling challenges, particularly for workload scaling. © 2014 IEEE.