An offline demand estimation method for multi-threaded applications
Parameterizing performance models for multi-threaded enterprise applications requires finding the service rates offered by worker threads to the incoming requests. Statistical inference on monitoring data is here helpful to reduce the overheads of application profiling and to infer missing informati...
- Autores:
- Tipo de recurso:
- Fecha de publicación:
- 2014
- Institución:
- Universidad del Rosario
- Repositorio:
- Repositorio EdocUR - U. Rosario
- Idioma:
- eng
- OAI Identifier:
- oai:repository.urosario.edu.co:10336/28497
- Acceso en línea:
- https://doi.org/10.1109/MASCOTS.2013.10
https://repository.urosario.edu.co/handle/10336/28497
- Palabra clave:
- Servers
Computational modeling
Time factors
Instruction sets
Maximum likelihood estimation
Estimation error
- Rights
- License
- Restringido (Acceso a grupos específicos)
Summary: | Parameterizing performance models for multi-threaded enterprise applications requires finding the service rates offered by worker threads to the incoming requests. Statistical inference on monitoring data is here helpful to reduce the overheads of application profiling and to infer missing information. While linear regression of utilization data is often used to estimate service rates, it suffers erratic performance and also ignores a large part of application monitoring data, e.g., response times. Yet inference from other metrics, such as response times or queue-length samples, is complicated by the dependence on scheduling policies. To address these issues, we propose novel scheduling-aware estimation approaches for multi-threaded applications based on linear regression and maximum likelihood estimators. The proposed methods estimate demands from samples of the number of requests in execution in the worker threads at the admission instant of a new request. Validation results are presented on simulated and real application datasets for systems with multi-class requests, class switching, and admission control. |
---|