Hi all,
I’m currently working on a Bayesian analysis project to model the energy consumption (in joules) of two different software fault detection tools. One tool is “traditional”, relying solely on CPU computation, while the other leverages a large language model (LLM) and makes heavy use of GPU resources.
I’ve already set up my benchmarking environment and can gather detailed execution data for both tools, including energy usage. However, I’m unsure how to approach the modeling, particularly with regard to choosing appropriate distributions and accounting for the different hardware components.
My main questions are:
1- What kind of likelihood distributions are typically used to model energy consumption data (e.g., for CPU and GPU usage separately or jointly)?
2- How can I structure the models to reflect the fact that one tool uses only CPU, while the other uses both CPU and GPU, yet still make their energy consumption comparable in a principled way?
If more context would be helpful, I’d be happy to elaborate.
Best regards