We give a model of parallel distributed genetic improvement. With modern low cost power monitors; high speed Ethernet LAN latency and network jitter have little effect. The model calculates a minimum usable mutation effect based on the analogue to digital converter (ADC)’s resolution and shows the optimal test duration is inversely proportional to smallest impact we wish to detect. Using the example of a 1 kHz 12 bit 0.4095 Amp ADC optimising software energy consumption we find: it will be difficult to detect mutations which an average effect less than 58 μA, and typically experiments should last well under a second.
CITATION STYLE
Langdon, W. B., Petke, J., & Bruce, B. R. (2016). Optimising quantisation noise in energy measurement. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9921 LNCS, pp. 249–259). Springer Verlag. https://doi.org/10.1007/978-3-319-45823-6_23
Mendeley helps you to discover research relevant for your work.