Energy Estimation Metric Documentation¶
Overview¶
The Energy Estimation Metric is an analytical tool for estimating the energy consumption of Spiking Neural Networks (SNNs) and traditional neural networks. It is based on the research paper “An Analytical Estimation of Spiking Neural Networks Energy Efficiency” (Lemaire et al., ICONIP 2022). This metric provides detailed insights into various energy consumption aspects of your neural network, including memory operations, computational operations, and spike-based calculations.
How It Works¶
The Energy Estimation Metric analyzes several key components of energy consumption:
Memory Operations Energy:
Memory potential (mem_pot): Energy for reading/writing neuron potentials
Memory weights (mem_weights): Energy for reading weights
Memory bias (mem_bias): Energy for reading biases
Memory I/O (mem_io): Energy for input/output operations
Computational Operations Energy:
Operations (ops): Energy for synaptic operations
Address computation (addr): Energy for event address calculation
Spike-Based Metrics:
Input spike rate: Average spikes per input per timestep
Output spike rate: Average spikes per output per timestep
The metric uses configurable energy values based on a 45nm ASIC implementation, with default values from Mark Horowitz’s ISSCC 2014 paper.
Configuration¶
To add the Energy Estimation Metric to your project, include it in your configuration file (config.toml) under the postprocessing section:
[[postprocessing]]
kind = "EnergyEstimationMetric"
params.mem_width = 8
params.fifo_size = 64
params.total_spikerate_exclude_nonbinary = true
params.op_estimation_type.add = "linear"
params.op_estimation_type.mul = "quadratic"
params.sram_estimation_type = "new"
Then you will need to run the following command:
qualia ./config.toml train
If you have already train your model you may want to not train it again. You need to add the following parameter to your model to directly run the Energy Estimation Metric:
[[model]]
name = "gsc_cnn_m5_smaller"
load = true # Tel qualia to load an existing model
train = false # Tel qualia to not train this model
Configuration Parameters¶
Required Parameters:
mem_width
: Bit width for memory operations (e.g., 8 for 8-bit quantization)fifo_size
: Size of the input/output FIFOs for each layertotal_spikerate_exclude_nonbinary
: Whether to exclude non-binary inputs/outputs from total spike rate computation
Optional Parameters:
op_estimation_type
: Method for estimating operation energy valuesSupported types: “ICONIP”, “saturation”, “linear”, “quadratic”
Can be specified separately for addition and multiplication
sram_estimation_type
: Algorithm for SRAM energy estimation“old”: ICONIP 2022 method (single access per data)
“new”: T. Louis method (packed data over 64-bit access)
Energy Estimation Methods¶
The metric supports four different methods for estimating operation energy:
ICONIP: Uses fixed 32-bit values from the ICONIP 2022 paper
Saturation: Uses closest predefined bitwidth values
Linear: Linear interpolation between 8-bit and 32-bit values
Quadratic: Quadratic interpolation between reference points
Usage¶
The Energy Estimation Metric automatically runs during model evaluation. It will generate two types of output:
Console Output: A detailed table showing energy consumption per layer
CSV Log: Detailed metrics saved in
logs/<bench.name>/EnergyEstimationMetric/
Reading the Results¶
The output includes the following metrics for each layer:
Layer | EMemPot | EMemWeights | EMemBias | EMemIO | EMemTotal | EOps | EAddr | EOpsAddr | ETotal | SNN | Input SR | Output SR
------------|---------|-------------|----------|--------|-----------|------|-------|----------|--------|-----|----------|----------
conv1 | 0.0023 | 0.0156 | 0.0012 | 0.0089 | 0.0280 | 0.15 | 0.02 | 0.17 | 0.198 | Yes | 0.142 | 0.086
All energy values are in nanojoules (nJ)
SR = Spike Rate (average spikes per neuron per timestep)
SNN indicates if the layer is processed as a spiking layer
Supported Layer Types¶
The metric supports analysis of:
Convolutional Layers:
Standard convolutions
Spiking convolutions
Quantized variants
Dense Layers:
Fully connected layers
Spiking dense layers
Quantized variants
Additional Layers:
Batch normalization
Pooling layers
Addition layers (for residual connections)
Complete Working Example¶
Let’s walk through a complete, practical example using the Google Speech Commands (GSC) dataset. This example demonstrates how to integrate energy estimation into a speech recognition system using a Spiking Neural Network:
[bench]
name = "GSC_SNN_example"
seed = 42
first_run = 1
last_run = 1
plugins = ["qualia_plugin_snn"] # Enable SNN support
[learningframework]
kind = "SpikingJellyMultiStep" # Multi-step processing for efficiency
[dataset]
kind = "GSC"
params.path = "data/speech_commands" # Path to extracted dataset
params.variant = "v2"
params.subset = "digits"
params.train_valid_split = true
[[preprocessing]]
kind = "Class2BinMatrix"
[[postprocessing]]
kind = "FuseBatchNorm" # All Batchnorm must be Fused before the estimation
export = true
params.evaluate = false
[[postprocessing]]
kind = "OperationCounter" # Information about Operations and Memory access
[[postprocessing]]
kind = "EnergyEstimationMetric"
params.mem_width = 8
params.fifo_size = 64
params.total_spikerate_exclude_nonbinary = true
params.op_estimation_type.add = "linear"
params.op_estimation_type.mul = "quadratic"
params.sram_estimation_type = "new"
# Model configuration
[model_template]
kind = "SCNN" # Spiking CNN
params.dims = 1
epochs = 10
batch_size = 512
params.timesteps = 2 # Number of timesteps for temporal processing
[model_template.optimizer]
kind = "Adam"
params.lr = 0.002
[[model]]
name = "gsc_scnn_m5_smaller_IF"
load = false
train = true
params.filters = [16, 16, 32, 64]
params.kernel_sizes = [40, 3, 3, 3]
params.paddings = [20, 1, 1, 1]
params.strides = [8, 4, 4, 4]
params.pool_sizes = [0, 0, 0, 0]
params.dropouts = [0, 0, 0, 0]
params.fc_units = []
params.gsp = true # End with a global sum pooling
params.batch_norm = true
params.neuron.kind = "IFNode" # Leaky Integrate-and-Fire neuron
params.neuron.params.v_threshold = 1.0
params.neuron.params.v_reset = false # Soft reset, foat value for Hard reset
params.neuron.params.detach_reset = true
params.neuron.params.step_mode = "m" # Multi-step mode
params.neuron.params.backend = "torch" # Use GPU acceleration if available
disabled = false
[[model]]
name = "gsc_scnn_m5_smaller_LIF"
load = false
train = true
params.filters = [16, 16, 32, 64]
params.kernel_sizes = [40, 3, 3, 3]
params.paddings = [20, 1, 1, 1]
params.strides = [8, 4, 4, 4]
params.pool_sizes = [0, 0, 0, 0]
params.dropouts = [0, 0, 0, 0]
params.fc_units = []
params.gsp = true
params.batch_norm = true
params.neuron.kind = "LIFNode" # Leaky Integrate-and-Fire neuron
params.neuron.params.tau = 2.0 # Membrane time constant
params.neuron.params.v_threshold = 1.0
params.neuron.params.v_reset = false # Soft reset, foat value for Hard reset
params.neuron.params.detach_reset = true
params.neuron.params.step_mode = "m" # Multi-step mode
params.neuron.params.backend = "torch" # Use GPU acceleration if available
disabled = false
#############################
# NEEDED RESOURCES #
#############################
# 1.9 Gb GPU RAM
# < 1 min train / model
# trageted test accuracy :
# gsc_scnn_m5_smaller_IF : 0.620
# gsc_scnn_m5_smaller_LIF : 0.657
#############################
Understanding the Example¶
Let’s break down the key components of this configuration:
Energy Estimation Setup: The energy estimation configuration is tailored for audio processing:
The 8-bit memory width matches common hardware implementations for audio processing
The 64-entry FIFO buffer provides balanced temporal data handling
Different operation estimation types for addition (ICONIP) and multiplication (saturation) reflect the varying complexity of operations
Modern SRAM estimation using the T. Louis method optimizes for actual hardware behavior
Network Architecture: The Spiking CNN is designed for efficient audio processing:
Four convolutional layers with increasing filter counts [16, 16, 32, 64]
Initial large kernel (size 40) for processing audio input
Regular pooling operations (size 4) for dimensionality reduction
Leaky Integrate-and-Fire neurons with carefully tuned parameters
Temporal Processing: The configuration accounts for the temporal nature of speech:
Multi-step processing enabled through SpikingJellyMultiStep
Four timesteps for temporal processing
Membrane time constant of 2.0 for the LIF neurons
When you run this configuration, the Energy Estimation Metric will analyze:
The energy impact of temporal audio processing
Energy consumption patterns across network layers
The effectiveness of the pooling strategy
Overall efficiency of the spiking neural implementation
This example serves as a practical template for implementing energy estimation in speech recognition applications, but the principles can be adapted for other types of neural networks and applications.
Best Practices¶
Memory Width Selection:
Match
mem_width
to your model’s quantization bitsUse 32 for full-precision models
Use 8 or 16 for quantized models
FIFO Size:
Choose based on your hardware implementation
Larger sizes increase memory energy but may improve throughput
Common values range from 32 to 128
Energy Estimation Method:
Use “ICONIP” for comparison with the paper results
Use “saturation” for more realistic estimates with quantization
Use “linear” or “quadratic” for more precise interpolation
SRAM Estimation:
Use “new” for modern hardware implementations
Use “old” for comparison with ICONIP 2022 results
Troubleshooting¶
Common issues and solutions:
High Energy Consumption:
Check if your model uses appropriate quantization
Verify spike rates are reasonable
Consider reducing network size or complexity
Unexpected Spike Rates:
Verify neuron thresholds are properly set
Check input normalization
Ensure correct number of timesteps
Missing Metrics:
Verify all required parameters are set
Check if layer types are supported
Ensure proper plugin installation
References¶
Lemaire et al., “An Analytical Estimation of Spiking Neural Networks Energy Efficiency,” ICONIP 2022
Horowitz, “Computing’s Energy Problem (and what we can do about it),” ISSCC 2014