Back to Insights
White Paper

Neuromorphic Sensor Architectures for Sub-Millisecond Fire Signature Recognition: A Comparative Study of Spiking Neural Networks on Event-Driven Hardware

Kaspar TavityanVerex Fire Solutions8 March 202616 min read

Abstract

Conventional fire detection systems, whether based on photoelectric scatter or AI-enhanced visual inference on edge GPUs, impose latency floors ranging from hundreds of milliseconds to tens of seconds. This paper presents a comparative evaluation of spiking neural networks deployed on Intel Loihi 2 neuromorphic hardware against a convolutional neural network baseline running on an NVIDIA Jetson Orin Nano, across three fire detection modalities: photoelectric scatter, thermal infrared imaging, and multi-spectral gas sensing. Our SNN architecture achieves a median detection latency of 340 microseconds from first anomalous sensor event, compared to 420 milliseconds for the GPU baseline, while consuming 23 milliwatts versus 7.4 watts. The false alarm rate on EN 54-7 standard test fires drops from 0.12% to 0.08%. These results suggest that neuromorphic computing offers a viable path toward battery-powered, distributed fire sensor nodes with multi-year operational lifespans and detection speeds three orders of magnitude faster than current commercial systems.

1. Introduction

The photoelectric smoke detector, for all its ubiquity, operates on a principle that imposes an unavoidable physical constraint. Smoke particles must drift into the sensing chamber in sufficient concentration to scatter light onto the photodiode. This takes time. Depending on airflow conditions, room geometry, and fire type, the interval between ignition and alarm can stretch from 30 seconds to several minutes. Ionisation detectors respond somewhat faster to flaming fires but remain fundamentally reactive: they wait for combustion byproducts to reach them, rather than seeking the earliest possible signatures of thermal or chemical anomaly.

AI-enhanced fire detection has made notable progress in recent years. Convolutional neural networks running on edge GPUs can classify thermal camera frames or multi-sensor telemetry with impressive accuracy. Yet even these systems face a latency floor imposed by their computational architecture. A frame must be captured, transferred to GPU memory, preprocessed into the model's expected input format, and then processed through dozens of convolutional layers. On an NVIDIA Jetson Orin Nano, a typical inference cycle for a fire classification CNN takes between 200 and 800 milliseconds depending on model depth, input resolution, and batch configuration. For many applications this is perfectly adequate. For others, particularly in environments where fire propagation is extremely rapid or where human occupancy is dense, it is not.

Neuromorphic computing represents a fundamentally different approach. Rather than processing data in synchronous clock cycles through dense matrix multiplications, neuromorphic processors operate asynchronously. Computation happens only when input events arrive, and only in the specific neurons those events affect. Intel's Loihi 2 chip, fabricated on Intel 4 process technology with 128 neuromorphic cores containing roughly one million artificial neurons, exemplifies this paradigm. IBM's NorthPole architecture takes a complementary approach, optimising for high throughput neural inference with near-memory computing, though it lacks the event-driven spiking behaviour that makes Loihi 2 particularly interesting for sensor processing.

The question we set out to answer is straightforward, if ambitious: can spiking neural networks running on neuromorphic hardware achieve meaningful improvements in fire detection latency while reducing power consumption to levels suitable for battery-powered distributed sensor nodes? We were particularly interested in whether the temporal coding capabilities of spiking networks could exploit information in the fine-grained timing of sensor events that conventional frame-based approaches simply discard.

2. Background on Spiking Neural Networks

Spiking neural networks constitute what Wolfgang Maass characterised in 1997 as the third generation of neural network models. Where first-generation networks used binary threshold units and second-generation networks use continuous-valued activations, spiking networks communicate through discrete, precisely timed pulses. This distinction is more than a theoretical curiosity. It changes what information the network can represent and how efficiently it can process temporal data.

In a conventional CNN, a neuron's output is a continuous floating-point value computed at every forward pass. In a spiking network, a neuron accumulates incoming signals over time according to the Leaky Integrate-and-Fire model. The neuron's membrane potential rises as input spikes arrive, decays gradually between inputs, and triggers an output spike only when it crosses a defined threshold. After firing, the membrane potential resets and the cycle begins again. The critical insight is that information can be encoded not merely in how many spikes a neuron fires over some interval, but in the precise timing of each spike relative to other spikes in the network. This temporal coding scheme allows spiking networks to represent fine-grained timing relationships that rate-coded networks cannot easily capture.

The computational efficiency advantage follows directly from this architecture. A spiking neuron that receives no input consumes no energy. In a fire detection context, this means the network operates at near-zero power during the long quiet periods when nothing unusual is happening, and ramps up power consumption only when the sensors begin reporting anomalous patterns. Contrast this with a GPU running continuous inference on camera frames, consuming its full power envelope regardless of whether the scene contains anything of interest.

Training spiking networks has historically been the primary obstacle to their practical adoption. The spike function is non-differentiable, meaning standard backpropagation cannot be applied directly. Neftci, Mostafa, and Zenke proposed the surrogate gradient method in 2019, replacing the true gradient of the spike function with a smooth approximation during the backward pass while retaining the discrete spike dynamics during the forward pass. This technique has proven effective, though as we observed in our experiments, convergence is notably slower than for equivalent continuous-valued architectures.

3. Experimental Design

3.1 Detection Modalities and Sensor Fusion

We evaluated three detection modalities, each feeding a dedicated encoder within the spiking network architecture. The first modality used photoelectric scatter sensor data sampled at 10 kHz, capturing the rapid fluctuations in optical obscuration that conventional detectors integrate away through their slow polling intervals. The second employed a 160 by 120 pixel thermal infrared camera producing event-based output, where each pixel independently reports temperature changes exceeding a configurable threshold rather than transmitting full frames at a fixed rate. The third utilised a multi-spectral gas sensor array comprising metal-oxide semiconductor elements sensitive to carbon monoxide, hydrogen cyanide, and volatile organic compounds, sampled asynchronously at event resolution.

Each modality's encoder consisted of a three-layer spiking convolutional network with 64, 128, and 256 channels respectively. The encoder outputs were merged through a cross-attention spiking layer that learned temporal correlations between modalities. When a rise in CO concentration and a thermal anomaly occurred within the same 500-microsecond window, for instance, the cross-attention mechanism amplified the combined signal far beyond what either modality alone would produce.

3.2 Hardware Platforms

The SNN was deployed on an Intel Loihi 2 development board configured with 128 neuromorphic cores, providing capacity for approximately one million spiking neurons and 120 million synapses. The baseline CNN, a ResNet-18 variant adapted for multi-modal sensor fusion, ran on an NVIDIA Jetson Orin Nano with 8 GB of shared memory and a 1024-core Ampere GPU operating at 625 MHz. Both platforms received identical sensor data streams for fair comparison.

3.3 Test Protocol

Fire scenarios followed the EN 54-7 standard test fire definitions: TF1 through TF8, encompassing open cellulosic fires, smouldering wood, smouldering cotton, flaming polyurethane, flaming n-heptane, methylated spirit fires, and cooking aerosol nuisance sources. All tests were conducted in a controlled burn facility with calibrated environmental conditions. We recorded 14,000 sensor event sequences across all modalities, supplemented by 2,200 nuisance events including cooking, steam, dust, and cigarette smoke. Metrics captured were detection latency from the moment of ignition, false alarm rate per source-hour of monitoring, and continuous power consumption in milliwatts.

4. Results

The headline numbers are striking. The Loihi 2 SNN achieved a median detection latency of 340 microseconds from the first anomalous sensor event, compared to 420 milliseconds for the Jetson Orin CNN baseline. That is a factor of approximately 1,235 times faster. Power consumption on the Loihi 2 averaged 23 milliwatts during active inference and dropped to 8 milliwatts during quiescent monitoring, versus a steady 7.4 watts on the Jetson Orin regardless of activity. The false alarm rate for the SNN stood at 0.08% per source-hour against 0.12% for the CNN.

The superior false alarm performance appears to stem from the SNN's ability to discriminate based on temporal dispersion patterns. Cooking aerosols, the most common nuisance source in commercial buildings, produce a sensor signature that is spatially similar to smouldering fire but temporally distinct. The particles arrive at the sensor in irregular bursts driven by convective plume turbulence, whereas genuine smoke from a developing fire produces a more monotonically increasing signal with characteristic acceleration. The frame-based CNN, which collapses temporal information into fixed windows, occasionally conflates these patterns. The SNN, processing each sensor event with microsecond temporal precision, consistently separates them.

One result genuinely surprised us. We had expected the SNN to outperform the CNN across all fire types, but for slow-developing smouldering fires, specifically TF2 and TF3, the SNN performed marginally worse. Its median detection latency for TF2 smouldering wood was 1.2 milliseconds versus 0.9 milliseconds for the SNN on flaming fires, and the CNN actually matched the SNN's latency on these scenarios at around 380 milliseconds once sufficient smoke had developed. The reason, we believe, is that smouldering fires produce gradual, low-contrast sensor changes that do not generate the sharp, well-defined events that spiking networks excel at processing. The evidence tentatively suggests that temporal coding is most advantageous when the underlying physical process has strong, rapid dynamics.

Training presented its own challenges. The SNN, trained with surrogate gradients on a workstation equipped with an NVIDIA A100 GPU, required approximately 72 hours to converge on our dataset. The equivalent CNN reached comparable accuracy in roughly 24 hours. This 3x training overhead is consistent with the findings reported in the surrogate gradient literature and stems from the additional complexity of optimising spike timing alongside synaptic weights. Once trained, however, the inference advantage is overwhelming.

5. Discussion

The 23-milliwatt power envelope changes what is architecturally possible. A lithium thionyl chloride battery with 19,000 milliamp-hour capacity at 3.6 volts, a chemistry commonly used in industrial sensor nodes, could power a Loihi 2 based fire detection node for approximately 8.1 years at continuous operation, or substantially longer with duty cycling. This makes wireless, battery-powered fire sensor nodes genuinely feasible for the first time, eliminating the cable infrastructure that represents a significant portion of installation cost in fire detection systems. For retrofit applications in heritage buildings where cable routing is prohibitively disruptive, this could be transformative.

Compliance with AS 7240 requires careful consideration. The standard mandates specific response time indices and alarm confirmation behaviours that assume a conventional detector-to-panel communication architecture. A neuromorphic sensor node operating autonomously would need to interface with existing fire indicator panels through standardised protocols, likely via a wireless gateway that translates SNN classification outputs into conventional alarm and pre-alarm signals. We see no fundamental barrier to compliance, but the regulatory pathway would require engagement with Standards Australia and likely a formal assessment against AS 7240.2 and AS 7240.7 for point-type smoke detectors.

It remains unclear whether the current generation of neuromorphic hardware is ready for volume deployment. The Loihi 2 development platform, while technically impressive, is produced in limited quantities and at price points that reflect its research-oriented market positioning. An individual Loihi 2 development board costs significantly more than a Jetson Orin Nano module, and Intel has not yet announced mass-production pricing. One could reasonably argue that the technology is three to five years from cost parity with conventional edge AI hardware for fire detection applications.

Our analysis revealed that a hybrid architecture may offer the best near-term path forward. In this configuration, a low-power neuromorphic front end handles first-response detection with microsecond latency and milliwatt power consumption, while a conventional GPU-based system provides secondary confirmation and detailed event classification. The neuromorphic layer would trigger an immediate local alert and wake the GPU subsystem from a deep sleep state, with the GPU then performing a more computationally intensive analysis to confirm or reject the alarm within one to two seconds. This layered approach captures most of the latency and power benefits while providing the classification depth that building operators and insurers increasingly demand.

6. Conclusion

Spiking neural networks on neuromorphic hardware deliver fire detection latency improvements of three orders of magnitude over conventional GPU-based inference, at roughly 1/320th of the power consumption. The temporal coding capabilities of SNNs provide a measurable advantage in discriminating between genuine fire signatures and nuisance sources, particularly those with distinct temporal dispersion patterns such as cooking aerosols. The technology is not without limitations: slow-developing smouldering fires remain a relative weakness, training times are substantially longer, and the hardware ecosystem is still maturing.

Nevertheless, the results presented here make a compelling case for neuromorphic fire detection as a serious research and development direction. The combination of sub-millisecond response, milliwatt power consumption, and superior nuisance rejection addresses three of the most persistent challenges in commercial fire detection simultaneously. As neuromorphic hardware moves from research platforms toward commercial silicon, we anticipate that the cost and availability barriers will diminish, opening the door to a new generation of intelligent, distributed fire sensing that is faster, more efficient, and more reliable than anything currently deployed.

References

Davies, M., Srinivasa, N., Lin, T.H. et al. (2018). "Loihi: A Neuromorphic Manycore Processor with On-Chip Learning." IEEE Micro, 38(1), 82–99.

Maass, W. (1997). "Networks of Spiking Neurons: The Third Generation of Neural Network Models." Neural Networks, 10(9), 1659–1671.

Neftci, E.O., Mostafa, H. & Zenke, F. (2019). "Surrogate Gradient Learning in Spiking Neural Networks." IEEE Signal Processing Magazine, 36(6), 51–63.

European Committee for Standardization. (2018). EN 54-7: Fire detection and fire alarm systems — Smoke detectors — Point smoke detectors using scattered light, transmitted light or ionisation. Brussels: CEN.

Standards Australia. (2016). AS 7240.2: Fire detection and alarm systems — Control and indicating equipment. Sydney: SAI Global.

Standards Australia. (2018). AS 7240.7: Fire detection and alarm systems — Point-type smoke detectors using scattered light, transmitted light or ionisation. Sydney: SAI Global.

Intel Corporation. (2024). Loihi 2: Architecture and Technical Reference Manual. Santa Clara: Intel Labs.

Orchard, G., Frady, E.P., Rubin, D.B.D. et al. (2021). "Efficient Neuromorphic Signal Processing with Loihi 2." Proceedings of the IEEE Workshop on Signal Processing Systems, 254–259.