article / 24 June 2025

BoutScout – Beyond AI for Images, Detecting Avian Behaviour with Sensors

In this case, you’ll explore how the BoutScout project is improving avian behavioural research through deep learning—without relying on images or video. By combining dataloggers, open-source hardware, and a powerful BiLSTM deep learning model trained on temperature data, the team has reduced the time needed to analyse weeks or even months of incubation behaviour to just seconds. This has enabled the discovery of new patterns in how tropical birds incubate their eggs across different elevations and climates. With tools soon to be released on PyPI, including a no-code platform for behavioural analysis, this work offers a fresh, scalable approach for conservationists and researchers working on breeding data at the tropics.

This case study presents an alternative route for studying avian incubation behaviour. While camera traps and visual recognition models dominate much of the conservation technology landscape, they are not always viable, especially in remote or densely vegetated ecosystems. In such settings, sensor-based approaches can offer a powerful, scalable, and less intrusive solution.

Here, researchers in Colombia (@jcguerra10, @bmgarrido, and me) introduce a modular, open-source system that combines environmental sensors with a deep learning model to detect incubation behaviour in wild birds. Instead of manually labelling thousands of hours of footage, our method uses simple temperature loggers to identify critical behaviours—when a parent bird enters or leaves the nest (on- and off-bouts) and when it initiates nocturnal incubation

The core of the system is a Bidirectional Long Short-Term Memory (BiLSTM) model, trained on over 2,200 nest-days from more than 120 Neotropical bird species. The model was developed following a three-step validation process: an initial 80/20 training-test split, followed by five-fold stratified cross-validation, and a final evaluation on an independent dataset. This robust methodology ensured reliable generalisation and performance, with theoretical error rates below 5%, based on macro F1 scores consistently around 0.95. These results suggest the model can accurately classify key incubation behaviours—on-bouts, off-bouts, and nocturnal incubation—even across species and ecosystems not included in the training data.

This figure shows how well the AI model performs when identifying different incubation behaviours in birds, using a robust validation strategy. (a) The graph illustrates the training loss over 5,600 epochs (learning cycles) during a 5-fold cross-validation. The black line represents the average performance across all folds, with shaded areas showing the variation. The model achieved rapid and stable learning by around epoch 30, meaning it quickly adapted to recognise behavioural patterns with minimal overfitting.(b) This panel displays precision–recall curves for each behavioural category: nocturnal incubation (black), off-bouts (blue), and on-bouts (orange). These curves highlight the model’s strong ability to distinguish each class, with macro-averaged F1 scores between 0.94 and 0.95, indicating excellent balance between precision (avoiding false positives) and recall (capturing all true behaviours). (c) The confusion matrices show how accurately the model classified each behaviour across all folds. True positives were highest for nocturnal behaviour (97–98%), while on- and off-bouts also showed high accuracy (91–95%). Most errors occurred between on- and off-bouts—cases that even trained biologists sometimes find hard to distinguish.Together, these results demonstrate that the model performs reliably and with very low error rates, offering a practical and non-invasive solution for analysing incubation behaviour at scale.

Exploratory Findings

In colaboration with two projects  Manu and Gradientes Colombia (leadered by PhD Gustavo Londoño), researchers also tested this method to study incubation strategies in Myioborus miniatus and M. melanocephalus, species distributed along Andean gradients. The model processed 328 full nest-days from both warbler species in just 28 seconds, with a classification accuracy exceeding 99%.This approach is not only accurate—it’s transformative. A task that used to take days or weeks of human effort is now fully automated, making large-scale behavioural analysis feasible, even in data-limited contexts.  

Exploratory analysis of incubation strategies in Myioborus miniatus and M. melanocephalus. (a) Principal Component Analysis (PCA) based on ecological, morphological, and behavioural variables, with group partitioning based on K-means clustering applied to the first three principal components (78.98% of total variance). The optimal number of clusters was selected by maximizing the silhouette score. The resulting clusters aligned strongly with species identity (Chi² = 153.97, p < 0.0001), with M. miniatus in dark violet and M. melanocephalus in dark orange. (b) Comparison of average on- (U = 8036.5, p = 0.0032) and off-bout durations (U = 9238.5, p < 0.0001) between species. Significant differences were detected using Mann–Whitney U tests. (c) Relationship between egg mass and elevation (left), showing a significant positive trend (GAM: pseudo R² = 0.2172, p < 0.001), and species-level comparison of egg mass (right), revealing significantly heavier eggs in M. melanocephalus (Mann–Whitney U = 1106.0, p < 0.0001). Asterisks in boxplots denote statistically significant differences between groups, * and ** indicate distinct groups supported by Mann–Whitney U tests. Illustrations made in watercolour by Lizarazo.
Exploratory analysis of incubation strategies in Myioborus miniatus and M. melanocephalus. (a) Principal Component Analysis (PCA) based on ecological, morphological, and behavioural variables, with group partitioning based on K-means clustering applied to the first three principal components (78.98% of total variance). The optimal number of clusters was selected by maximizing the silhouette score. The resulting clusters aligned strongly with species identity (Chi² = 153.97, p < 0.0001), with M. miniatus in dark violet and M. melanocephalus in dark orange. (b) Comparison of average on- (U = 8036.5, p = 0.0032) and off-bout durations (U = 9238.5, p < 0.0001) between species. Significant differences were detected using Mann–Whitney U tests. (c) Relationship between egg mass and elevation (left), showing a significant positive trend (GAM: pseudo R² = 0.2172, p < 0.001), and species-level comparison of egg mass (right), revealing significantly heavier eggs in M. melanocephalus (Mann–Whitney U = 1106.0, p < 0.0001). Asterisks in boxplots denote statistically significant differences between groups, * and ** indicate distinct groups supported by Mann–Whitney U tests. Illustrations made in watercolour by Lizarazo.

For decades, analysing bird incubation behaviour from nest temperature or camera traps data has been a slow and labour-intensive process. Early methods with nest temperature required biologists to manually review spreadsheets, spending up to eight hours just to annotate a week's worth of data for a single nest. Even with the advent of tools like RAVEN, incR, or NestIQ, analysis times remained lengthy—up to eight hours per 15 days of data—and still depended on manual adjustments and calibration. This not only made large-scale studies impractical but also introduced room for human error. These challenges highlighted a clear need for faster, more reliable tools capable of handling diverse datasets without the burden of constant fine-tuning.

How to Use It

The trained model is available on GitHub:
👉 GitHub Repo

  • You can load your data and perform inference in PyTorch using the .pth file at the final model.
  • Basic instructions and examples are included in the repository.

 

Why It Matters

By eliminating the need for complex visual data, this sensor-driven approach makes behaviour modelling more accessible, especially in the tropics. It lowers the cost, increases the scalability, and preserves the privacy of sensitive ecosystems. The use of open-source tools also ensures that other researchers can adapt and improve upon this work, including for new behavioural traits or additional sensors.

_______________________________________________________________________________________________________

🦉 Coming soon: the model will be available as a PyPI package with a Dash-based web interface, enabling use with zero coding required.

Let us know how you’d like to apply it in your own work—or if you want to contribute to future versions!

_________________________________________________________________________________________________________

Caveats

The model's performance is fundamentally tied to its training data. Accuracy may decrease when applied to species, climates, or sensor configurations significantly different from our Neotropical dataset. The model's reliance solely on temperature can be a limitation in environments where ambient temperatures are high, obscuring the thermal signatures of off-bouts. Future work should aim to incorporate complementary data streams from pressure to motion sensors, for which our modular architecture is well-suited. To enhance its impact and accessibility, our immediate goal is to package BoutScout into a user-friendly tool. 

 

We thank Universidad Icesi and PhD Gustavo A Londoño who guided us throughout this journey—for their unwavering support, mentorship, and belief in the potential of this work. Also a thanks to Angie Pamela López and Mariana Quintero, both experienced conservation biologists, provided invaluable assistance in labelling incubation behaviours across over 200 nests, ensuring the quality of the training dataset.


Add the first post in this thread.

Want to share your own conservation tech experiences and expertise with our growing global community? Login or register to start posting!