## Deep Learning for Biodiversity Monitoring Note: Titel der Arbeit --- ## Outline ----
Note: Present the structure of the 30-minute talk. * Presentation - Topic - Objectives - Methodology - Results - Outlook * Discussion * Feedback --- ## Topic ---- ### What I signed up for "Tiere in Fotofallendataset mit KI automatisch erkennen"
Image: Bavarian State Institute of Forestry (LWF)
---- ### What it turned into "Deep Learning for Biodiversity Monitoring: Automated Classification of Small Mammals Captured in Foto Trap Boxes"
Figure: Author's own example
Note: already way in the core of the topic ---- ### Biodiversity
Figure: Global trends in biodiversity loss (Brondízio et al., 2019, IPBES)
Source: Brondízio, E. S., Settele, J., Díaz, S., & Ngo, H. T. (Eds.). (2019).
The Global Assessment Report of the Intergovernmental Science-Policy Platform on Biodiversity and Ecosystem Services
. IPBES, Bonn.
Note: This figure illustrates long-term declines across vertebrate groups. ---- ### Small Mammals ecologically important and often overlooked
Image: From the Mammalia Dataset
---- ### Wildlife@Campus Project
Figure: Author's own
---- ### Why Automation Matters 📷 **Millions of Images** ⚙️ **Automation is Key**
Image: Illustration by ChatGPT 4o
--- ## Objectives ---- ### These were the core objectives:
--- ## Methodology ----
----
Figure: Percentage of images discarded
Note: Explain fig well ----
---- ### Tested Architectures
EfficientNet-B0
4M
Scaled CNN baseline
DenseNet-169
12M
Dense CNN, feature reuse
ResNet-50
23M
Residual blocks, deep
ViT-B/16
85M
Transformer, patch-wise
pretrained and from scratch Note: **EfficientNet-B0**
Scales depth, width, and resolution in a balanced way
Lightweight baseline with ~4 million parameters
**DenseNet-169**
Uses dense connections between layers (each layer receives input from all previous layers)
Enables feature reuse and efficient gradient flow, ~12 million parameters
**ResNet-50**
Deep architecture with residual (skip) connections
Robust performance, ~23 million parameters
**ViT-B/16**
Vision Transformer using 16×16 image patches
Uses self-attention (each patch attends to all others, capturing global context), ~85 million parameters ---- ### Cross Validation
---- ### Evaluation
Note: Why different aggregation approaches — to retain information about distribution of fold performance. **Balanced Accuracy**: Average of recall scores across classes. Useful for imbalanced datasets. **Accuracy**: Overall proportion of correctly classified samples. Can be misleading if classes are imbalanced. **Recall**: True positives / (true positives + false negatives). Measures how well a class is detected (sensitivity). **F1 Score**: Harmonic mean of precision and recall. Balances false positives and false negatives. **Support**: Number of ground truth samples per class. Indicates class distribution in the test set. --- ## Results ---- ### Comparing Different Model Architectures
Table: Balanced accuracy of all models – mean ± standard deviation; best values highlighted
Figure: Balanced accuracy across folds
---- ### Pretrained EfficientNet-B0
Table: Class-wise precision, recall, F1-score, and support for the pretrained EfficientNet-B0
Figure: Confusion matrix EfficientNet-B0
---- ### Stoats: hard to detect – easy to classify
Figure: Not detected stoats
---- ### Stoats: hard to detect – easy to classify
Figure: Classification examples
---- ### Looking into some errors
Figure: Detected snails – classified as mammals
---- ### Looking into some errors
Figure: Missed detections
--- ## Outlook ---- ### Directions for Improvements
Notes: - Introduce a non-target class for OOD detection - Add additional species, e.g. _Glis glis_ - Improve detection quality, e.g. via fine-tuning - Explore sequence-aware or temporally informed classification approaches ---- ### Automated Sequence Detection
Utilizing OCR for sequence detection:
Figure: Top strip of a random sample
Output string was: _2019-09-04 1:02:09 AM M 1/3 #9 10°C_ --- ## Discussion ----
--- ## Feedback Note: - Fedback to my supervisors - thank the audience - feedback to me