Multichannel Alignment


Challenge results

More details can be found on the task page:

Task description page

Teams ranking

Submission information Rank Test set Validation
Rank Submission label Name Technical
Report
Official
rank
Rank
score
MSE (avg) MSE (aru) MSE (Zebra Finch) MSE (avg) MSE (aru) MSE (Zebra Finch)
NoSync Nosync Baseline biodcase2025Task1 3 1.35 1.35 0.85 1.84 1.15 0.98 1.31
CCBaseline Crosscorrelation Baseline biodcase2025Task1 5 3.32 3.32 1.10 5.55 8.45 6.86 10.03
DLBaseline Deep Learning Baseline biodcase2025Task1 2 0.58 0.58 0.62 0.55 0.89 0.52 1.26
landmarks Landmark based synchronization harjuLandmark2025 4 1.60 1.60 1.17 2.04 0.84 0.41 1.27
BEATsCA BEATs with Cross-Attention nihalBeats2025 1 0.30 0.30 0.14 0.45 0.31 0.10 0.52

Technical reports

BioDCASE Baleen Whale Deep Learning Detection and Classification Network

Alongi, Gabriela and Ferguson, Liz and Sugarman, Peter

Abstract

DeepAcoustics is a deep learning-based tool designed for the detection and classification of underwater acoustic signals, with a focus on marine mammal vocalizations. Developed with a modular architecture, DeepAcoustics allows users to configure, train, and apply custom neural network models for sound event detection. DeepAcoustics supports a range of deep learning architectures for acoustic detection and classification, including Tiny YOLO, ResNet, and Darknet-based networks, allowing users to select the model best suited for their specific data and computational needs.

PDF

BioDCASE 2025 Task 1

Hoffman, Benjamin and Gill, Lisa and Heath, Becky, and Narula, Gagan

Abstract

Coming soon!

Landmark-based synchronization for Drifting Audio Recordings

Harju, Manu and Mesaros, Annamaria
Tampere University

Abstract

This technical report describes our submissions to the multichannel alignment task in the BioDCASE 2025 Challenge. Our system is based on matching and aligning audio landmarks, which are simple structures extracted from the spectrogram representations. Our code and the configuration used is available on GitHub.

PDF

BEATs with Cross-Attention for Multi-Channel Audio Alignment

Nihal, Ragib Amin and Yen, Benjamin and Ashizawa, Takeshi and Nakadai, Kazuhiro
Institute of Science Tokyo

Abstract

We modified the provided BEATs baseline with three main changes. First, we added cross-attention layers that allow audio embeddings from different channels to interact before making alignment predictions. Second, improved the training process with better data sampling, conservative augmentation (amplitude scaling and noise addition), and AdamW optimization with learning rate scheduling. Third, replaced the baseline's binary counting similarity metric with confidence-weighted scoring that uses the full range of model outputs. The system uses the same candidate generation approach as the baseline but processes alignment decisions differently. On validation data, the method achieved MSE of 0.099 for ARU and 0.521 for zebra finch.

PDF