anonymous-submission000 commited on
Commit
7dd635d
·
verified ·
1 Parent(s): 48d91cc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -12
README.md CHANGED
@@ -137,18 +137,6 @@ configs:
137
 
138
  **Bird3M** is the first synchronized, multi-modal, multi-individual dataset designed for comprehensive behavioral analysis of freely interacting birds, specifically zebra finches, in naturalistic settings. It addresses the critical need for benchmark datasets that integrate precisely synchronized multi-modal recordings to support tasks such as 3D pose estimation, multi-animal tracking, sound source localization, and vocalization attribution. The dataset facilitates research in machine learning, neuroscience, and ethology by enabling the development of robust, unified models for long-term tracking and interpretation of complex social behaviors.
139
 
140
- ### Key Features
141
- - **Duration**: 22.2 hours of synchronized multi-modal recordings, including a fully annotated subset with 4,420 video frames and 2.5 hours of contextual audio and sensor data.
142
- - **Modalities**:
143
- - **Multi-view video**: Three orthogonal color cameras (top, side, back) at 47 fps, supplemented by a monochrome nest camera.
144
- - **Multi-channel audio**: Wall-mounted microphones (16 kHz) and body-mounted accelerometers (24,414 Hz, down-sampled to 16 kHz).
145
- - **Radio signals**: FM radio phases and magnitudes from four orthogonal antennas.
146
- - **Annotations**:
147
- - **Visual**: 57,396 3D keypoints (5 per bird: beak tip, head center, backpack center, tail base, tail end) across 4,420 frames, with 2D keypoints, visibility labels, and bounding boxes.
148
- - **Audio**: 4,902 vocalization segments with onset/offset times and vocalizer identities, linked across microphone and accelerometer channels.
149
- - **Experimental Setup**: Data from 15 experiments (2–8 birds each) conducted in the **Birdpark** system, with sessions lasting 4–120 days.
150
- - **Applications**: Supports 3D localization, pose estimation, multi-animal tracking, sound source localization/separation, and cross-modal behavioral analyses (e.g., vocalization directness).
151
-
152
  ### Purpose
153
  Bird3M bridges the gap in publicly available datasets for multi-modal animal behavior analysis by providing:
154
  1. A benchmark for unified machine learning models tackling multiple behavioral tasks.
 
137
 
138
  **Bird3M** is the first synchronized, multi-modal, multi-individual dataset designed for comprehensive behavioral analysis of freely interacting birds, specifically zebra finches, in naturalistic settings. It addresses the critical need for benchmark datasets that integrate precisely synchronized multi-modal recordings to support tasks such as 3D pose estimation, multi-animal tracking, sound source localization, and vocalization attribution. The dataset facilitates research in machine learning, neuroscience, and ethology by enabling the development of robust, unified models for long-term tracking and interpretation of complex social behaviors.
139
 
 
 
 
 
 
 
 
 
 
 
 
 
140
  ### Purpose
141
  Bird3M bridges the gap in publicly available datasets for multi-modal animal behavior analysis by providing:
142
  1. A benchmark for unified machine learning models tackling multiple behavioral tasks.