You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+10-6Lines changed: 10 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -15,15 +15,15 @@ Keep the directory structure the way it is and put your data in the 'audio' and
15
15
│ ├── 01_generate_spectrograms.ipynb
16
16
│ ├── ...
17
17
│ └── ...
18
-
├── audio <- ! put your input soundfiles in this folder !
18
+
├── audio <- ! put your input soundfiles in this folder or unzip the provided example audio!
19
19
│ ├── call_1.wav
20
20
│ ├── call_2.wav
21
21
│ └── ...
22
22
├── functions <- contains functions that will be called in analysis scripts
23
23
│ ├── audio_functions.py
24
24
│ ├── ...
25
25
│ └── ...
26
-
├── data <- ! put a .csv metadata file of your input in this folder !
26
+
├── data <- ! put a .csv metadata file of your input in this folder or use the provided example csv!
27
27
│ └── info_file.csv
28
28
├── parameters
29
29
│ └── spec_params.py <- this file contains parameters for spectrogramming (fft_win, fft_hop...)
@@ -125,10 +125,14 @@ and select the first jupyter notebook file to start your analysis (see section "
125
125
126
126
#### 2.2.1. Audio files
127
127
128
-
Use the provided example data (LINK) or put your own dataset of sound files in the subfolder "/audio" (make sure that the /audio folder contains __only__ your input files, nothing else).
129
-
Each sound file should contain a single vocalization or syllable.
128
+
All audio input files need to be in a subfolder /audio. This folder should not contain any other files.
129
+
130
+
To use the provided example data of meerkat calls, please unzip the file 'audio_please_unzip.zip' and verify that all audio files have been unpacked into an /audio folder according to the structure described in Section1.
131
+
132
+
To use your own data, create a subfolder "/audio" and put your sound files there (make sure that the /audio folder contains __only__ your input files, nothing else). Each sound file should contain a single vocalization or syllable.
130
133
(You may have to detect and extract such vocal elements first, if working with acoustic recordings.)
131
134
135
+
132
136
Ideally, start and end of the sound file correspond exactly to start and end of the vocalization.
133
137
If there are delays in the onset of the vocalizations, these should be the same for all sound files.
134
138
Otherwise, vocalizations may appear dissimilar or distant in latent space simply because their onset times are different.
@@ -137,10 +141,10 @@ but note that it comes at the cost of increased computation time.
137
141
138
142
#### 2.2.2. [Optional: Info file]
139
143
140
-
Use the .csv file provided with the example data (LINK) or, if you are using your own data,´ add a ";"-separated .csv file with headers containing the filenames of the input audio, some labels and any other additional metadata (if available) in the subfolder "/data".
144
+
Use the provided info_file.csv file for the example audio data or, if you are using your own data,´ add a ";"-separated info_file.csv file with headers containing the filenames of the input audio, some labels and any other additional metadata (if available) in the subfolder "/data".
141
145
If some or all labels are unknown, there should still be a label column and unkown labels should be marked with "unknown".
0 commit comments