@@ -46,18 +46,18 @@ Each CSV file includes the following columns:
46
46
<br >
47
47
48
48
| Dataset | Type | Classes | Split | Size |
49
- | :-- | :-- | :--: | :--: | : --: |
50
- | [ Beijing-Opera] ( #beijing-opera ) | Instrument Classification | 4 | Five-Fold | 68 MB
51
- | [ CREMA-D] ( #crema-d ) | Emotion Recognition | 6 | Train-Test | 653M
52
- | [ ESC50] ( #esc50 ) | Sound Event Classification | 50 | Five-Fold | 777M
53
- | [ ESC50-Actions] ( #esc50-actions ) | Sound Event Classification | 10 | Five-Fold | 772M
54
- | [ GT-Music-Genre] ( #gt-music-genre ) | Music Analysis | 10 | Train-Test | 1.4G
55
- | [ NS-Instruments] ( #ns-instruments ) | Instrument Classification | 10 | Train-Test | 14G
56
- | [ RAVDESS] ( #ravdess ) | Emotion Recognition | 8 | Train-Test | 683M
57
- | [ SESA] ( #sesa ) | Surveillance Sound Classification | 4 | Train-Test | 51M
58
- | [ TUT2017] ( #tut2017 ) | Acoustic Scene Classification | 15 | Four-Fold | 12G
59
- | [ UrbanSound8K] ( #urbansound8k ) | Sound Event Classification | 10 | Ten-Fold | 6.8G
60
- | [ VocalSound] ( #vocalsound ) | Vocal Sound Classification | 6 | Train-Test | 6.9G
49
+ | :-- | :-- | :--: | :--: | --: |
50
+ | [ Beijing-Opera] ( #beijing-opera ) | Instrument Classification | 4 | Five-Fold | 69 MB |
51
+ | [ CREMA-D] ( #crema-d ) | Emotion Recognition | 6 | Train-Test | 606 MB |
52
+ | [ ESC50] ( #esc50 ) | Sound Event Classification | 50 | Five-Fold | 881 MB |
53
+ | [ ESC50-Actions] ( #esc50-actions ) | Sound Event Classification | 10 | Five-Fold | 881 MB |
54
+ | [ GT-Music-Genre] ( #gt-music-genre ) | Music Analysis | 10 | Train-Test | 1.3 GB |
55
+ | [ NS-Instruments] ( #ns-instruments ) | Instrument Classification | 10 | Train-Test | 18.5 GB
56
+ | [ RAVDESS] ( #ravdess ) | Emotion Recognition | 8 | Train-Test | 1.1 GB |
57
+ | [ SESA] ( #sesa ) | Surveillance Sound Classification | 4 | Train-Test | 70 MB |
58
+ | [ TUT2017] ( #tut2017 ) | Acoustic Scene Classification | 15 | Four-Fold | 12.3 GB |
59
+ | [ UrbanSound8K] ( #urbansound8k ) | Sound Event Classification | 10 | Ten-Fold | 6.8 GB |
60
+ | [ VocalSound] ( #vocalsound ) | Vocal Sound Classification | 6 | Train-Test | 8.2 GB |
61
61
62
62
<br ><br >
63
63
<hr ><hr >
@@ -78,8 +78,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
78
78
huggingface_hub.snapshot_download(repo_id = " MahiA/Beijing-Opera" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " Beijing-Opera" ))
79
79
```
80
80
| Type | Classes | Split | Size |
81
- | :-- | :--: | :--: | : --: |
82
- | Instrument Classification | 4 | Five-Fold | 68 MB |
81
+ | :-- | :--: | :--: | --: |
82
+ | Instrument Classification | 4 | Five-Fold | 69 MB |
83
83
84
84
<br >
85
85
<hr >
@@ -94,8 +94,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
94
94
huggingface_hub.snapshot_download(repo_id = " MahiA/CREMA-D" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " CREMA-D" ))
95
95
```
96
96
| Type | Classes | Split | Size |
97
- | :-- | :--: | :--: | : --: |
98
- | Emotion Recognition | 6 | Train-Test | |
97
+ | :-- | :--: | :--: | --: |
98
+ | Emotion Recognition | 6 | Train-Test | 606 MB |
99
99
100
100
<br >
101
101
<hr >
@@ -110,8 +110,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
110
110
huggingface_hub.snapshot_download(repo_id = " MahiA/ESC50" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " ESC50" ))
111
111
```
112
112
| Type | Classes | Split | Size |
113
- | :-- | :--: | :--: | : --: |
114
- | Sound Event Classification | 50 | Five-Fold | |
113
+ | :-- | :--: | :--: | --: |
114
+ | Sound Event Classification | 50 | Five-Fold | 881 MB |
115
115
116
116
<br >
117
117
<hr >
@@ -126,8 +126,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
126
126
huggingface_hub.snapshot_download(repo_id = " MahiA/ESC50-Actions" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " ESC50-Actions" ))
127
127
```
128
128
| Type | Classes | Split | Size |
129
- | :-- | :--: | :--: | : --: |
130
- | Sound Event Classification | 10 | Five-Fold | |
129
+ | :-- | :--: | :--: | --: |
130
+ | Sound Event Classification | 10 | Five-Fold | 881 MB |
131
131
132
132
<br >
133
133
<hr >
@@ -142,8 +142,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
142
142
huggingface_hub.snapshot_download(repo_id = " MahiA/GT-Music-Genre" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " GT-Music-Genre" ))
143
143
```
144
144
| Type | Classes | Split | Size |
145
- | :-- | :--: | :--: | : --: |
146
- | Music Analysis | 10 | Train-Test | |
145
+ | :-- | :--: | :--: | --: |
146
+ | Music Analysis | 10 | Train-Test | 1.3 GB |
147
147
148
148
<br >
149
149
<hr >
@@ -158,8 +158,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
158
158
huggingface_hub.snapshot_download(repo_id = " MahiA/NS-Instruments" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " NS-Instruments" ))
159
159
```
160
160
| Type | Classes | Split | Size |
161
- | :-- | :--: | :--: | : --: |
162
- | Instrument Classification | 10 | Train-Test | |
161
+ | :-- | :--: | :--: | --: |
162
+ | Instrument Classification | 10 | Train-Test | 18.5 GB |
163
163
164
164
<br >
165
165
<hr >
@@ -174,8 +174,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
174
174
huggingface_hub.snapshot_download(repo_id = " MahiA/RAVDESS" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " RAVDESS" ))
175
175
```
176
176
| Type | Classes | Split | Size |
177
- | :-- | :--: | :--: | : --: |
178
- | Emotion Recognition | 8 | Train-Test | |
177
+ | :-- | :--: | :--: | --: |
178
+ | Emotion Recognition | 8 | Train-Test | 1.1 GB |
179
179
180
180
<br >
181
181
<hr >
@@ -190,8 +190,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
190
190
huggingface_hub.snapshot_download(repo_id = " MahiA/SESA" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " SESA" ))
191
191
```
192
192
| Type | Classes | Split | Size |
193
- | :-- | :--: | :--: | : --: |
194
- | Surveillance Sound Classification | 4 | Train-Test | |
193
+ | :-- | :--: | :--: | --: |
194
+ | Surveillance Sound Classification | 4 | Train-Test | 70 MB |
195
195
196
196
<br >
197
197
<hr >
@@ -206,8 +206,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
206
206
huggingface_hub.snapshot_download(repo_id = " MahiA/TUT2017" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " TUT2017" ))
207
207
```
208
208
| Type | Classes | Split | Size |
209
- | :-- | :--: | :--: | : --: |
210
- | Acoustic Scene Classification | 15 | Four-Fold | |
209
+ | :-- | :--: | :--: | --: |
210
+ | Acoustic Scene Classification | 15 | Four-Fold | 12.3 GB |
211
211
212
212
213
213
<br >
@@ -223,8 +223,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
223
223
huggingface_hub.snapshot_download(repo_id = " MahiA/UrbanSound8K" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " UrbanSound8K" ))
224
224
```
225
225
| Type | Classes | Split | Size |
226
- | :-- | :--: | :--: | : --: |
227
- | Sound Event Classification | 10 | Ten-Fold | |
226
+ | :-- | :--: | :--: | --: |
227
+ | Sound Event Classification | 10 | Ten-Fold | 6.8 GB |
228
228
229
229
<br >
230
230
<hr >
@@ -239,8 +239,8 @@ if not os.path.exists(audio_datasets_path): print(f"Given {audio_datasets_path=}
239
239
huggingface_hub.snapshot_download(repo_id = " MahiA/VocalSound" , repo_type = " dataset" , local_dir = os.path.join(audio_datasets_path, " VocalSound" ))
240
240
```
241
241
| Type | Classes | Split | Size |
242
- | :-- | :--: | :--: | : --: |
243
- | Vocal Sound Classification | 6 | Train-Test | |
242
+ | :-- | :--: | :--: | --: |
243
+ | Vocal Sound Classification | 6 | Train-Test | 8.2 GB |
244
244
245
245
<br >
246
246
<hr >
0 commit comments