Skip to content

Commit 696737e

Browse files
author
Greg Cohen
committed
Added 8 new datasets
1 parent 08f61eb commit 696737e

File tree

7 files changed

+1262
-0
lines changed

7 files changed

+1262
-0
lines changed

datasets/APEIRON.md

Lines changed: 176 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,176 @@
1+
---
2+
{
3+
"name": "APEIRON",
4+
"aliases": [],
5+
"year": 2024,
6+
"modalities": [
7+
"Vision"
8+
],
9+
"sensors": [
10+
"Prophesee Gen4"
11+
],
12+
"other_sensors": [
13+
"Stereolabs ZED2i",
14+
"PixHawk 2.4.8"
15+
],
16+
"category": "Robotic and Moving Vehicle Datasets",
17+
"tags": [
18+
"Flying Datasets",
19+
"Drone Detection"
20+
],
21+
"description": "Multimodal aerial dataset with Wireless Signals",
22+
"dataset_properties": {
23+
"available_online": true,
24+
"has_real_data": true,
25+
"has_simulated_data": false,
26+
"has_ground_truth": true,
27+
"has_frames": true,
28+
"has_biases": true,
29+
"distribution_methods": [
30+
"Direct Download"
31+
],
32+
"file_formats": [
33+
"Binary"
34+
],
35+
"availability_comment": "",
36+
"dataset_links": [
37+
{
38+
"name": "Direct Download",
39+
"url": "https://c3lab.github.io/Apeiron/",
40+
"format": "Binary",
41+
"available": true
42+
}
43+
],
44+
"size_gb": 34.2,
45+
"size_type": "Compressed"
46+
},
47+
"paper": {
48+
"title": "APEIRON",
49+
"doi": "10.1145/3625468.3652186",
50+
"authors": [
51+
"Nunzio Barone",
52+
"Walter Brescia",
53+
"Saverio Mascolo",
54+
"Luca De Cicco"
55+
],
56+
"abstract": "Unmanned Aerial Vehicles (UAVs), commonly denoted as drones, are being increasingly adopted as platforms to enable applications such as surveillance, disaster response, environmental monitoring, live drone broadcasting, and Internet-of-Drones (IoD). In this context, drone systems are required to carry out tasks autonomously in potentially unknown and challenging environments. As such, deep learning algorithms are widely adopted to implement efficient perception from sensors, making the availability of comprehensive datasets capturing real-world environments important. In this work, we introduce APEIRON, a rich multimodal aerial dataset that simultaneously collects perception data from a stereocamera and an event based camera sensor, along with measurements of wireless network links obtained using an LTE module. The assembled dataset consists of both perception and network data, making it suitable for typical perception or communication applications, as well as cross-disciplinary applications that require both types of data. We believe that this dataset will help promoting multi-disciplinary research at the intersection of multimedia systems, computer networks, and robotics fields. APEIRON is available at https://c3lab.github.io/Apeiron/.",
57+
"open_access": true
58+
},
59+
"citation_counts": [
60+
{
61+
"source": "crossref",
62+
"count": 2,
63+
"updated": "2025-09-13T15:43:31.596979"
64+
}
65+
],
66+
"links": [
67+
{
68+
"type": "paper",
69+
"url": "https://dl.acm.org/doi/abs/10.1145/3625468.3652186"
70+
},
71+
{
72+
"type": "github_page",
73+
"url": "https://github.com/c3lab/Apeiron"
74+
}
75+
],
76+
"full_name": "",
77+
"additional_metadata": {},
78+
"referenced_papers": [
79+
{
80+
"doi": "10.1016/j.petrol.2021.109633",
81+
"source": "crossref"
82+
},
83+
{
84+
"doi": "10.1109/ICRA.2019.8793887",
85+
"source": "crossref"
86+
},
87+
{
88+
"doi": "10.1145/3522784.3522786",
89+
"source": "crossref"
90+
},
91+
{
92+
"doi": "10.1109/TVT.2023.3318641",
93+
"source": "crossref"
94+
},
95+
{
96+
"doi": "10.1109/TPAMI.2020.3008413",
97+
"source": "crossref"
98+
},
99+
{
100+
"doi": "10.1109/ACCESS.2016.2537208",
101+
"source": "crossref"
102+
},
103+
{
104+
"doi": "10.1145/3083187.3083214",
105+
"source": "crossref"
106+
},
107+
{
108+
"doi": "10.1109/TASE.2020.3016276",
109+
"source": "crossref"
110+
},
111+
{
112+
"doi": "10.1109/ACCESS.2022.3220694",
113+
"source": "crossref"
114+
},
115+
{
116+
"doi": "10.1109/MCOM.005.2200784",
117+
"source": "crossref"
118+
},
119+
{
120+
"doi": "10.1109/ICUAS57906.2023.10156376",
121+
"source": "crossref"
122+
},
123+
{
124+
"doi": "10.1109/ICCCN52240.2021.9522344",
125+
"source": "crossref"
126+
},
127+
{
128+
"doi": "10.1109/TVT.2021.3126229",
129+
"source": "crossref"
130+
},
131+
{
132+
"doi": "10.1016/j.comcom.2020.03.012",
133+
"source": "crossref"
134+
},
135+
{
136+
"doi": "10.1109/JIOT.2022.3184323",
137+
"source": "crossref"
138+
},
139+
{
140+
"doi": "10.1109/VTCSpring.2019.8746290",
141+
"source": "crossref"
142+
},
143+
{
144+
"doi": "10.1109/MRA.2011.943233",
145+
"source": "crossref"
146+
},
147+
{
148+
"doi": "10.1109/ICDCS.2017.200",
149+
"source": "crossref"
150+
},
151+
{
152+
"title": "Cooperative Exploration of Heterogeneous UAVs in Mountainous Environments by Constructing Steady Communication",
153+
"source": "crossref"
154+
}
155+
],
156+
"bibtex": {
157+
"collection": "MMSys \u201924",
158+
"pages": "401\u2013407",
159+
"month": "apr",
160+
"year": 2024,
161+
"author": "Barone, Nunzio and Brescia, Walter and Mascolo, Saverio and De Cicco, Luca",
162+
"publisher": "ACM",
163+
"booktitle": "Proceedings of the ACM Multimedia Systems Conference 2024 on ZZZ",
164+
"doi": "10.1145/3625468.3652186",
165+
"url": "http://dx.doi.org/10.1145/3625468.3652186",
166+
"title": "APEIRON: a Multimodal Drone Dataset Bridging Perception and Network Data in Outdoor Environments",
167+
"series": "MMSys \u201924",
168+
"type": "inproceedings",
169+
"key": "Barone_2024"
170+
}
171+
}
172+
---
173+
174+
# Dataset Description
175+
176+
APEIRON is a rich multimodal aerial dataset that simultaneously collects perception data from a stereocamera and an event based camera sensor, along with measurements of wireless network links obtained using an LTE module. The assembled dataset consists of both perception and network data, making it suitable for typical perception or communication applications, as well as cross-disciplinary applications that require both types of data.

datasets/CoSEC.md

Lines changed: 139 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,139 @@
1+
---
2+
{
3+
"name": "CoSEC",
4+
"aliases": [],
5+
"year": 2024,
6+
"modalities": [
7+
"Vision"
8+
],
9+
"sensors": [
10+
"Prophesee Gen4"
11+
],
12+
"other_sensors": [
13+
"FLIR Blackfly S USB3",
14+
"Velodyne VLP-32C",
15+
"InertialLabs INS-D-E1"
16+
],
17+
"category": "Robotic and Moving Vehicle Datasets",
18+
"tags": [
19+
"Beamsplitters",
20+
"Autonomous Driving",
21+
"Stereo"
22+
],
23+
"description": "Stereo Autonomous Driving Dataset",
24+
"dataset_properties": {
25+
"available_online": false,
26+
"has_real_data": true,
27+
"has_simulated_data": false,
28+
"has_ground_truth": true,
29+
"has_frames": true,
30+
"has_biases": false,
31+
"distribution_methods": [],
32+
"file_formats": [],
33+
"availability_comment": "No download link in paper. No assosciated project page or Github repository.",
34+
"dataset_links": [],
35+
"size_gb": 0.0,
36+
"size_type": "Compressed"
37+
},
38+
"paper": {
39+
"title": "CoSEC: A Coaxial Stereo Event Camera Dataset for Autonomous Driving",
40+
"doi": "10.48550/arXiv.2408.08500",
41+
"authors": [
42+
"Shihan Peng",
43+
"Hanyu Zhou",
44+
"Hao Dong",
45+
"Zhiwei Shi",
46+
"Haoyue Liu",
47+
"Yuxing Duan",
48+
"Yi Chang",
49+
"Luxin Yan"
50+
],
51+
"abstract": "Conventional frame camera is the mainstream sensor of the autonomous driving scene perception, while it is limited in adverse conditions, such as low light. Event camera with high dynamic range has been applied in assisting frame camera for the multimodal fusion, which relies heavily on the pixel-level spatial alignment between various modalities. Typically, existing multimodal datasets mainly place event and frame cameras in parallel and directly align them spatially via warping operation. However, this parallel strategy is less effective for multimodal fusion, since the large disparity exacerbates spatial misalignment due to the large event-frame baseline. We argue that baseline minimization can reduce alignment error between event and frame cameras. In this work, we introduce hybrid coaxial event-frame devices to build the multimodal system, and propose a coaxial stereo event camera (CoSEC) dataset for autonomous driving. As for the multimodal system, we first utilize the microcontroller to achieve time synchronization, and then spatially calibrate different sensors, where we perform intra- and inter-calibration of stereo coaxial devices. As for the multimodal dataset, we filter LiDAR point clouds to generate depth and optical flow labels using reference depth, which is further improved by fusing aligned event and frame data in nighttime conditions. With the help of the coaxial device, the proposed dataset can promote the all-day pixel-level multimodal fusion. Moreover, we also conduct experiments to demonstrate that the proposed dataset can improve the performance and generalization of the multimodal fusion.",
52+
"open_access": false
53+
},
54+
"citation_counts": [
55+
{
56+
"source": "scholar",
57+
"count": 3,
58+
"updated": "2025-09-13T12:38:03.416087"
59+
}
60+
],
61+
"links": [
62+
{
63+
"type": "preprint",
64+
"url": "https://arxiv.org/abs/2408.08500"
65+
}
66+
],
67+
"full_name": "",
68+
"additional_metadata": {
69+
"num_recordings": "128",
70+
"total_duration": "753 seconds"
71+
},
72+
"bibtex": {
73+
"copyright": "arXiv.org perpetual, non-exclusive license",
74+
"year": 2024,
75+
"publisher": "arXiv",
76+
"title": "CoSEC: A Coaxial Stereo Event Camera Dataset for Autonomous Driving",
77+
"keywords": "Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences",
78+
"author": "Peng, Shihan and Zhou, Hanyu and Dong, Hao and Shi, Zhiwei and Liu, Haoyue and Duan, Yuxing and Chang, Yi and Yan, Luxin",
79+
"url": "https://arxiv.org/abs/2408.08500",
80+
"doi": "10.48550/ARXIV.2408.08500",
81+
"type": "misc",
82+
"key": "https://doi.org/10.48550/arxiv.2408.08500"
83+
}
84+
}
85+
---
86+
87+
# Dataset Structure
88+
89+
### Table
90+
91+
#### Train
92+
93+
| Time | Area | Sequences | Duration (s) |
94+
|-------|---------|-----------|--------------|
95+
| Day | City | 11 | 320 |
96+
| | Campus | 12 | 338 |
97+
| | Park | 11 | 314 |
98+
| | Suburbs | 15 | 411 |
99+
| | Village | 9 | 269 |
100+
| | **Total Day** | **58** | **1652** |
101+
| Night | City | 9 | 253 |
102+
| | Campus | 10 | 287 |
103+
| | Park | 8 | 231 |
104+
| | Suburbs | 12 | 334 |
105+
| | Village | 5 | 148 |
106+
| | **Total Night** | **44** | **1253** |
107+
108+
**Total Train**
109+
Sequences: **102**
110+
Duration (s): **2905**
111+
112+
---
113+
114+
#### Test
115+
116+
| Time | Area | Sequences | Duration (s) |
117+
|-------|---------|-----------|--------------|
118+
| Day | City | 2 | 58 |
119+
| | Campus | 3 | 94 |
120+
| | Park | 3 | 84 |
121+
| | Suburbs | 5 | 138 |
122+
| | Village | 2 | 57 |
123+
| | **Total Day** | **15** | **431** |
124+
| Night | City | 2 | 61 |
125+
| | Campus | 2 | 66 |
126+
| | Park | 2 | 56 |
127+
| | Suburbs | 3 | 85 |
128+
| | Village | 2 | 54 |
129+
| | **Total Night** | **11** | **322** |
130+
131+
**Total Test**
132+
Sequences: **26**
133+
Duration (s): **753**
134+
135+
---
136+
137+
### Grand Total
138+
Sequences: **128**
139+
Duration (s): **3658** seconds.

0 commit comments

Comments
 (0)