You can find our paper on: https://arxiv.org/abs/2309.12568
For a visual overview of our paper please, visit: https://www.youtube.com/watch?v=5j8mAK9ecjs
You can download SCAND ROSBAGS from the
https://dataverse.tdl.org/dataset.xhtml?persistentId=doi:10.18738/T8/0PRYRH.
All this files will be used to extract necessary sensor data for training the model.
You can use wget to download files with their corresponding URL.
- For parsing data, create a folder
recorded-dataandbagfilesfolder in the root. - Place all the rosbag files in
bagfilesdirectory. - Run
/scripts/parser/parse_runner.py - All recorded file will be parsed inside
recorded-datafolder.
This step will parse all the necessary information from rosbag files for training.
In the recorded-data folder you will be able to see all RGB Images and a snapshot.pickle file which contains LiDAR and other necessary information.
Corresponding to each rosbag file, there should be folder in recorded-data.
Once all the training data are parsed create two folders inside recorded-data that are train and val.
You can split the parsed folder in recorded-data between these two directory to create appropriate split.
Refer to the labels from SCAND ROSBAGS to identify different social scenarios to split the data appropriately.
-
Once you have created the split you are ready to train the model.
-
Run
/scripts/multimodal.pyto start the training process. -
The code uses
comet.mlto track all the training metrics. -
You can turn off all the experiment logs, if prefer to do the training without any monitoring.
-
If you wish to use
comet.mlreplace theAPI_KEYwith yourAPI_KEYkey. -
Visit
https://www.comet.com/docs/v2/api-and-sdk/rest-api/overview/#obtaining-your-api-keyto get yourAPI_KEY.
- The model will be saved after an interval of 10 epochs. You can modify
multimodal.pyto store the model at appropriate checkpoint. - The testing inference will run at an interval of 2 epochs.
