|
5986 | 5986 | </span> |
5987 | 5987 | </a> |
5988 | 5988 |
|
| 5989 | +</li> |
| 5990 | + |
| 5991 | + <li class="md-nav__item"> |
| 5992 | + <a href="#training-a-yolox-model-for-autoware_traffic_light_fine_detector" class="md-nav__link"> |
| 5993 | + <span class="md-ellipsis"> |
| 5994 | + |
| 5995 | + Training a YOLOX model for autoware_traffic_light_fine_detector |
| 5996 | + |
| 5997 | + </span> |
| 5998 | + </a> |
| 5999 | + |
| 6000 | + <nav class="md-nav" aria-label="Training a YOLOX model for autoware_traffic_light_fine_detector"> |
| 6001 | + <ul class="md-nav__list"> |
| 6002 | + |
| 6003 | + <li class="md-nav__item"> |
| 6004 | + <a href="#example-dataset-voc-format" class="md-nav__link"> |
| 6005 | + <span class="md-ellipsis"> |
| 6006 | + |
| 6007 | + 📦 Example Dataset (VOC Format) |
| 6008 | + |
| 6009 | + </span> |
| 6010 | + </a> |
| 6011 | + |
| 6012 | +</li> |
| 6013 | + |
| 6014 | + </ul> |
| 6015 | + </nav> |
| 6016 | + |
5989 | 6017 | </li> |
5990 | 6018 |
|
5991 | 6019 | </ul> |
|
12928 | 12956 | </span> |
12929 | 12957 | </a> |
12930 | 12958 |
|
| 12959 | +</li> |
| 12960 | + |
| 12961 | + <li class="md-nav__item"> |
| 12962 | + <a href="#training-a-yolox-model-for-autoware_traffic_light_fine_detector" class="md-nav__link"> |
| 12963 | + <span class="md-ellipsis"> |
| 12964 | + |
| 12965 | + Training a YOLOX model for autoware_traffic_light_fine_detector |
| 12966 | + |
| 12967 | + </span> |
| 12968 | + </a> |
| 12969 | + |
| 12970 | + <nav class="md-nav" aria-label="Training a YOLOX model for autoware_traffic_light_fine_detector"> |
| 12971 | + <ul class="md-nav__list"> |
| 12972 | + |
| 12973 | + <li class="md-nav__item"> |
| 12974 | + <a href="#example-dataset-voc-format" class="md-nav__link"> |
| 12975 | + <span class="md-ellipsis"> |
| 12976 | + |
| 12977 | + 📦 Example Dataset (VOC Format) |
| 12978 | + |
| 12979 | + </span> |
| 12980 | + </a> |
| 12981 | + |
| 12982 | +</li> |
| 12983 | + |
| 12984 | + </ul> |
| 12985 | + </nav> |
| 12986 | + |
12931 | 12987 | </li> |
12932 | 12988 |
|
12933 | 12989 | </ul> |
@@ -13061,6 +13117,38 @@ <h2 id="training-centerpoint-3d-object-detection-model">Training CenterPoint 3D |
13061 | 13117 | <p>In order to assist you with your training process, we have also included an example dataset in the TIER IV dataset format.</p> |
13062 | 13118 | <p>This dataset contains 600 lidar frames and covers 5 classes, including 6905 cars, 3951 pedestrians, 75 cyclists, 162 buses, and 326 trucks.</p> |
13063 | 13119 | <p>You can utilize this example dataset to facilitate your training efforts.</p> |
| 13120 | +<h2 id="training-a-yolox-model-for-autoware_traffic_light_fine_detector">Training a YOLOX model for <code>autoware_traffic_light_fine_detector</code><a class="headerlink" href="#training-a-yolox-model-for-autoware_traffic_light_fine_detector" title="Permanent link">#</a></h2> |
| 13121 | +<p>To train a custom YOLOX model for use with the <strong>Autoware Traffic Light Fine Detector</strong>, please refer to the official YOLOX and Autoware training guides listed below. These documents provide the required setup, data preparation steps, and export instructions.</p> |
| 13122 | +<details class="abstract" open="open"> |
| 13123 | +<summary>Relevant documentation</summary> |
| 13124 | +<div class="grid cards" markdown> |
| 13125 | +<ul> |
| 13126 | +<li> |
| 13127 | +<p><strong>Autoware traffic_light_fine_detector README</strong></p> |
| 13128 | +<hr> |
| 13129 | +<p>Training overview, model requirements, and ONNX export details: |
| 13130 | +<a href="https://github.com/autowarefoundation/autoware_universe/tree/main/perception/autoware_traffic_light_fine_detector/README.md">autoware_universe/perception/autoware_traffic_light_fine_detector/README.md</a></p> |
| 13131 | +</li> |
| 13132 | +</ul> |
| 13133 | +<ul> |
| 13134 | +<li> |
| 13135 | +<p><strong>YOLOX Custom Dataset Training Guide</strong></p> |
| 13136 | +<hr> |
| 13137 | +<p>Instructions for preparing datasets, configuring experiments, and launching training: |
| 13138 | +<a href="https://github.com/Megvii-BaseDetection/YOLOX/blob/main/docs/train_custom_data.md">YOLOX/train_custom_data.md</a></p> |
| 13139 | +</li> |
| 13140 | +</ul> |
| 13141 | +</div> |
| 13142 | +</details> |
| 13143 | +<h3 id="example-dataset-voc-format">📦 Example Dataset (VOC Format)<a class="headerlink" href="#example-dataset-voc-format" title="Permanent link">#</a></h3> |
| 13144 | +<p>To assist with training, an example dataset containing <strong>1062 cropped traffic-light images with Pascal VOC annotations</strong> is available here:</p> |
| 13145 | +<p><a class="md-button" href="https://autoware-files.s3.us-west-2.amazonaws.com/dataset/traffic_light_detection_sample_dataset.tar.gz"><i class="fa-solid fa-cloud-arrow-down"></i> Download the sample traffic light dataset (3 MB)</a></p> |
| 13146 | +<ul> |
| 13147 | +<li>Use the YOLOX documentation to set up your environment and prepare your dataset (VOC or COCO).</li> |
| 13148 | +<li>Train a YOLOX model using your custom traffic-light data or the provided sample dataset.</li> |
| 13149 | +<li>After training, export the model to <strong>ONNX</strong> following the instructions in the Autoware <code>traffic_light_fine_detector</code> README.</li> |
| 13150 | +<li>Replace or integrate the exported ONNX model within the Autoware package for deployment.</li> |
| 13151 | +</ul> |
13064 | 13152 | <!-- prettier-ignore --></div> |
13065 | 13153 |
|
13066 | 13154 |
|
|
0 commit comments