Skip to content

Commit 635f073

Browse files
authored
Merge pull request #424 from roboflow/fix/how-to-segment-videos-with-segment-anything-3
fix `How to Segment Videos with Segment Anything 3 (SAM3)` notebook
2 parents d76ef0f + 3db0ffa commit 635f073

File tree

1 file changed

+30
-44
lines changed

1 file changed

+30
-44
lines changed

notebooks/how-to-segment-videos-with-segment-anything-3.ipynb

Lines changed: 30 additions & 44 deletions
Original file line numberDiff line numberDiff line change
@@ -81,23 +81,23 @@
8181
"base_uri": "https://localhost:8080/"
8282
},
8383
"id": "ILPFDHXR8Cfh",
84-
"outputId": "b521ae11-7ada-4858-9fb6-b7978c157695"
84+
"outputId": "de1d8bb0-5268-40c2-e233-c416cccd7c1b"
8585
},
8686
"outputs": [
8787
{
88-
"name": "stdout",
8988
"output_type": "stream",
89+
"name": "stdout",
9090
"text": [
91-
"Thu Nov 20 12:06:00 2025 \n",
91+
"Tue Feb 3 16:05:05 2026 \n",
9292
"+-----------------------------------------------------------------------------------------+\n",
9393
"| NVIDIA-SMI 550.54.15 Driver Version: 550.54.15 CUDA Version: 12.4 |\n",
9494
"|-----------------------------------------+------------------------+----------------------+\n",
9595
"| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |\n",
9696
"| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |\n",
9797
"| | | MIG M. |\n",
9898
"|=========================================+========================+======================|\n",
99-
"| 0 NVIDIA A100-SXM4-80GB Off | 00000000:00:05.0 Off | 0 |\n",
100-
"| N/A 35C P0 57W / 400W | 0MiB / 81920MiB | 0% Default |\n",
99+
"| 0 NVIDIA A100-SXM4-40GB Off | 00000000:00:04.0 Off | 0 |\n",
100+
"| N/A 29C P0 43W / 400W | 0MiB / 40960MiB | 0% Default |\n",
101101
"| | | Disabled |\n",
102102
"+-----------------------------------------+------------------------+----------------------+\n",
103103
" \n",
@@ -124,15 +124,15 @@
124124
"base_uri": "https://localhost:8080/"
125125
},
126126
"id": "Vw3admXIqMlC",
127-
"outputId": "61d7c09c-6f72-41ce-edb8-9d7de99e208d"
127+
"outputId": "5420cfd5-332f-4ea4-8c31-c454940efe0e"
128128
},
129129
"outputs": [
130130
{
131-
"name": "stdout",
132131
"output_type": "stream",
132+
"name": "stdout",
133133
"text": [
134-
"PyTorch version: 2.8.0+cu126\n",
135-
"Torchvision version: 0.23.0+cu126\n",
134+
"PyTorch version: 2.9.0+cu126\n",
135+
"Torchvision version: 0.24.0+cu126\n",
136136
"CUDA is available: True\n"
137137
]
138138
}
@@ -176,9 +176,22 @@
176176
"execution_count": null,
177177
"id": "eM_DGtMk_gFM",
178178
"metadata": {
179-
"id": "eM_DGtMk_gFM"
179+
"id": "eM_DGtMk_gFM",
180+
"outputId": "8a3874c0-be38-4440-8052-c78885641d3b",
181+
"colab": {
182+
"base_uri": "https://localhost:8080/"
183+
}
180184
},
181-
"outputs": [],
185+
"outputs": [
186+
{
187+
"output_type": "stream",
188+
"name": "stdout",
189+
"text": [
190+
"\u001b[?25l \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m0.0/212.4 kB\u001b[0m \u001b[31m?\u001b[0m eta \u001b[36m-:--:--\u001b[0m\r\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m212.4/212.4 kB\u001b[0m \u001b[31m6.1 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
191+
"\u001b[?25h"
192+
]
193+
}
194+
],
182195
"source": [
183196
"!pip install -q supervision jupyter_bbox_widget"
184197
]
@@ -275,12 +288,12 @@
275288
"base_uri": "https://localhost:8080/"
276289
},
277290
"id": "QynbB5AgFhML",
278-
"outputId": "a59946c8-0c2a-4f4d-bf47-61a96e6f21ed"
291+
"outputId": "e139269c-6149-45e5-8bbd-f324fdbf9a11"
279292
},
280293
"outputs": [
281294
{
282-
"name": "stdout",
283295
"output_type": "stream",
296+
"name": "stdout",
284297
"text": [
285298
"HOME: /content\n"
286299
]
@@ -332,38 +345,11 @@
332345
"execution_count": null,
333346
"id": "y51dHwSKO2CX",
334347
"metadata": {
335-
"colab": {
336-
"base_uri": "https://localhost:8080/"
337-
},
338-
"id": "y51dHwSKO2CX",
339-
"outputId": "fd7ae78f-a881-41de-d193-3bab847f2cb8"
348+
"id": "y51dHwSKO2CX"
340349
},
341-
"outputs": [
342-
{
343-
"name": "stderr",
344-
"output_type": "stream",
345-
"text": [
346-
"\u001b[0m\u001b[32mINFO 2025-11-20 14:02:37,450 35343 sam3_video_predictor.py: 299:\u001b[0m using the following GPU IDs: [0]\n",
347-
"\u001b[0m\u001b[32mINFO 2025-11-20 14:02:37,564 35343 sam3_video_predictor.py: 315:\u001b[0m \n",
348-
"\n",
349-
"\n",
350-
"\t*** START loading model on all ranks ***\n",
351-
"\n",
352-
"\n",
353-
"\u001b[0m\u001b[32mINFO 2025-11-20 14:02:37,564 35343 sam3_video_predictor.py: 317:\u001b[0m loading model on rank=0 with world_size=1 -- this could take a while ...\n",
354-
"\u001b[0m\u001b[32mINFO 2025-11-20 14:02:46,745 35343 sam3_video_base.py: 124:\u001b[0m setting max_num_objects=10000 and num_obj_for_compile=16\n",
355-
"\u001b[0m\u001b[32mINFO 2025-11-20 14:02:51,644 35343 sam3_video_predictor.py: 319:\u001b[0m loading model on rank=0 with world_size=1 -- DONE locally\n",
356-
"\u001b[0m\u001b[32mINFO 2025-11-20 14:02:51,645 35343 sam3_video_predictor.py: 330:\u001b[0m \n",
357-
"\n",
358-
"\n",
359-
"\t*** DONE loading model on all ranks ***\n",
360-
"\n",
361-
"\n"
362-
]
363-
}
364-
],
350+
"outputs": [],
365351
"source": [
366-
"predictor = build_sam3_video_predictor(gpus_to_use=DEVICES)"
352+
"predictor = build_sam3_video_predictor(bpe_path=\"/content/sam3/sam3/assets/bpe_simple_vocab_16e6.txt.gz\", gpus_to_use=DEVICES)"
367353
]
368354
},
369355
{
@@ -1631,4 +1617,4 @@
16311617
},
16321618
"nbformat": 4,
16331619
"nbformat_minor": 5
1634-
}
1620+
}

0 commit comments

Comments
 (0)