8181 "base_uri": "https://localhost:8080/"
8282 },
8383 "id": "ILPFDHXR8Cfh",
84- "outputId": "b521ae11-7ada-4858-9fb6-b7978c157695 "
84+ "outputId": "de1d8bb0-5268-40c2-e233-c416cccd7c1b "
8585 },
8686 "outputs": [
8787 {
88- "name": "stdout",
8988 "output_type": "stream",
89+ "name": "stdout",
9090 "text": [
91- "Thu Nov 20 12:06:00 2025 \n",
91+ "Tue Feb 3 16:05:05 2026 \n",
9292 "+-----------------------------------------------------------------------------------------+\n",
9393 "| NVIDIA-SMI 550.54.15 Driver Version: 550.54.15 CUDA Version: 12.4 |\n",
9494 "|-----------------------------------------+------------------------+----------------------+\n",
9595 "| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |\n",
9696 "| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |\n",
9797 "| | | MIG M. |\n",
9898 "|=========================================+========================+======================|\n",
99- "| 0 NVIDIA A100-SXM4-80GB Off | 00000000:00:05 .0 Off | 0 |\n",
100- "| N/A 35C P0 57W / 400W | 0MiB / 81920MiB | 0% Default |\n",
99+ "| 0 NVIDIA A100-SXM4-40GB Off | 00000000:00:04 .0 Off | 0 |\n",
100+ "| N/A 29C P0 43W / 400W | 0MiB / 40960MiB | 0% Default |\n",
101101 "| | | Disabled |\n",
102102 "+-----------------------------------------+------------------------+----------------------+\n",
103103 " \n",
124124 "base_uri": "https://localhost:8080/"
125125 },
126126 "id": "Vw3admXIqMlC",
127- "outputId": "61d7c09c-6f72-41ce-edb8-9d7de99e208d "
127+ "outputId": "5420cfd5-332f-4ea4-8c31-c454940efe0e "
128128 },
129129 "outputs": [
130130 {
131- "name": "stdout",
132131 "output_type": "stream",
132+ "name": "stdout",
133133 "text": [
134- "PyTorch version: 2.8 .0+cu126\n",
135- "Torchvision version: 0.23 .0+cu126\n",
134+ "PyTorch version: 2.9 .0+cu126\n",
135+ "Torchvision version: 0.24 .0+cu126\n",
136136 "CUDA is available: True\n"
137137 ]
138138 }
176176 "execution_count": null,
177177 "id": "eM_DGtMk_gFM",
178178 "metadata": {
179- "id": "eM_DGtMk_gFM"
179+ "id": "eM_DGtMk_gFM",
180+ "outputId": "8a3874c0-be38-4440-8052-c78885641d3b",
181+ "colab": {
182+ "base_uri": "https://localhost:8080/"
183+ }
180184 },
181- "outputs": [],
185+ "outputs": [
186+ {
187+ "output_type": "stream",
188+ "name": "stdout",
189+ "text": [
190+ "\u001b[?25l \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m0.0/212.4 kB\u001b[0m \u001b[31m?\u001b[0m eta \u001b[36m-:--:--\u001b[0m\r\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m212.4/212.4 kB\u001b[0m \u001b[31m6.1 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
191+ "\u001b[?25h"
192+ ]
193+ }
194+ ],
182195 "source": [
183196 "!pip install -q supervision jupyter_bbox_widget"
184197 ]
275288 "base_uri": "https://localhost:8080/"
276289 },
277290 "id": "QynbB5AgFhML",
278- "outputId": "a59946c8-0c2a-4f4d-bf47-61a96e6f21ed "
291+ "outputId": "e139269c-6149-45e5-8bbd-f324fdbf9a11 "
279292 },
280293 "outputs": [
281294 {
282- "name": "stdout",
283295 "output_type": "stream",
296+ "name": "stdout",
284297 "text": [
285298 "HOME: /content\n"
286299 ]
332345 "execution_count": null,
333346 "id": "y51dHwSKO2CX",
334347 "metadata": {
335- "colab": {
336- "base_uri": "https://localhost:8080/"
337- },
338- "id": "y51dHwSKO2CX",
339- "outputId": "fd7ae78f-a881-41de-d193-3bab847f2cb8"
348+ "id": "y51dHwSKO2CX"
340349 },
341- "outputs": [
342- {
343- "name": "stderr",
344- "output_type": "stream",
345- "text": [
346- "\u001b[0m\u001b[32mINFO 2025-11-20 14:02:37,450 35343 sam3_video_predictor.py: 299:\u001b[0m using the following GPU IDs: [0]\n",
347- "\u001b[0m\u001b[32mINFO 2025-11-20 14:02:37,564 35343 sam3_video_predictor.py: 315:\u001b[0m \n",
348- "\n",
349- "\n",
350- "\t*** START loading model on all ranks ***\n",
351- "\n",
352- "\n",
353- "\u001b[0m\u001b[32mINFO 2025-11-20 14:02:37,564 35343 sam3_video_predictor.py: 317:\u001b[0m loading model on rank=0 with world_size=1 -- this could take a while ...\n",
354- "\u001b[0m\u001b[32mINFO 2025-11-20 14:02:46,745 35343 sam3_video_base.py: 124:\u001b[0m setting max_num_objects=10000 and num_obj_for_compile=16\n",
355- "\u001b[0m\u001b[32mINFO 2025-11-20 14:02:51,644 35343 sam3_video_predictor.py: 319:\u001b[0m loading model on rank=0 with world_size=1 -- DONE locally\n",
356- "\u001b[0m\u001b[32mINFO 2025-11-20 14:02:51,645 35343 sam3_video_predictor.py: 330:\u001b[0m \n",
357- "\n",
358- "\n",
359- "\t*** DONE loading model on all ranks ***\n",
360- "\n",
361- "\n"
362- ]
363- }
364- ],
350+ "outputs": [],
365351 "source": [
366- "predictor = build_sam3_video_predictor(gpus_to_use=DEVICES)"
352+ "predictor = build_sam3_video_predictor(bpe_path=\"/content/sam3/sam3/assets/bpe_simple_vocab_16e6.txt.gz\", gpus_to_use=DEVICES)"
367353 ]
368354 },
369355 {
16311617 },
16321618 "nbformat": 4,
16331619 "nbformat_minor": 5
1634- }
1620+ }
0 commit comments