Skip to content

Commit c16db5c

Browse files
authored
Update index.html
1 parent c5b8367 commit c16db5c

File tree

1 file changed

+57
-32
lines changed

1 file changed

+57
-32
lines changed

index.html

Lines changed: 57 additions & 32 deletions
Original file line numberDiff line numberDiff line change
@@ -158,6 +158,24 @@
158158
font-family: 'Courier New', monospace;
159159
}
160160

161+
.paper-preview {
162+
width: 100%;
163+
border-radius: 8px;
164+
border: 1px solid #ddd;
165+
background-color: #fff;
166+
}
167+
.zoomable {
168+
width: 100%;
169+
border-radius: 8px;
170+
border: 1px solid #ddd;
171+
cursor: zoom-in;
172+
transition: transform 0.3s ease, border-color 0.3s;
173+
}
174+
.zoomable:hover {
175+
border-color: #0077cc;
176+
transform: scale(1.01);
177+
}
178+
161179
.zoom-overlay {
162180
position: fixed;
163181
top: 0;
@@ -202,6 +220,10 @@
202220
background-color: #444;
203221
color: #fff;
204222
}
223+
/* Ensure paper preview SVG stays visible in dark mode */
224+
.dark-mode .paper-preview {
225+
background-color: #fff;
226+
}
205227

206228
#theme-toggle {
207229
position: fixed;
@@ -257,20 +279,18 @@ <h2>CVPR 2025</h2>
257279

258280
<h2>Abstract</h2>
259281
<p>
260-
Can objects that are not visible in an image—but are in the
261-
vicinity of the camera—be detected? This study introduces
262-
the novel tasks of 2D, 2.5D and 3D unobserved object de-
263-
tection for predicting the location of nearby objects that are
264-
occluded or lie outside the image frame. We adapt several
265-
state-of-the-art pre-trained generative models to address
266-
this task, including 2D and 3D diffusion models and vision–
267-
language models, and show that they can be used to infer the
268-
presence of objects that are not directly observed. To bench-
269-
mark this task, we propose a suite of metrics that capture
270-
different aspects of performance. Our empirical evaluation
271-
on indoor scenes from the RealEstate10k and NYU Depth
272-
V2 datasets demonstrate results that motivate the use of
273-
generative models for the unobserved object detection task.
282+
Can objects that are not visible in an image—but are in the vicinity of the
283+
camera—be detected? This study introduces the novel tasks of 2D, 2.5D and 3D
284+
unobserved object detection for predicting the location of nearby objects
285+
that are occluded or lie outside the image frame. We adapt several
286+
state-of-the-art pre-trained generative models to address this task,
287+
including 2D and 3D diffusion models and vision–language models, and show
288+
that they can be used to infer the presence of objects that are not
289+
directly observed. To benchmark this task, we propose a suite of metrics
290+
that capture different aspects of performance. Our empirical evaluation on
291+
indoor scenes from the RealEstate10k and NYU Depth V2 datasets demonstrate
292+
results that motivate the use of generative models for the unobserved
293+
object detection task.
274294
</p>
275295

276296
<h2>Task Definition</h2>
@@ -284,11 +304,15 @@ <h2>Task Definition</h2>
284304
/>
285305
</div>
286306
<div class="task-text">
287-
<p>
288-
The task of <strong>unobserved object detection</strong> is to detect objects that are present in the scene but not captured within the camera frustum. In this paper, we address this by predicting a conditional distribution over a bounded spatial region and a set of semantic labels from a single RGB image. We refer to this distribution as a <strong>spatio-semantic distribution</strong> visualized as a heatmap.
289-
</p>
290-
291-
307+
<p>
308+
The task of <strong>unobserved object detection</strong> is to detect
309+
objects that are present in the scene but not captured within the
310+
camera frustum. In this paper, we address this by predicting a
311+
conditional distribution over a bounded spatial region and a set of
312+
semantic labels from a single RGB image. We refer to this distribution
313+
as a <strong>spatio-semantic distribution</strong>, visualized as a
314+
heatmap.
315+
</p>
292316
</div>
293317
</div>
294318

@@ -298,7 +322,7 @@ <h2>Paper</h2>
298322
<img
299323
src="assets/images/first.svg"
300324
alt="First Page of the Paper"
301-
style="width: 100%; border-radius: 8px; border: 1px solid #ddd;"
325+
class="paper-preview"
302326
loading="lazy"
303327
/>
304328
</a>
@@ -326,12 +350,12 @@ <h2>Results</h2>
326350
alt="Detection Results"
327351
class="zoomable"
328352
loading="lazy"
329-
style="width: 100%; border-radius: 8px; border: 1px solid #ddd; cursor: zoom-in; transition: transform 0.3s ease, border-color 0.3s;"
330353
/>
331354
<figcaption>
332-
<strong>Figure:</strong> Each row shows spatial predictions by different models across object
333-
types (TV, fridge, sink, laptop). The white star represents ground truth.
334-
Warmer heatmap indicates higher object likelihood.
355+
<strong>Figure:</strong> Each row shows spatial predictions by different
356+
models across object types (TV, fridge, sink, laptop). The white star
357+
represents ground truth. Warmer heatmap indicates higher object
358+
likelihood.
335359
</figcaption>
336360
</figure>
337361

@@ -343,18 +367,19 @@ <h2>Poster</h2>
343367
alt="Project Poster"
344368
class="zoomable"
345369
loading="lazy"
346-
style="width: 100%; border-radius: 8px; border: 1px solid #ddd; cursor: zoom-in; transition: transform 0.3s ease, border-color 0.3s;"
347370
/>
348371
<figcaption>Click to zoom the full poster.</figcaption>
349372
</figure>
350373

351374
<h2>Acknowledgments</h2>
352375
<p>
353-
Subhransu is supported by the University Research Scholarship at the Australian National University.
354-
This research was partially funded by the U.S. Government under DARPA TIAMAT HR00112490421. The
355-
views and conclusions expressed in this document are solely those of the authors and do not represent
356-
the official policies or endorsements, either expressed or implied, of the U.S. Government. This
357-
research was also funded by the Australian Research Council under the scheme ITRH IH210100030.
376+
Subhransu is supported by the University Research Scholarship at the
377+
Australian National University. This research was partially funded by the
378+
U.S. Government under DARPA TIAMAT HR00112490421. The views and conclusions
379+
expressed in this document are solely those of the authors and do not
380+
represent the official policies or endorsements, either expressed or
381+
implied, of the U.S. Government. This research was also funded by the
382+
Australian Research Council under the scheme ITRH IH210100030.
358383
</p>
359384

360385
<h2>Cite As</h2>
@@ -373,7 +398,7 @@ <h2>Cite As</h2>
373398
</div>
374399

375400
<footer>
376-
Contact: Corresponding Author —
401+
Contact: Corresponding Author —
377402
<a href="mailto:Subhransu.Bhattacharjee@anu.edu.au">Subhransu.Bhattacharjee@anu.edu.au</a>
378403
</footer>
379404
</div>
@@ -398,7 +423,7 @@ <h2>Cite As</h2>
398423
});
399424
});
400425

401-
// Zoom functionality for any element with class "zoomable"
426+
// Zoom functionality for any element with class "zoomable" or task image
402427
document.querySelectorAll('.zoomable, .task-image-wrapper img').forEach((img) => {
403428
img.addEventListener('click', () => {
404429
const overlay = document.createElement('div');

0 commit comments

Comments
 (0)