<rss xmlns:atom="http://www.w3.org/2005/Atom" version="2.0">
    <channel>
        <title>即插即用型 - 分类 - Zhaoylee&#39;s Blogs</title>
        <link>https://zhaoylee.github.io/Blogs_lovelt/categories/%E5%8D%B3%E6%8F%92%E5%8D%B3%E7%94%A8%E5%9E%8B/</link>
        <description>即插即用型 - 分类 - Zhaoylee&#39;s Blogs</description>
        <generator>Hugo -- gohugo.io</generator><language>zh-CN</language><lastBuildDate>Sat, 04 Apr 2026 12:31:36 &#43;0800</lastBuildDate><atom:link href="https://zhaoylee.github.io/Blogs_lovelt/categories/%E5%8D%B3%E6%8F%92%E5%8D%B3%E7%94%A8%E5%9E%8B/" rel="self" type="application/rss+xml" /><item>
    <title>MONOLSS  Learnable Sample Selection For Monocular 3D Detection</title>
    <link>https://zhaoylee.github.io/Blogs_lovelt/posts/monolss--learnable-sample-selection-for-monocular-3d-detection/</link>
    <pubDate>Sat, 04 Apr 2026 12:31:36 &#43;0800</pubDate>
    <author>zhaoylee</author>
    <guid>https://zhaoylee.github.io/Blogs_lovelt/posts/monolss--learnable-sample-selection-for-monocular-3d-detection/</guid>
    <description><![CDATA[<div class="featured-image">
                <img src="https://cdn.jsdelivr.net/gh/zhaoylee/BlogImage@main/blogs/20260330230353149.png" referrerpolicy="no-referrer">
            </div>博客的简述]]></description>
</item>
<item>
    <title>Mix-Teaching: A Simple, Unified and Effective  Semi-Supervised Learning Framework for  Monocular 3D Object Detection</title>
    <link>https://zhaoylee.github.io/Blogs_lovelt/posts/mix-teaching--a-simple-unified-and-effective--semi-supervised-learning-framework-for--monocular-3d-object-detection/</link>
    <pubDate>Mon, 30 Mar 2026 23:17:02 &#43;0800</pubDate>
    <author>zhaoylee</author>
    <guid>https://zhaoylee.github.io/Blogs_lovelt/posts/mix-teaching--a-simple-unified-and-effective--semi-supervised-learning-framework-for--monocular-3d-object-detection/</guid>
    <description><![CDATA[<div class="featured-image">
                <img src="https://cdn.jsdelivr.net/gh/zhaoylee/BlogImage@main/blogs/20260330231443214.png" referrerpolicy="no-referrer">
            </div>本文提出 Mix-Teaching，首个专为单目 3D 目标检测设计的半监督学习统一框架。针对伪标签“低精度与低召回率”导致的确认偏差痛点，创新性提出“分解-重组”的跨帧实例级混合机制，并结合基于不确定性的过滤策略，优雅且高效地释放了无标注数据的潜力。]]></description>
</item>
<item>
    <title>Adaptive Dual Uncertainty Optimization: Boosting Monocular 3D Object Detection under Test Time Shifts</title>
    <link>https://zhaoylee.github.io/Blogs_lovelt/posts/adaptive-dual-uncertainty-optimization---boosting-monocular-3d-object-detection-under-test-time-shifts/</link>
    <pubDate>Mon, 30 Mar 2026 10:44:00 &#43;0800</pubDate>
    <author>zhaoylee</author>
    <guid>https://zhaoylee.github.io/Blogs_lovelt/posts/adaptive-dual-uncertainty-optimization---boosting-monocular-3d-object-detection-under-test-time-shifts/</guid>
    <description><![CDATA[<div class="featured-image">
                <img src="https://cdn.jsdelivr.net/gh/zhaoylee/BlogImage@main//blogs/20260330113435293.png" referrerpolicy="no-referrer">
            </div>针对 M3OD 在未知域“测试时偏移”导致的性能断崖，本文提出双重不确定性优化 (DUO)。核心是通过无监督 Focal Loss 压制语义模糊，并用语义感知法向量约束修复空间几何坍塌，大幅提升落地鲁棒性。]]></description>
</item>
<item>
    <title>OBMO: One Bounding Box Multiple Objects for Monocular 3D Object Detection</title>
    <link>https://zhaoylee.github.io/Blogs_lovelt/posts/obmo-one-bounding-box-multiple-objects-for-monocular-3d-object-detection/</link>
    <pubDate>Tue, 24 Mar 2026 10:48:33 &#43;0800</pubDate>
    <author>zhaoylee</author>
    <guid>https://zhaoylee.github.io/Blogs_lovelt/posts/obmo-one-bounding-box-multiple-objects-for-monocular-3d-object-detection/</guid>
    <description><![CDATA[<div class="featured-image">
                <img src="/Blogs_lovelt/cover.jpg" referrerpolicy="no-referrer">
            </div>简述:一个框对应多个物体位置，通过施加软标签，帮助网络训练稳定，从而提升一定的性能;]]></description>
</item>
<item>
    <title>OCM3D: Object-Centric Monocular 3D Object Detection</title>
    <link>https://zhaoylee.github.io/Blogs_lovelt/posts/ocm3d--object-centric-monocular-3d-object-detection/</link>
    <pubDate>Mon, 16 Mar 2026 09:12:18 &#43;0800</pubDate>
    <author>zhaoylee</author>
    <guid>https://zhaoylee.github.io/Blogs_lovelt/posts/ocm3d--object-centric-monocular-3d-object-detection/</guid>
    <description><![CDATA[<div class="featured-image">
                <img src="/Blogs_lovelt/cover.jpg" referrerpolicy="no-referrer">
            </div><hr>
<blockquote>
<p><strong>🏛️ 会议/期刊</strong>：arxiv<br>
<strong>📅 发表年份</strong>：2021<br>
<strong>💻 开源代码</strong>：<a href="https://github.com/mrsempress/OBMO_GUPNet/blob/main/tools/offline_OBMO.py" target="_blank" rel="noopener noreffer ">OBMO_GUPNet</a><br>
<strong>📄 论文题目</strong>：<a href="https://arxiv.org/pdf/2104.06041" target="_blank" rel="noopener noreffer ">OCM3D: Object-Centric Monocular 3D Object Detection</a></p>
</blockquote>
<hr>
<h3 id="1-文献背景研究目的与核心问题">1. 文献背景、研究目的与核心问题</h3>
<ul>
<li>
<p><strong>研究背景</strong>：单目 3D 目标检测（Monocular 3D Object Detection）是一个高度病态（ill-posed）的问题。主流方法通常依赖纯图像或将其转化为伪激光雷达（Pseudo-LiDAR）点云。然而，前者难以捕捉像素间的 3D 空间几何关系，后者则受困于单目深度估计带来的巨大点云噪声。</p>]]></description>
</item>
<item>
    <title>LR3D: Improving Distant 3D Object Detection Using 2D Box Supervision</title>
    <link>https://zhaoylee.github.io/Blogs_lovelt/posts/lr3d--improving-distant-3d-object-detection-using-2d-box-supervision/</link>
    <pubDate>Sun, 15 Mar 2026 22:23:00 &#43;0800</pubDate>
    <author>zhaoylee</author>
    <guid>https://zhaoylee.github.io/Blogs_lovelt/posts/lr3d--improving-distant-3d-object-detection-using-2d-box-supervision/</guid>
    <description><![CDATA[<div class="featured-image">
                <img src="/Blogs_lovelt/cover.jpg" referrerpolicy="no-referrer">
            </div><hr>
<blockquote>
<p><strong>🏛️ 会议/期刊</strong>：CVPR<br>
<strong>📅 发表年份</strong>：2024<br>
<strong>💻 开源代码</strong>：<a href="%e5%a1%ab%e5%86%99%e4%bd%a0%e7%9a%84URL" rel="">无</a><br>
<strong>📄 论文题目</strong>：<a href="https://openaccess.thecvf.com/content/CVPR2024/papers/Yang_Improving_Distant_3D_Object_Detection_Using_2D_Box_Supervision_CVPR_2024_paper.pdf" target="_blank" rel="noopener noreffer ">Improving Distant 3D Object Detection Using 2D Box Supervision</a></p>
</blockquote>
<hr>
<p>这篇由 NVIDIA 等机构的研究人员发表在 CVPR 2024 的重磅论文 <strong>《Improving Distant 3D Object Detection Using 2D Box Supervision》(简称 LR3D)</strong>，切入了一个目前高阶自动驾驶极其头疼的落地难题：<strong>远距离感知（Long-Range Detection）</strong>。它展示了如何用最廉价的标注，榨取单目视觉在远距离上的极限潜力。</p>]]></description>
</item>
<item>
    <title>OBMO: One Bounding Box Multiple Objects
for Monocular 3D Object Detection</title>
    <link>https://zhaoylee.github.io/Blogs_lovelt/posts/obmo--one-bounding-box-multiple-objects-for-monocular-3d-object-detection/</link>
    <pubDate>Sun, 15 Mar 2026 21:59:12 &#43;0800</pubDate>
    <author>zhaoylee</author>
    <guid>https://zhaoylee.github.io/Blogs_lovelt/posts/obmo--one-bounding-box-multiple-objects-for-monocular-3d-object-detection/</guid>
    <description><![CDATA[<div class="featured-image">
                <img src="/Blogs_lovelt/cover.jpg" referrerpolicy="no-referrer">
            </div><hr>
<blockquote>
<p><strong>🏛️ 会议/期刊</strong>：IEEE TIP<br>
<strong>📅 发表年份</strong>：2023<br>
<strong>💻 开源代码</strong>：<a href="https://github.com/mrsempress/OBMO_patchnet" target="_blank" rel="noopener noreffer ">mrsempress/OBMO_patchnet</a><br>
<strong>📄 论文题目</strong>：<a href="https://arxiv.org/pdf/2212.10049" target="_blank" rel="noopener noreffer ">OBMO: One Bounding Box Multiple Objects for Monocular 3D Object Detection</a></p>
</blockquote>
<hr>
<p>这篇发表于 IEEE TIP (2023) 的经典论文 <strong>《OBMO: One Bounding Box Multiple Objects for Monocular 3D Object Detection》</strong> 切入点非常犀利。它没有在复杂的网络主干上做文章，而是直击单目 3D 目标检测在“底层数学物理逻辑”上的痛点，提出了一种极其优雅的“即插即用（Plug-and-play）”训练策略。</p>]]></description>
</item>
<item>
    <title>MonoXiver： Monocular 3D Object Detection with Bounding Box Denoising in 3D by Perceiver</title>
    <link>https://zhaoylee.github.io/Blogs_lovelt/posts/monoxiver--monocular-3d-object-detection-with-bounding-box-denoising-in-3d-by-perceivermonocular-3d-object-detection-with-bounding-box-denoising-in-3d-by-perceiver/</link>
    <pubDate>Sun, 15 Mar 2026 21:14:36 &#43;0800</pubDate>
    <author>zhaoylee</author>
    <guid>https://zhaoylee.github.io/Blogs_lovelt/posts/monoxiver--monocular-3d-object-detection-with-bounding-box-denoising-in-3d-by-perceivermonocular-3d-object-detection-with-bounding-box-denoising-in-3d-by-perceiver/</guid>
    <description><![CDATA[<div class="featured-image">
                <img src="/Blogs_lovelt/cover.jpg" referrerpolicy="no-referrer">
            </div><hr>
<blockquote>
<p><strong>🏛️ 会议/期刊</strong>：ICCV<br>
<strong>📅 发表年份</strong>：2023<br>
<strong>💻 开源代码</strong>：<a href="https://github.com/Xianpeng919/monoxiver" target="_blank" rel="noopener noreffer ">Xianpeng919/monoxiver (ICCV'23)</a><br>
<strong>📄 论文题目</strong>：<a href="https://openaccess.thecvf.com/content/ICCV2023/papers/Liu_Monocular_3D_Object_Detection_with_Bounding_Box_Denoising_in_3D_ICCV_2023_paper.pdf" target="_blank" rel="noopener noreffer ">Monocular 3D Object Detection with Bounding Box Denoising in 3D by Perceiver</a></p>
</blockquote>
<hr>
<p>这篇发表于 ICCV 2023 的论文 <strong>《Monocular 3D Object Detection with Bounding Box Denoising in 3D by Perceiver》(简称 MonoXiver)</strong>，提供了一个非常经典且极具工程价值的“自上而下（Top-down）”纠错思路。它并没有试图发明一种全新的主干网络，而是设计了一个强大的“插件”，专门用来拯救那些定位不准的预测框。</p>]]></description>
</item>
<item>
    <title>PLOT: Pseudo-Labeling via Object Tracking for Monocular 3D Object Detection</title>
    <link>https://zhaoylee.github.io/Blogs_lovelt/posts/plot---pseudo-labeling-via-object-tracking-for-monocular-3d-object-detection/</link>
    <pubDate>Sun, 15 Mar 2026 20:52:51 &#43;0800</pubDate>
    <author>zhaoylee</author>
    <guid>https://zhaoylee.github.io/Blogs_lovelt/posts/plot---pseudo-labeling-via-object-tracking-for-monocular-3d-object-detection/</guid>
    <description><![CDATA[<div class="featured-image">
                <img src="/Blogs_lovelt/cover.jpg" referrerpolicy="no-referrer">
            </div><hr>
<blockquote>
<p><strong>🏛️ 会议/期刊</strong>：ICLR<br>
<strong>📅 发表年份</strong>：2026<br>
<strong>💻 开源代码</strong>：<a href="%e5%a1%ab%e5%86%99%e4%bd%a0%e7%9a%84URL" rel="">无</a><br>
<strong>📄 论文题目</strong>：<a href="https://openreview.net/pdf?id=3knS4J9isg" target="_blank" rel="noopener noreffer ">PLOT: Pseudo-Labeling via Object Tracking for Monocular 3D Object Detection</a></p>
</blockquote>
<hr>
<h3 id="一-背景研究目的与核心问题">一、 背景、研究目的与核心问题</h3>
<ul>
<li>
<p><strong>研究背景：</strong> 单目 3D 目标检测模型极度“吃数据”。然而，人工标注 3D 边界框极其昂贵且耗时，导致目前带 3D 标签的数据集规模很小，严重限制了模型的泛化能力。</p>]]></description>
</item>
<item>
    <title>Object-Scene-Camera Decomposition and Recomposition for Data-Efficient Monocular 3D Object Detection</title>
    <link>https://zhaoylee.github.io/Blogs_lovelt/posts/dr-traversal-m3d--object-scene-camera-decomposition-and-recomposition-for-data-efficient-monocular-3d-object-detection/</link>
    <pubDate>Sun, 15 Mar 2026 19:59:15 &#43;0800</pubDate>
    <author>zhaoylee</author>
    <guid>https://zhaoylee.github.io/Blogs_lovelt/posts/dr-traversal-m3d--object-scene-camera-decomposition-and-recomposition-for-data-efficient-monocular-3d-object-detection/</guid>
    <description><![CDATA[<div class="featured-image">
                <img src="/Blogs_lovelt/cover.jpg" referrerpolicy="no-referrer">
            </div><hr>
<blockquote>
<p><strong>🏛️ 会议/期刊</strong>：IJCV<br>
<strong>📅 发表年份</strong>：2026<br>
<strong>💻 开源代码</strong>：<a href="https://github.com/kwong292521/DR-Traversal-M3D" target="_blank" rel="noopener noreffer ">GitHub 链接</a><br>
<strong>📄 论文题目</strong>：<a href="https://arxiv.org/pdf/2602.20627" target="_blank" rel="noopener noreffer ">Object-Scene-Camera Decomposition and Recomposition for Data-Efficient Monocular 3D Object Detection</a></p>
</blockquote>
<hr>
<h3 id="一-背景研究目的与核心问题">一、 背景、研究目的与核心问题</h3>
<ul>
<li>
<p><strong>研究背景：</strong> 单目 3D 目标检测本身是一个病态（ill-posed）问题，由于缺乏深度信息，模型极度依赖海量、多样化的训练数据来学习鲁棒的特征。</p>]]></description>
</item>
</channel>
</rss>
