WisPaper
WisPaper
学术搜索
学术问答
价格
TrueCite
[CVPR 2026] Group3D: Preventing 3D "Over-merging" with MLLM-Driven Semantic Gating
总结
问题
方法
结果
要点
摘要

This paper introduces Group3D, a training-free multi-view open-vocabulary 3D object detection framework. It leverages Multimodal Large Language Models (MLLMs) to perform semantic grouping of 2D fragments, achieving state-of-the-art zero-shot performance on ScanNet and ARKitScenes without requiring 3D supervision or ground-truth depth.

TL;DR

The inability to distinguish between a "table" and the "floor" using only incomplete 2D images is a classic failure mode in multi-view 3D detection. Group3D solves this by using an MLLM (like GPT-5) to act as a "semantic gatekeeper." By grouping synonymous terms and separating structurally distinct ones before merging 3D fragments, it achieves a massive +20% mAP boost over previous benchmarks like Zoo3D in zero-shot settings.

Problem & Motivation: The Geometry-Only Trap

Existing multi-view open-vocabulary detectors usually follow a "Geometry First, Semantics Later" pipeline. They lift 2D segments into 3D space and merge them based on spatial overlap.

However, in pure RGB settings (without LiDAR), depth estimation is noisy. If a chair is right next to a desk, their fragments often overlap geometrically. A purely geometric algorithm sees one big "blob" and merges them. Once merged, it’s nearly impossible for a downstream classifier to "un-stick" them. The authors recognize that semantic knowledge (knowing that a chair is not a desk) should be a prerequisite for merging, not an afterthought.

Methodology: The Architecture of Group3D

Group3D operates through a sophisticated memory-based pipeline that bridges 2D observations and 3D reasoning.

1. Scene Memory Construction

  • Scene Vocabulary Memory: The model queries an MLLM (GPT-5.1) across views to build a list of objects present.
  • 3D Fragment Memory: 2D masks from SAM are back-projected into 3D space using depth and pose estimation.

2. Semantic Compatibility Grouping

This is the core innovation. The MLLM doesn't just list objects; it partitions the vocabulary into Compatibility Groups.

  • Compatible: {chair, sofa, couch}. These represent the same physical object under different labels.
  • Incompatible: {wall, door}. These might be close together but are physically distinct.

Group3D Overall Pipeline Figure 1: The Group3D framework integrates MLLM-driven semantic groups to gate the 3D fragment merging process.

3. Group-Gated Merging

During the merging phase, two fragments $F_i$ and $F_j$ are merged ONLY IF:

  1. They satisfy the Semantic Condition: $g(\ell_i) = g(\ell_j)$ (they belong to the same group).
  2. They satisfy the Geometric Condition: Their voxel-level intersection-over-union (IoU) or containment ratio exceeds a threshold.

This "Semantic Gate" prevents the irreversible fusion of semantically distinct but spatially adjacent objects.

Experiments & Results

Group3D tested on ScanNet and ARKitScenes, showcasing its dominance in the multi-view RGB domain.

  • ScanNet20: Achieved 51.1% mAP25, a staggering leap from the 19.8% of OpenM3D.
  • Pose-Free Robustness: Even when the camera's position is unknown (relying on SfM-style reconstruction), Group3D hits 41.2% mAP25, proving that semantic gating can "fix" noisy geometry.

Performance Comparison Table Table 1: Quantitative comparison showing Group3D's significant lead in both pose-known and pose-free settings.

Ablation Insight: Why Grouping Matters?

The ablation study (Table 5) reveals that merging by "Same Category" is too strict (missing matches due to label noise), whereas "No Category" (Geometry Only) is too loose. Semantic Compatibility Grouping strikes the perfect balance by allowing lexical variation (e.g., merging "trash bin" with "waste basket") while maintaining physical boundaries.

Critical Analysis & Conclusion

Takeaways

Group3D proves that MLLMs are no longer just for "labeling" results—they can be used as structural logic inside the perception pipeline. By using language to resolve geometric ambiguity, the model handles long-tail categories and noisy depth maps much better than traditional algorithms.

Limitations

  1. Computational Overhead: Relying on per-frame MLLM queries can be slow for real-time applications.
  2. Dependency on Foundation Models: The performance is bounded by the quality of the 2D segmentor (SAM) and the reconstruction model.

Future Work

The next frontier is likely moving from static semantic grouping to dynamic scene reasoning, where the MLLM understands the relationships between objects (e.g., "the cup on the table") to further refine 3D instance boundaries.


For more details, visit the project page: https://ubin108.github.io/Group3D/

发现相似论文

试试这些示例

  • Search for recent papers in open-vocabulary 3D object detection that utilize MLLMs for cross-view temporal association or instance merging.
  • Which study first introduced the concept of using CLIP-based features for 3D proposal grounding, and how does Group3D's MLLM-driven grouping technically differ from late-stage CLIP alignment?
  • Examine how semantic compatibility grouping techniques have been applied to robotics or autonomous driving tasks where 3D sensor data is sparse or noisy.
目录
[CVPR 2026] Group3D: Preventing 3D "Over-merging" with MLLM-Driven Semantic Gating
1. TL;DR
2. Problem & Motivation: The Geometry-Only Trap
3. Methodology: The Architecture of Group3D
3.1. 1. Scene Memory Construction
3.2. 2. Semantic Compatibility Grouping
3.3. 3. Group-Gated Merging
4. Experiments & Results
4.1. Ablation Insight: Why Grouping Matters?
5. Critical Analysis & Conclusion
5.1. Takeaways
5.2. Limitations
5.3. Future Work