WisPaper
WisPaper
学术搜索
学术问答
价格
TrueCite
[CVPR 2025] TAG: Solving the "Near-Miss" Problem in Robotic Manipulation via Target-Agnostic Guidance
总结
问题
方法
结果
要点
摘要

The paper introduces TAG (Target-Agnostic Guidance), a novel inference-time guidance mechanism designed to improve the robustness of Vision-Language-Action (VLA) models. By contrasting standard observations with "target-erased" counterfactuals, TAG effectively filters out distractor-induced biases, achieving state-of-the-art performance on benchmarks like LIBERO and VLABench (e.g., +26% SR on complex manipulation tasks).

Executive Summary

TL;DR: Even the most advanced Vision-Language-Action (VLA) models, such as π0, often fail at the finishing line—grasping a distractor instead of the target or landing just a few millimeters off. This paper identifies instance-level grounding as a core bottleneck and proposes Target-Agnostic Guidance (TAG). By contrasting real observations with "object-erased" counterfactuals, TAG provides a residual steering signal that effectively amplifies target evidence and filters out background clutter.

Academic Positioning: This work builds on the powerful foundations of Flow-Matching VLA models like π0/π0.5. It shifts the focus from better action tokenization or transformer scaling to specialized inference-time guidance to tackle the "last-mile" precision problem.

Problem & Motivation

The "Plausible Grasp" Paradox

Modern VLA policies (RT-2, OpenVLA, π0) are excellent at motion planning. They can reach the table, avoid collisions, and orient their grippers. However, in cluttered scenes—like a workspace filled with multiple poker cards—they often produce a perfect trajectory but land on the wrong card.

Visualizing Attention Distraction

The authors identify that these failures stem from over-reliance on background appearance statistics. The model learns to act based on global context (the scene) but fails to isolate the target-specific "grounding" signal when multiple similar objects (distractors) are present.

Methodology: Target-Agnostic Guidance (TAG)

The Vision-Centric CFG

Inspired by Classifier-Free Guidance (CFG) in diffusion models, TAG proposes that a policy prediction should be steered by the difference between a target-present observation and a target-agnostic one.

Mathematically, given a velocity field $v_ heta$, TAG computes: $$v_{ ext{TAG}} = v_ heta(x_ au, I_{ ext{uncond}}) + w \cdot (v_ heta(x_ au, I_{ ext{cond}}) - v_ heta(x_ au, I_{ ext{uncond}}))$$

Where $I_{ ext{uncond}}$ is an "erased" version of the image. This subtraction effectively cancels out environmental noise and prior biases, leaving only the action component driven by the specific task target.

The Counterfactual Synthesis Pipeline

To make this work, the authors created an automated "data factory":

  1. Target Parsing: Qwen3-VL extracts a visual description of the target from the instruction.
  2. Detection & Tracking: Grounding DINO and SAM 2 localize the target across video frames.
  3. Inpainting: MiniMaxRemover "removes" the target and fills in the background, creating a counterfactual sequence.

Automated Synthesis Pipeline

Experiments & Results

LIBERO and VLABench Breakthroughs

The effectiveness of TAG is most evident in highly cluttered environments.

  • In LIBERO-Long, which requires long-horizon planning prone to compounding errors, TAG increased the π0.5 success rate from 89.6% to 97.0%.
  • In VLABench Track 1 (e.g., Selecting aSpecific Poker card), the success rate more than doubled (29.4% to 55.4%).

Performance Comparison Table

Ablation Study: Why Static Backgrounds Matter?

One key finding was that static backgrounds ($I_{bg}$) performed far better than real-time dynamic erasure. Frame-wise masking during inference introduced temporal artifacts that confused the spatial reasoning of the model. Using a clean, pre-intervention background provided a much more stable "anchor" for the subtraction process.

Critical Analysis & Conclusion

Takeaway

TAG provides a modular, "drop-in" improvement for any VLA model that uses iterative refinement (like Diffusion or Flow-Matching). It proves that explicit visual disentanglement at inference time is a powerful tool for achieving high-precision grounding without needing massive new architectural changes.

Limitations & Future Work

  1. Computation Overhead: Running the model twice (conditional and unconditional) at each inference step increases latency—though the authors imply this is manageable.
  2. Imperfect Erasure: If the inpainting model fails or removes too much (like the robotic arm itself), the guidance signal might become misleading.

In conclusion, TAG is a significant step toward making robots more reliable in messy, real-world environments. By simply training the model to know what the world looks like without the target, we empower it to act much more decisively with the target.

发现相似论文

试试这些示例

  • Search for recent papers on Classifier-Free Guidance applications for non-generative robotic trajectory optimization or VLA models.
  • Which study first introduced the LIBERO-Plus benchmark for robustness analysis, and how does TAG's performance compare to other top-ranking models on its "Noise" and "Camera" perturbation categories?
  • Explore research that uses Segment Anything Model (SAM 2) or Grounding DINO to create counterfactual training data for embodied AI reinforcement learning.
目录
[CVPR 2025] TAG: Solving the "Near-Miss" Problem in Robotic Manipulation via Target-Agnostic Guidance
1. Executive Summary
2. Problem & Motivation
2.1. The "Plausible Grasp" Paradox
3. Methodology: Target-Agnostic Guidance (TAG)
3.1. The Vision-Centric CFG
3.2. The Counterfactual Synthesis Pipeline
4. Experiments & Results
4.1. LIBERO and VLABench Breakthroughs
4.2. Ablation Study: Why Static Backgrounds Matter?
5. Critical Analysis & Conclusion
5.1. Takeaway
5.2. Limitations & Future Work