Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Clarification on Global Condition vs. Local Condition in Diffusion Policy #94

Open
Gear420 opened this issue Jun 25, 2024 · 2 comments
Open

Comments

@Gear420
Copy link

Gear420 commented Jun 25, 2024

I am currently exploring the diffusion_policy library and I am interested in understanding the distinctions and applications of global condition and local condition within this framework.

What constitutes a global condition versus a local condition in the context of trajectory prediction and decision-making tasks?
How are these conditions processed and utilized differently in the library's architecture?
Are there specific examples or use cases in the codebase where global condition and local condition are distinctly applied?

@RochMollero
Copy link

Same here: how is it related to 'impating' which is the term used in the paper, which is not defined. in the code there seem 3 types of inputs:

input
local_cond
global_cond

i kinda think i nderstand what they are but some confirmation would be great

@a-nooj
Copy link

a-nooj commented Jul 29, 2024

Here's my understanding -

Inpainting = hardcoding parts of the action to whatever your want. Eg: overwriting the start and goal states after each diffusion step to match your planning problem

global_cond = adding supplementary information that applies to the action trajectory as a whole. Eg: RGB image that gives an overview of the scene or language input that tells the robot what to do generally ("pick up the red cup")
(This is simply appended to the diffusion timestep embedding to generate a global feature.)

local_cond = adding "supplementary information" that is specific to each timestep in the action. Eg: collision cost for a single robot waypoint in the whole trajectory
(This is passed through a mini neural net to generate an embedding. This embedding is concatenated with the global feature and added to the interim outputs of the downsampling and upsampling modules.)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants