Hi, thanks for your inspiring work! I have a question regarding the mismatch between the 3d affordance annotation and the actual affordance depicted in the 2d interaction image. As introduced, a 2d interaction image and a 3d object is paired up in training set, the problem is: when 2d interaction images of the same affordance type (i.e. grasp) actually have two different contact regions as shown below, how can we make sure that they have the correct 3D affordance ground-truth of shape (2048,) given that the 3D affordace gts are collected from 3D affordance Net?


Hi, thanks for your inspiring work! I have a question regarding the mismatch between the 3d affordance annotation and the actual affordance depicted in the 2d interaction image. As introduced, a 2d interaction image and a 3d object is paired up in training set, the problem is: when 2d interaction images of the same affordance type (i.e. grasp) actually have two different contact regions as shown below, how can we make sure that they have the correct 3D affordance ground-truth of shape (2048,) given that the 3D affordace gts are collected from 3D affordance Net?

