Hi Anders,
Thank you very much for your reply. I have tried following your pipeline and looking at your code snippets. Unfortunately I am unable to obtain useful results. Based on manual measurements I expect a translation of [0, 70, -140] (mm), but instead I get the following:
t_cam2gripper_TSAI =
[44.19581170751601;
-36.76606175788393;
334.8397719246281]
t_cam2gripper_HORAUD =
[72.54143289335866;
-8.567504303752303;
172.897423299859]
t_cam2gripper_PARK =
[28.76007401985598;
7.45904983438578;
437.0946431017954]
t_cam2gripper_ANDREFF =
[77.60418709348254;
-22.88842792831192;
353.6155533951903]
t_cam2gripper_DAN =
[-nan(ind);
-nan(ind);
-nan(ind)]
I have ensured that the images are loaded in the correct order. I just don’t know if it is my code being wrong or if it is the robot pose I’m reading wrong from the UR5.
I have uploaded my material here: Source Material
Any help would be highly appreciated!