I’m making a camera calibration toolkit with OpenCV and C++. I’m “using calibrateCamera” and “undistort” functions from opencv2/calib3d. I get good results for two image set but I get terrible result for 5 image set. All image sets are good and has enough picture to calibrate camera.
For examle when I try my code with that image set I can find and draw corners of chessboard correctly but the other stuff are wrong. Distortion coefs are not true, the optical center of camera is negative so undistortion is weird.
What do you think about that?