2017-10-18 103 views
0

投射点我想:使用OpenCV的

  1. 做射线平面相交针对几个点是在世界坐标,
  2. 然后我得到的是相交
  3. 那么那些点我试着将这些点从该世界坐标投影到图像坐标,

但我得到的点在0.4,0.1,0.5等范围内。

下面是我在做什么,希望你发现错误

 Mat cameraIntrinsics(3, 3, CV_32F); 

     cameraIntrinsics.at<float>(0, 0) = 1.6003814935684204; 
     cameraIntrinsics.at<float>(0, 1) = 0; 
     cameraIntrinsics.at<float>(0, 2) = -0.0021958351135253906; 
     cameraIntrinsics.at<float>(1, 0) = 0; 
     cameraIntrinsics.at<float>(1, 1) = 1.6003814935684204; 
     cameraIntrinsics.at<float>(1, 2) = -0.0044271680526435375; 
     cameraIntrinsics.at<float>(2, 0) = 0; 
     cameraIntrinsics.at<float>(2, 1) = 0; 
     cameraIntrinsics.at<float>(2, 2) = 1; 


     Mat invCameraIntrinsics = cameraIntrinsics.inv(); 

     std::vector<cv::Point3f> points3D; 
     std::vector<Ray> rays; 
     for (int i = 0; i < corners.size(); i++) 
     { 
     cv::Point3f pt; 

     pt.z = -1.0f; 

     pt.x = corners[i].x; 
     pt.y = corners[i].y; 

     points3D.push_back(pt); 

     Ray ray; 

     ray.origin = Vec3f(0, 0, 0); 
     ray.direction = Vec3f(pt.x, pt.y, pt.z); 

     rays.push_back(ray); 
     } 

     std::vector<cv::Point3f> pointsTransformed3D; 


     cv::transform(points3D, pointsTransformed3D, invCameraIntrinsics); 


     std::vector<cv::Vec3f> contacts; 

     for (int i = 0; i < pointsTransformed3D.size(); i++) 
     { 
     Vec3f pt(pointsTransformed3D[i].x, pointsTransformed3D[i].y, pointsTransformed3D[i].z); 

     cv::Vec3f contact; 
     std::pair<bool, double> test = linePlaneIntersection(contact, rays[i].direction, rays[i].origin, Vec3f(0, 1, 0), pt); 
     if (test.first == true) 
     { 
      cv::Vec3f contact(rays[i].origin + (rays[i].direction) * test.second); 
      contacts.push_back(contact); 
     } 
     } 


     Mat rotationMatrix(3, 3, CV_32F); 

     rotationMatrix.at<float>(0, 0) = 0.9115078799790896; 
     rotationMatrix.at<float>(0, 1) = -0.1883612409043686; 
     rotationMatrix.at<float>(0, 2) = -0.3656137684237178; 
     rotationMatrix.at<float>(1, 0) = -0.3046835686704949; 
     rotationMatrix.at<float>(1, 1) = 0.2878667580409447; 
     rotationMatrix.at<float>(1, 2) = -0.9079100465339108; 
     rotationMatrix.at<float>(2, 0) = 0.2762631132059388; 
     rotationMatrix.at<float>(2, 1) = 0.9389636694462479; 
     rotationMatrix.at<float>(2, 2) = 0.2050022432604093; 

     cv::Mat rVec(3, 1, CV_32F); // Rotation vector 
     Rodrigues(rotationMatrix, rVec); 
     double norm = cv::norm(rVec); 

     float theta = (float)(sqrt(rVec.at<float>(0)*rVec.at<float>(0) + rVec.at<float>(1)*rVec.at<float>(1) + rVec.at<float>(2)*rVec.at<float>(2)) * 180/3.14567898726); 

     cv::Mat tVec(3, 1, CV_32F); // Translation vector 
     tVec.at<float>(0) = 21.408294677734375; 
     tVec.at<float>(1) = 531.1319580078125; 
     tVec.at<float>(2) = 705.74224853515625; 

     cv::Mat distCoeffs(5, 1, CV_32F); // Distortion vector 
     distCoeffs.at<float>(0) = 0; 
     distCoeffs.at<float>(1) = 0; 
     distCoeffs.at<float>(2) = 0; 
     distCoeffs.at<float>(3) = 0; 
     distCoeffs.at<float>(4) = 0; 

     std::vector<cv::Point2d> projectedPoints; 
     std::vector < cv::Point3d> ContactPoints; 

     for (int i = 0; i < contacts.size(); i++) 
     { 
     cv::Point3d pt; 

     pt.x = contacts[i][0]; 
     pt.y = contacts[i][1]; 
     pt.z = contacts[i][2]; 

     ContactPoints.push_back(pt); 
     } 


     cv::projectPoints(ContactPoints, rVec, tVec, cameraIntrinsics, distCoeffs, projectedPoints); 

     for (size_t i = 0; i < projectedPoints.size(); i++) 
     { 
     cv::Point2d pt; 

     pt.x = projectedPoints[i].x; 
     pt.y = projectedPoints[i].y; 

     cv::circle(src, pt, 10, cv::Scalar(255, 0, 255), -1); 
     } 

     imshow("My window", src); 
    } 
    } 

cv:waitKey(0); 
    return 0; 
} 
+0

该cameraIntriniscs看起来真的很奇怪...焦距太小(它通常在100s甚至1000s ...并且中心也太小,通常它是在图像中心lke)。这将是就像在一个角落......你是如何获得它的? – api55

+0

@ api55我从ARKIT –

+0

嗯,这很奇怪...他们看起来完全错误,至少对于OpenCV来说,它可能是某种程度上的缩放,或者它使用不同的符号,或两者兼而有之。[Here](http://ksimek.github.io/2013/08/13/intrinsic/)可以找到更多关于相机矩阵的信息。当焦距小于真实焦距时,你会看到较小的物体......这对你的行为是有意义的,而且,在你的情况下,光学中心为0,尝试使用1600作为焦距,图像的中间为中心(例如,如果它是640x480,它会在319.5,239.5),你可能会得到一些让事情更有意义 – api55

回答

0

你内在的参数实在是太小:1.6像素的焦距是物理准针孔摄像头的废话。

+0

我实际上是遵循那篇文章http://nghiaho.com/?page_id=363 如果你想在这里是完整的源代码:https://pastebin.com/9MKx0uCT并使用该图像https ://imgur.com/a/eFzbM –