You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
434 lines
16 KiB
434 lines
16 KiB
/* |
|
* Image.cpp |
|
* |
|
* Copyright (c) 2014-2015 SEACAVE |
|
* |
|
* Author(s): |
|
* |
|
* cDc <cdc.seacave@gmail.com> |
|
* |
|
* |
|
* This program is free software: you can redistribute it and/or modify |
|
* it under the terms of the GNU Affero General Public License as published by |
|
* the Free Software Foundation, either version 3 of the License, or |
|
* (at your option) any later version. |
|
* |
|
* This program is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
|
* GNU Affero General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Affero General Public License |
|
* along with this program. If not, see <http://www.gnu.org/licenses/>. |
|
* |
|
* |
|
* Additional Terms: |
|
* |
|
* You are required to preserve legal notices and author attributions in |
|
* that material or in the Appropriate Legal Notices displayed by works |
|
* containing it. |
|
*/ |
|
|
|
#include "Common.h" |
|
#include "Image.h" |
|
|
|
using namespace MVS; |
|
|
|
|
|
// D E F I N E S /////////////////////////////////////////////////// |
|
|
|
|
|
// S T R U C T S /////////////////////////////////////////////////// |
|
|
|
IMAGEPTR Image::OpenImage(const String& fileName) |
|
{ |
|
#if 0 |
|
if (Util::isFullPath(fileName)) |
|
return IMAGEPTR(CImage::Create(fileName, CImage::READ)); |
|
return IMAGEPTR(CImage::Create((Util::getCurrentFolder()+fileName).c_str(), CImage::READ)); |
|
#else |
|
return IMAGEPTR(CImage::Create(fileName, CImage::READ)); |
|
#endif |
|
} // OpenImage |
|
/*----------------------------------------------------------------*/ |
|
|
|
IMAGEPTR Image::ReadImageHeader(const String& fileName) |
|
{ |
|
IMAGEPTR pImage(OpenImage(fileName)); |
|
if (pImage == NULL || FAILED(pImage->ReadHeader())) { |
|
LOG("error: failed loading image header"); |
|
pImage.Release(); |
|
} |
|
return pImage; |
|
} // ReadImageHeader |
|
/*----------------------------------------------------------------*/ |
|
|
|
IMAGEPTR Image::ReadImage(const String& fileName, Image8U3& image) |
|
{ |
|
IMAGEPTR pImage(OpenImage(fileName)); |
|
if (pImage != NULL && !ReadImage(pImage, image)) |
|
pImage.Release(); |
|
return pImage; |
|
} // ReadImage |
|
/*----------------------------------------------------------------*/ |
|
|
|
bool Image::ReadImage(IMAGEPTR pImage, Image8U3& image) |
|
{ |
|
if (FAILED(pImage->ReadHeader())) { |
|
LOG("error: failed loading image header"); |
|
return false; |
|
} |
|
image.create(pImage->GetHeight(), pImage->GetWidth()); |
|
if (FAILED(pImage->ReadData(image.data, PF_R8G8B8, 3, (CImage::Size)image.step))) { |
|
LOG("error: failed loading image data"); |
|
return false; |
|
} |
|
return true; |
|
} // ReadImage |
|
/*----------------------------------------------------------------*/ |
|
|
|
|
|
bool Image::LoadImage(const String& fileName, unsigned nMaxResolution) |
|
{ |
|
name = fileName; |
|
// open image file |
|
IMAGEPTR pImage(OpenImage(fileName)); |
|
if (pImage == NULL) { |
|
LOG("error: failed opening input image '%s'", name.c_str()); |
|
return false; |
|
} |
|
// create and fill image data |
|
if (!ReadImage(pImage, image)) { |
|
LOG("error: failed loading image '%s'", name.c_str()); |
|
return false; |
|
} |
|
// resize image if needed |
|
scale = ResizeImage(nMaxResolution); |
|
return true; |
|
} // LoadImage |
|
/*----------------------------------------------------------------*/ |
|
|
|
// open the stored image file name and read again the image data |
|
bool Image::ReloadImage(unsigned nMaxResolution, bool bLoadPixels) |
|
{ |
|
IMAGEPTR pImage(bLoadPixels ? ReadImage(name, image) : ReadImageHeader(name)); |
|
if (pImage == NULL) { |
|
LOG("error: failed reloading image '%s'", name.c_str()); |
|
return false; |
|
} |
|
if (!bLoadPixels) { |
|
// init image size |
|
width = pImage->GetWidth(); |
|
height = pImage->GetHeight(); |
|
} |
|
// resize image if needed |
|
scale = ResizeImage(nMaxResolution); |
|
return true; |
|
} // ReloadImage |
|
/*----------------------------------------------------------------*/ |
|
|
|
// free the image data |
|
void Image::ReleaseImage() |
|
{ |
|
image.release(); |
|
} // ReleaseImage |
|
/*----------------------------------------------------------------*/ |
|
|
|
// resize image if needed |
|
// return scale |
|
float Image::ResizeImage(unsigned nMaxResolution) |
|
{ |
|
if (!image.empty()) { |
|
width = image.width(); |
|
height = image.height(); |
|
} |
|
if (nMaxResolution == 0 || MAXF(width,height) <= nMaxResolution) |
|
return 1.f; |
|
const REAL scale(width > height ? (REAL)nMaxResolution/width : (REAL)nMaxResolution/height); |
|
const cv::Size scaledSize(Image8U::computeResize(GetSize(), scale)); |
|
width = (uint32_t)scaledSize.width; |
|
height = (uint32_t)scaledSize.height; |
|
if (!image.empty()) |
|
cv::resize(image, image, scaledSize, 0, 0, cv::INTER_AREA); |
|
return static_cast<float>(scale); |
|
} // ResizeImage |
|
/*----------------------------------------------------------------*/ |
|
|
|
// compute image scale for a given max and min resolution, using the current image file data |
|
unsigned Image::RecomputeMaxResolution(unsigned& level, unsigned minImageSize, unsigned maxImageSize) const |
|
{ |
|
IMAGEPTR pImage(ReadImageHeader(name)); |
|
if (pImage == NULL) { |
|
// something went wrong, use the current known size (however it will most probably fail later) |
|
return Image8U3::computeMaxResolution(width, height, level, minImageSize, maxImageSize); |
|
} |
|
// re-compute max image size |
|
return Image8U3::computeMaxResolution(pImage->GetWidth(), pImage->GetHeight(), level, minImageSize, maxImageSize); |
|
} // RecomputeMaxResolution |
|
/*----------------------------------------------------------------*/ |
|
|
|
|
|
// resize image at desired scale |
|
Image Image::GetImage(const PlatformArr& platforms, double scale, bool bUseImage) const |
|
{ |
|
Image scaledImage(*this); |
|
const cv::Size scaledSize(Image8U::computeResize(GetSize(), scale)); |
|
scaledImage.width = (unsigned)scaledSize.width; |
|
scaledImage.height = (unsigned)scaledSize.height; |
|
scaledImage.camera = GetCamera(platforms, scaledImage.GetSize()); |
|
if (!image.empty()) { |
|
if (bUseImage) |
|
cv::resize(image, scaledImage.image, scaledSize, 0, 0, scale>1?cv::INTER_CUBIC:cv::INTER_AREA); |
|
else |
|
scaledImage.image.release(); |
|
} |
|
return scaledImage; |
|
} // GetImage |
|
// compute the camera extrinsics from the platform pose and the relative camera pose to the platform |
|
Camera Image::GetCamera(const PlatformArr& platforms, const Image8U::Size& resolution) const |
|
{ |
|
ASSERT(platformID != NO_ID); |
|
ASSERT(cameraID != NO_ID); |
|
ASSERT(poseID != NO_ID); |
|
|
|
// compute the normalized absolute camera pose |
|
const Platform& platform = platforms[platformID]; |
|
Camera camera(platform.GetCamera(cameraID, poseID)); |
|
|
|
// compute the unnormalized camera |
|
camera.K = camera.GetK<REAL>(resolution.width, resolution.height); |
|
camera.ComposeP(); |
|
|
|
return camera; |
|
} // GetCamera |
|
void Image::UpdateCamera(const PlatformArr& platforms) |
|
{ |
|
camera = GetCamera(platforms, Image8U::Size(width, height)); |
|
} // UpdateCamera |
|
// computes camera's field of view for the given direction |
|
REAL Image::ComputeFOV(int dir) const |
|
{ |
|
switch (dir) { |
|
case 0: // width |
|
return 2*ATAN(REAL(width)/(camera.K(0,0)*2)); |
|
case 1: // height |
|
return 2*ATAN(REAL(height)/(camera.K(1,1)*2)); |
|
case 2: // diagonal |
|
return 2*ATAN(SQRT(REAL(SQUARE(width)+SQUARE(height)))/(camera.K(0,0)+camera.K(1,1))); |
|
} |
|
return 0; |
|
} // ComputeFOV |
|
/*----------------------------------------------------------------*/ |
|
|
|
|
|
// stereo-rectify the image pair |
|
// - input leftPoints and rightPoints contains the pairs of corresponding image projections and their depth (optional) |
|
// - H converts a pixel from original to rectified left-image |
|
// - Q converts [x' y' disparity 1] in rectified coordinates to [x*z y*z z 1]*w in original image coordinates |
|
// where disparity=x1-x2 |
|
bool Image::StereoRectifyImages(const Image& image1, const Image& image2, const Point3fArr& points1, const Point3fArr& points2, Image8U3& rectifiedImage1, Image8U3& rectifiedImage2, Image8U& mask1, Image8U& mask2, Matrix3x3& H, Matrix4x4& Q) |
|
{ |
|
ASSERT(image1.IsValid() && image2.IsValid()); |
|
ASSERT(image1.GetSize() == image1.image.size() && image2.GetSize() == image2.image.size()); |
|
ASSERT(points1.size() && points2.size()); |
|
|
|
#if 0 |
|
{ // display projection pairs |
|
std::vector<Point2f> matches1, matches2; |
|
FOREACH(i, points1) { |
|
matches1.emplace_back(reinterpret_cast<const Point2f&>(points1[i])); |
|
matches2.emplace_back(reinterpret_cast<const Point2f&>(points2[i])); |
|
} |
|
RECTIFY::DrawMatches(const_cast<Image8U3&>(image1.image), const_cast<Image8U3&>(image2.image), matches1, matches2); |
|
} |
|
#endif |
|
|
|
// compute rectification |
|
Matrix3x3 K1, K2, R1, R2; |
|
#if 0 |
|
const REAL t(Camera::StereoRectify(image1.GetSize(), image1.camera, image2.GetSize(), image2.camera, R1, R2, K1, K2)); |
|
#elif 1 |
|
const REAL t(Camera::StereoRectifyFusiello(image1.GetSize(), image1.camera, image2.GetSize(), image2.camera, R1, R2, K1, K2)); |
|
#else |
|
Pose pose; |
|
ComputeRelativePose(image1.camera.R, image1.camera.C, image2.camera.R, image2.camera.C, pose.R, pose.C); |
|
cv::Mat P1, P2; |
|
cv::stereoRectify(image1.camera.K, cv::noArray(), image2.camera.K, cv::noArray(), image1.GetSize(), pose.R, Vec3(pose.GetTranslation()), R1, R2, P1, P2, Q, 0/*cv::CALIB_ZERO_DISPARITY*/, -1); |
|
K1 = P1(cv::Rect(0,0,3,3)); |
|
K2 = P2(cv::Rect(0,0,3,3)); |
|
const Point3 _t(R2 * pose.GetTranslation()); |
|
ASSERT((ISZERO(_t.x) || ISZERO(_t.y)) && ISZERO(_t.z)); |
|
const REAL t(ISZERO(_t.x)?_t.y:_t.x); |
|
#if 0 |
|
cv::Mat map1, map2; |
|
cv::initUndistortRectifyMap(image1.camera.K, cv::noArray(), R1, K1, image1.GetSize(), CV_16SC2, map1, map2); |
|
cv::remap(image1.image, rectifiedImage1, map1, map2, cv::INTER_CUBIC); |
|
cv::initUndistortRectifyMap(image2.camera.K, cv::noArray(), R2, K2, image1.GetSize(), CV_16SC2, map1, map2); |
|
cv::remap(image2.image, rectifiedImage2, map1, map2, cv::INTER_CUBIC); |
|
return; |
|
#endif |
|
#endif |
|
if (ISZERO(t)) |
|
return false; |
|
|
|
// adjust rectified camera matrices such that the entire area common to both source images is contained in the rectified images |
|
cv::Size size1(image1.GetSize()), size2(image2.GetSize()); |
|
if (!points1.empty()) |
|
Camera::SetStereoRectificationROI(points1, size1, image1.camera, points2, size2, image2.camera, R1, R2, K1, K2); |
|
ASSERT(size1 == size2); |
|
|
|
// compute rectification homography (from original to rectified image) |
|
const Matrix3x3 H1(K1 * R1 * image1.camera.GetInvK()); H = H1; |
|
const Matrix3x3 H2(K2 * R2 * image2.camera.GetInvK()); |
|
|
|
#if 0 |
|
{ // display epipolar lines before and after rectification |
|
Pose pose; |
|
ComputeRelativePose(image1.camera.R, image1.camera.C, image2.camera.R, image2.camera.C, pose.R, pose.C); |
|
const Matrix3x3 F(CreateF(pose.R, pose.C, image1.camera.K, image2.camera.K)); |
|
std::vector<Point2f> matches1, matches2; |
|
#if 1 |
|
FOREACH(i, points1) { |
|
matches1.emplace_back(reinterpret_cast<const Point2f&>(points1[i])); |
|
matches2.emplace_back(reinterpret_cast<const Point2f&>(points2[i])); |
|
} |
|
#endif |
|
RECTIFY::DrawRectifiedImages(image1.image.clone(), image2.image.clone(), F, H1, H2, matches1, matches2); |
|
} |
|
#endif |
|
|
|
// rectify images (apply homographies) |
|
rectifiedImage1.create(size1); |
|
cv::warpPerspective(image1.image, rectifiedImage1, H1, rectifiedImage1.size()); |
|
rectifiedImage2.create(size2); |
|
cv::warpPerspective(image2.image, rectifiedImage2, H2, rectifiedImage2.size()); |
|
|
|
// mark valid regions covered by the rectified images |
|
struct Compute { |
|
static void Mask(Image8U& mask, const cv::Size& sizeh, const cv::Size& size, const Matrix3x3& H) { |
|
mask.create(sizeh); |
|
mask.memset(0); |
|
std::vector<Point2f> corners(4); |
|
corners[0] = Point2f(0,0); |
|
corners[1] = Point2f((float)size.width,0); |
|
corners[2] = Point2f((float)size.width,(float)size.height); |
|
corners[3] = Point2f(0,(float)size.height); |
|
cv::perspectiveTransform(corners, corners, H); |
|
std::vector<std::vector<Point2i>> contours(1); |
|
for (int i=0; i<4; ++i) |
|
contours.front().emplace_back(ROUND2INT(corners[i])); |
|
cv::drawContours(mask, contours, 0, cv::Scalar(255), cv::FILLED); |
|
} |
|
}; |
|
Compute::Mask(mask1, size1, image1.GetSize(), H1); |
|
Compute::Mask(mask2, size2, image2.GetSize(), H2); |
|
|
|
// from the formula that relates disparity to depth as z=B*f/d where B=-t and d=x_l-x_r |
|
// and the formula that converts the image projection from right to left x_r=K1*K2.inv()*x_l |
|
// compute the inverse projection matrix that transforms image coordinates in image 1 and its |
|
// corresponding disparity value to the 3D point in camera 1 coordinates as: |
|
ASSERT(ISEQUAL(K1(1,1),K2(1,1))); |
|
Q = Matrix4x4::ZERO; |
|
// Q * [x, y, disparity, 1] = [X, Y, Z, 1] * w |
|
ASSERT(ISEQUAL(K1(0,0),K2(0,0)) && ISZERO(K1(0,1)) && ISZERO(K2(0,1))); |
|
Q(0,0) = Q(1,1) = REAL(1); |
|
Q(0,3) = -K1(0,2); |
|
Q(1,3) = -K1(1,2); |
|
Q(2,3) = K1(0,0); |
|
Q(3,2) = -REAL(1)/t; |
|
Q(3,3) = (K1(0,2)-K2(0,2))/t; |
|
|
|
// compute Q that converts disparity from rectified to depth in original image |
|
Matrix4x4 P(Matrix4x4::IDENTITY); |
|
cv::Mat(image1.camera.K*R1.t()).copyTo(cv::Mat(4,4,cv::DataType<Matrix4x4::Type>::type,P.val)(cv::Rect(0,0,3,3))); |
|
Q = P*Q; |
|
return true; |
|
} |
|
|
|
// adjust H and Q such that they correspond to rectified images scaled by the given factor |
|
void Image::ScaleStereoRectification(Matrix3x3& H, Matrix4x4& Q, REAL scale) |
|
{ |
|
{ |
|
Matrix3x3 S(Matrix3x3::IDENTITY); |
|
S(0,0) = S(1,1) = scale; |
|
H = S*H*S.inv(); |
|
} |
|
{ |
|
Matrix4x4 Sinv(Matrix4x4::IDENTITY); |
|
Sinv(0,0) = Sinv(1,1) = Sinv(2,2) = REAL(1)/scale; |
|
Matrix4x4 S(Matrix4x4::IDENTITY); |
|
S(0,0) = S(1,1) = scale*scale; |
|
S(2,2) = S(3,3) = scale; |
|
Q = S*Q*Sinv; |
|
} |
|
} |
|
|
|
|
|
// note: disparity has to be inverted as the formula is z=B*f/d where d=x_l-x_r |
|
// while we store d as the value that fulfills x_l+d=x_r |
|
// |
|
// converts the given disparity at the rectified image coordinates to |
|
// the depth corresponding to the un-rectified image coordinates |
|
template <typename TYPE> |
|
float TDisparity2Depth(const Matrix4x4& Q, const TPoint2<TYPE>& u, float d) |
|
{ |
|
const REAL w(Q(3,0)*u.x + Q(3,1)*u.y - Q(3,2)*d + Q(3,3)); |
|
if (ISZERO(w)) |
|
return 0; |
|
const REAL z(Q(2,0)*u.x + Q(2,1)*u.y - Q(2,2)*d + Q(2,3)); |
|
const float depth((float)(z/w)); |
|
return depth < ZEROTOLERANCE<float>() ? float(0) : depth; |
|
} |
|
float Image::Disparity2Depth(const Matrix4x4& Q, const ImageRef& u, float d) |
|
{ |
|
return TDisparity2Depth(Q, u, d); |
|
} |
|
float Image::Disparity2Depth(const Matrix4x4& Q, const Point2f& u, float d) |
|
{ |
|
return TDisparity2Depth(Q, u, d); |
|
} |
|
// same as above, but converts also the coordinates from rectified to original image |
|
template <typename TYPE> |
|
float TDisparity2Depth(const Matrix4x4& Q, const TPoint2<TYPE>& u, float d, Point2f& pt) |
|
{ |
|
const REAL w(Q(3,0)*u.x + Q(3,1)*u.y - Q(3,2)*d + Q(3,3)); |
|
if (ISZERO(w)) |
|
return 0; |
|
const REAL z((Q(2,0)*u.x + Q(2,1)*u.y - Q(2,2)*d + Q(2,3))/w); |
|
if (z < ZEROTOLERANCE<REAL>()) |
|
return 0; |
|
const REAL nrm(REAL(1)/(w*z)); |
|
pt.x = (float)((Q(0,0)*u.x + Q(0,1)*u.y - Q(0,2)*d + Q(0,3))*nrm); |
|
pt.y = (float)((Q(1,0)*u.x + Q(1,1)*u.y - Q(1,2)*d + Q(1,3))*nrm); |
|
return (float)z; |
|
} |
|
float Image::Disparity2Depth(const Matrix4x4& Q, const ImageRef& u, float d, Point2f& pt) |
|
{ |
|
return TDisparity2Depth(Q, u, d, pt); |
|
} |
|
float Image::Disparity2Depth(const Matrix4x4& Q, const Point2f& u, float d, Point2f& pt) |
|
{ |
|
return TDisparity2Depth(Q, u, d, pt); |
|
} |
|
// converts the given disparity at the rectified image coordinates into |
|
// the distance from the camera position to the corresponding 3D point; |
|
// note: the distance is the same in both rectified and un-rectified cameras |
|
// - K is the camera matrix of the un-rectified image |
|
float Image::Disparity2Distance(const Matrix3x3& K, const Matrix4x4& Q, const Point2f& u, float d) |
|
{ |
|
Point2f pt; |
|
const float depth(Disparity2Depth(Q, u, d, pt)); |
|
return (float)(SQRT(SQUARE(pt.x-K(0,2))+SQUARE(pt.y-K(1,2))+SQUARE(K(0,0)))*depth/K(0,0)); |
|
} |
|
// converts the given depth at the un-rectified image coordinates to |
|
// the disparity corresponding to the rectified image coordinates |
|
bool Image::Depth2Disparity(const Matrix4x4& Q, const Point2f& u, float d, float& disparity) |
|
{ |
|
const REAL w((Q(3,0)*u.x + Q(3,1)*u.y + Q(3,2))*d + Q(3,3)); |
|
if (ISZERO(w)) |
|
return false; |
|
const REAL z((Q(2,0)*u.x + Q(2,1)*u.y + Q(2,2))*d + Q(2,3)); |
|
disparity = -(float)(z/w); |
|
return true; |
|
} |
|
/*----------------------------------------------------------------*/
|
|
|