使用cv::ximgproc::FastBilateralSolverFilter优化realsense D455的深度图

1 源码

OpenCV中内置了很多滤波器,这里我们讨论cv::ximgproc其中包含的滤波器。
https://docs.opencv.org/3.4/da/d17/group__ximgproc__filters.html
需要注意的是,默认安装的OpenCV中不包含cv::ximgproc,请从源码重修编译。
在这里插入图片描述
在这里贴上我的测试源码。

/*
cv::ximgproc::AdaptiveManifoldFilter
[86] Eduardo SL Gastal and Manuel M Oliveira. Adaptive manifolds for real-time high-dimensional filtering. ACM Transactions on Graphics (TOG), 31(4):33, 2012.

cv::ximgproc::EdgeAwareInterpolator
[182] Jerome Revaud, Philippe Weinzaepfel, Zaid Harchaoui, and Cordelia Schmid. Epicflow: Edge-preserving interpolation of correspondences for optical flow. In Computer Vision and Pattern Recognition (CVPR), IEEE Conference on, pages 1164–1172, 2015.

cv::ximgproc::DTFilter
[85] Eduardo SL Gastal and Manuel M Oliveira. Domain transform for edge-aware image and video processing. In ACM Transactions on Graphics (TOG), volume 30, page 69. ACM, 2011.

cv::ximgproc::DisparityWLSFilter

cv::ximgproc::FastBilateralSolverFilter
[14] Jonathan T Barron and Ben Poole. The fast bilateral solver. In European Conference on Computer Vision (ECCV), pages 617–632. Springer International Publishing, 2016.

cv::ximgproc::FastGlobalSmootherFilter
[158] Dongbo Min, Sunghwan Choi, Jiangbo Lu, Bumsub Ham, Kwanghoon Sohn, and Minh N Do. Fast global image smoothing based on weighted least squares. Image Processing, IEEE Transactions on, 23(12):5638–5653, 2014.
[66] Zeev Farbman, Raanan Fattal, Dani Lischinski, and Richard Szeliski. Edge-preserving decompositions for multi-scale tone and detail manipulation. In ACM Transactions on Graphics (TOG), volume 27, page 67. ACM, 2008.

cv::ximgproc::GuidedFilter
[99] Kaiming He, Jian Sun, and Xiaoou Tang. Guided image filtering. In Computer Vision–ECCV 2010, pages 1–14. Springer, 2010.

cv::ximgproc::RidgeDetectionFilter
[64] Niki Estner. Best way of segmenting veins in leaves.
[151] Wolfram Mathematica. Ridge filter mathematica.

*/



#include <ros/ros.h>
#include <sensor_msgs/Image.h>
#include <cv_bridge/cv_bridge.h>
#include <image_transport/image_transport.h>
#include <opencv2/imgproc/imgproc.hpp>
#include <opencv2/highgui/highgui.hpp>
#include <opencv2/ximgproc.hpp>
#include <message_filters/subscriber.h>
#include <message_filters/time_synchronizer.h>
#include <message_filters/sync_policies/approximate_time.h>

#include <pcl/point_types.h>
#include <pcl_ros/point_cloud.h>
#include <pcl/common/common.h>
#include <pcl/common/pca.h>
#include <Eigen/Dense>

class ImageProcessor
{
public:
    ImageProcessor()
    : it_(nh_), image_sub_(nh_, "/camera/color/image_raw", 1), depth_sub_(nh_, "/camera/aligned_depth_to_color/image_raw", 1)
    {
        typedef message_filters::sync_policies::ApproximateTime<sensor_msgs::Image, sensor_msgs::Image> MySyncPolicy;
        // ApproximateTime或者ExactTime
        sync_.reset(new message_filters::Synchronizer<MySyncPolicy>(MySyncPolicy(10), image_sub_, depth_sub_));
        sync_->registerCallback(boost::bind(&ImageProcessor::callback, this, _1, _2));

        pub_ = it_.advertise("/fbs/depth_processed", 1);
        cloud_pub_ = nh_.advertise<pcl::PointCloud<pcl::PointXYZRGB>>("/fbs/output_cloud", 1);
        cloud2_pub_ = nh_.advertise<pcl::PointCloud<pcl::PointXYZRGB>>("/fbs/output_cloud2", 1);

    }

    void callback(const sensor_msgs::ImageConstPtr& color_msg, const sensor_msgs::ImageConstPtr& depth_msg)
    {
        cv::Mat color_image, depth_image;

        try {
            color_image = cv_bridge::toCvShare(color_msg, "rgb8")->image;
            depth_image = cv_bridge::toCvShare(depth_msg, sensor_msgs::image_encodings::TYPE_16UC1)->image;

            cv::Mat confidence;
            generateConfidence(depth_image, confidence);

            // 将confidence图像标准化到0-255的范围
            cv::Mat normalized_confidence;
            confidence.convertTo(normalized_confidence, CV_8UC1, 255.0); // 将浮点值乘以255以进行转换

            // // 显示信心图
            // cv::namedWindow("Confidence", cv::WINDOW_AUTOSIZE); // 创建一个窗口
            // cv::imshow("Confidence", normalized_confidence);    // 显示信心图
            // cv::waitKey(1); // 等待1毫秒,让显示函数有机会更新窗口

            cv::Mat filtered_image;
            applyFastBilateralSolverFilter(depth_image, color_image, confidence, filtered_image);
            // applyFastGlobalSmootherFilter(depth_image, color_image, filtered_image);
            // applyGuidedFilter(depth_image, color_image, filtered_image);
            // applyRidgeDetectionFilter(depth_image, filtered_image);
            
            // 转换回ROS图像并发布
            sensor_msgs::ImagePtr msg_out = cv_bridge::CvImage(std_msgs::Header(), sensor_msgs::image_encodings::TYPE_16UC1, filtered_image).toImageMsg();
            pub_.publish(msg_out);

            // 发布点云
            pcl::PointCloud<pcl::PointXYZRGB>::Ptr cloud(new pcl::PointCloud<pcl::PointXYZRGB>());
            depthToPointCloud(depth_image, color_image, *cloud);
            cloud->header.stamp = pcl_conversions::toPCL(color_msg->header.stamp); // Use the correct message for the timestamp
            cloud_pub_.publish(cloud);

            // 发布点云
            // pcl::PointCloud<pcl::PointXYZRGB>::Ptr cloud(new pcl::PointCloud<pcl::PointXYZRGB>());
            depthToPointCloud(filtered_image, color_image, *cloud);
            // cloud->header.stamp = pcl_conversions::toPCL(color_msg->header.stamp); // Use the correct message for the timestamp
            cloud2_pub_.publish(cloud);

        } catch (const cv_bridge::Exception& e) {
            ROS_ERROR("cv_bridge exception: %s", e.what());
            return;
        }
    }

    void generateConfidence(const cv::Mat &depth, cv::Mat &confidence)
    {
        confidence = cv::Mat::zeros(depth.size(), CV_32F);
        confidence.setTo(1, depth != 0);
    }

    void applyFastBilateralSolverFilter(const cv::Mat &depth, const cv::Mat &color, const cv::Mat &confidence, cv::Mat &output)
    {
        auto filter = cv::ximgproc::createFastBilateralSolverFilter(color, 10, 1, 10);
        filter->filter(depth, confidence, output);
    }

    void applyFastGlobalSmootherFilter(const cv::Mat &depth, const cv::Mat &color, cv::Mat &output)
    {
        // 将depth转换为16S格式
        cv::Mat depth16S;
        depth.convertTo(depth16S, CV_16S);

        // 创建Fast Global Smoother滤波器
        auto filter = cv::ximgproc::createFastGlobalSmootherFilter(color, 128, 10);

        // 使用16S格式的depth进行滤波
        filter->filter(depth16S, output);
    }

    void applyGuidedFilter(const cv::Mat &depth, const cv::Mat &color, cv::Mat &output)
    {
        // 将depth图像转换为CV_32F类型
        cv::Mat depth32F;
        cv::Mat output32F;
        depth.convertTo(depth32F, CV_32F, 1.0 / 255.0);
        
        auto filter = cv::ximgproc::createGuidedFilter(color, 5, 0.01);
        filter->filter(depth32F, output32F);

        output32F.convertTo(output, CV_16U, 255.0);
    }

    void applyRidgeDetectionFilter(const cv::Mat &depth, cv::Mat &output)
    {
        // 将图像转换为浮点类型,这是脊线检测滤波器的要求
        cv::Mat depth32F;
        depth.convertTo(depth32F, CV_32FC1, 1.0 / 255.0);

        // 创建脊线检测滤波器
        int ddepth = -1; // 使用相同的图像深度作为输入和输出
        double scale = 1; // Sobel算子的比例因子
        double delta = 0; // 添加到结果中的可选增量值
        int borderType = cv::BORDER_DEFAULT; // 边缘填充方式

        cv::Ptr<cv::ximgproc::RidgeDetectionFilter> ridgeFilter = cv::ximgproc::RidgeDetectionFilter::create();

        // 应用脊线检测滤波器
        cv::Mat output32F;
        ridgeFilter->getRidgeFilteredImage(depth32F, output32F);

        // 将结果转换回8位图像以便显示
        output32F.convertTo(output, CV_16UC1, 255.0);
    }

    void depthToPointCloud(const cv::Mat& depth_img, const cv::Mat& rgb_img, pcl::PointCloud<pcl::PointXYZRGB>& cloud) {
        // 假设 fx, fy, cx, cy 是相机内参
        float fx = 644.7825927734375; // 焦距x
        float fy = 644.0340576171875; // 焦距y
        float cx = 636.1322631835938; // 主点x
        float cy = 367.4698486328125; // 主点y

        cloud.width = depth_img.cols;
        cloud.height = depth_img.rows;
        cloud.is_dense = false;
        cloud.points.resize(cloud.width * cloud.height);
        cloud.header.frame_id = "map";


        for (int v = 0; v < depth_img.rows; v++) {
            for (int u = 0; u < depth_img.cols; u++) {
                pcl::PointXYZRGB& point = cloud.points[v * depth_img.cols + u];

                float depth = depth_img.at<ushort>(v, u) / 1000.0f; // 深度单位转换为米
                if (depth == 0) continue; // 无效点跳过

                // 从像素坐标转换到相机坐标系
                point.x = (u - cx) * depth / fx;
                point.y = (v - cy) * depth / fy;
                point.z = depth;

                // 设置颜色
                cv::Vec3b rgb = rgb_img.at<cv::Vec3b>(v, u);
                uint32_t rgb_val = ((uint32_t)rgb[0] << 16 | (uint32_t)rgb[1] << 8 | (uint32_t)rgb[2]);
                point.rgb = *reinterpret_cast<float*>(&rgb_val);
            }
        }
    }

private:
    ros::NodeHandle nh_;
    image_transport::ImageTransport it_;
    image_transport::Publisher pub_;
    message_filters::Subscriber<sensor_msgs::Image> image_sub_, depth_sub_;
    boost::shared_ptr<message_filters::Synchronizer<message_filters::sync_policies::ApproximateTime<sensor_msgs::Image, sensor_msgs::Image>>> sync_;
    ros::Publisher cloud_pub_;
    ros::Publisher cloud2_pub_;
};

int main(int argc, char** argv)
{
    ros::init(argc, argv, "image_processor");
    ImageProcessor ip;
    ros::spin();
    return 0;
}


这里FBS和其他的滤波器的本质区别是,FBS引入了信心图(confidence),这使得其在进行滤波时不会被无效像素所影响,因此我建议,当使用无信心图的滤波器时,尽量使用其他后处理方法填充空洞,减少无效像素;当使用FBS时,建议深度图宁缺毋滥,尽量保持真值。

2 效果展示

这是realsense D455的原始深度图与经过FBS的深度图是对比。

nullrawFBS
深度图请添加图片描述请添加图片描述
深度图请添加图片描述请添加图片描述
点云在这里插入图片描述在这里插入图片描述

看起来似乎不错,但我们将其转换为点云就会发现,原始深度图干干净净,不同深度处的物体分离性很好;但是经过优化的深度图产生了大量的离群点,这些是FBS“猜”错的点。
请添加图片描述
这个视角下我们看到,FBS确实补全了部分缺失的深度值,但也带来了很大问题,最主要的问题就是这些错误的点。

  • 5
    点赞
  • 7
    收藏
    觉得还不错? 一键收藏
  • 1
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值