首页 > 代码库 > Learning OpenCV Lecture 4 (Transforming Images with Morphological Operations)
Learning OpenCV Lecture 4 (Transforming Images with Morphological Operations)
In this chapter, we will cover:
- Eroding and dilating images using morphological filters
- Opening and closing images using morphological filters
- Detecting edges and corners using morphological filters
- Segmenting images using watersheds 分水岭算法
- Extracting foreground objects with the GrabCut algorithm
Eroding、dilating、Opening、closing
#include <opencv2/core/core.hpp>#include <opencv2/imgproc/imgproc.hpp>#include <opencv2/highgui/highgui.hpp>int main(){ // Read input image cv::Mat image= cv::imread("../binary.bmp" ); if (!image.data) return 0; // Display the image cv::namedWindow( "Image"); cv::imshow( "Image",image); // Erode the image cv::Mat eroded; cv::erode(image,eroded,cv::Mat()); // Display the eroded image cv::namedWindow( "Eroded Image"); cv::imshow( "Eroded Image",eroded); // Dilate the image cv::Mat dilated; cv::dilate(image,dilated,cv::Mat()); // Display the dialted image cv::namedWindow( "Dilated Image"); cv::imshow( "Dilated Image",dilated); // Erode the image with a larger s.e. cv::Mat element(7,7,CV_8U,cv::Scalar(1)); cv::erode(image,eroded,element); // Display the eroded image cv::namedWindow( "Eroded Image (7x7)"); cv::imshow( "Eroded Image (7x7)",eroded); // Erode the image 3 times. cv::erode(image,eroded,cv::Mat(),cv::Point(-1,-1),3); // Display the eroded image cv::namedWindow( "Eroded Image (3 times)"); cv::imshow( "Eroded Image (3 times)",eroded); // Close the image cv::Mat element5(5,5,CV_8U,cv::Scalar(1)); cv::Mat closed; cv::morphologyEx(image,closed,cv::MORPH_CLOSE,element5); // Display the opened image cv::namedWindow( "Closed Image"); cv::imshow( "Closed Image",closed); // Open the image cv::Mat opened; cv::morphologyEx(image,opened,cv::MORPH_OPEN,element5); // Display the opened image cv::namedWindow( "Opened Image"); cv::imshow( "Opened Image",opened); // Close and Open the image cv::morphologyEx(image,image,cv::MORPH_CLOSE,element5); cv::morphologyEx(image,image,cv::MORPH_OPEN,element5); // Display the close/opened image cv::namedWindow( "Closed and Opened Image"); cv::imshow( "Closed and Opened Image",image); cv::imwrite( "binaryGroup.bmp",image); // Read input image image= cv::imread("../binary.bmp"); // Open and Close the image cv::morphologyEx(image,image,cv::MORPH_OPEN,element5); cv::morphologyEx(image,image,cv::MORPH_CLOSE,element5); // Display the close/opened image cv::namedWindow( "Opened and Closed Image"); cv::imshow( "Opened and Closed Image",image); cv::waitKey(); return 0;}
results:
Detecting edges and corners using morphological filters
morphoFeatures.h
#if !defined MORPHOF#define MORPHOF#include <opencv2/core/core.hpp>#include <opencv2/highgui/highgui.hpp>#include <opencv2/imgproc/imgproc.hpp>class MorphoFeatures {private: // threshold to produce binary image int threshold; // structuring elements used in corner detection cv::Mat cross; cv::Mat diamond; cv::Mat square; cv::Mat x;public: MorphoFeatures() : threshold(-1), cross(5, 5, CV_8U, cv::Scalar(0)), diamond(5, 5, CV_8U, cv::Scalar(0)), square(5, 5, CV_8U, cv::Scalar(0)), x(5, 5, CV_8U, cv::Scalar(0)) { // Creating the cross-shaped structuring element for (int i = 0; i < 5; i++) { cross.at<uchar>(2, i) = 1; cross.at<uchar>(i, 2) = 1; } // Creating the diamond-shaped structuring element diamond.at<uchar>(0, 0) = 0; diamond.at<uchar>(0, 1) = 0; diamond.at<uchar>(1, 0) = 0; diamond.at<uchar>(4, 4) = 0; diamond.at<uchar>(3, 4) = 0; diamond.at<uchar>(4, 3) = 0; diamond.at<uchar>(4, 0) = 0; diamond.at<uchar>(4, 1) = 0; diamond.at<uchar>(3, 0) = 0; diamond.at<uchar>(0, 4) = 0; diamond.at<uchar>(0, 3) = 0; diamond.at<uchar>(1, 4) = 0; // Creating the x-shaped structuring element for (int i = 0; i < 5; i++) { x.at<uchar>(i, i) = 1; x.at<uchar>(4 - i, i) = 1; } } void setThreshold(int t) { if (t > 0) threshold = t; } int getThreshold() const { return threshold; } cv::Mat getEdges(const cv::Mat &image) { // Get the gradient image cv::Mat result; cv::morphologyEx(image, result, cv::MORPH_GRADIENT, cv::Mat()); // Apply threshold to obtain a binary image applyThreshold(result); return result; } void applyThreshold(cv::Mat &result) { // Apply threshold on result if (threshold > 0) { cv::threshold(result, result, threshold, 255, cv::THRESH_BINARY_INV); } } cv::Mat getCorners(const cv::Mat &image) { cv::Mat result; // Dilate with a cross cv::dilate(image, result, cross); // Erode with a diamond cv::erode(result, result, diamond); cv::Mat result2; // Dilate with a x cv::dilate(image, result2, x); // Erode with a square cv::erode(result2, result2, square); // Corners are obtained by differencing // the two closed images cv::absdiff(result2, result, result); // Apply threshold to obtain a binary image applyThreshold(result); return result; } void drawOnImage(const cv::Mat &binary, cv::Mat &image) { cv::Mat_<uchar>::const_iterator it = binary.begin<uchar>(); cv::Mat_<uchar>::const_iterator itend = binary.end<uchar>(); // for each pixel for (int i = 0; it != itend; ++it, ++i) { if (!*it) { cv::circle(image, cv::Point(i%image.step, i/image.step), 5, cv::Scalar(255, 0, 0)); } } }};#endif
morph.cpp
#include <iostream>#include "morphoFeatures.h"int main() { cv::Mat image = cv::imread( "../building.jpg"); cv::cvtColor(image, image, CV_BGR2GRAY); // Create the morphological features instance MorphoFeatures morpho; morpho.setThreshold(40); // Get the edges cv::Mat edges; edges = morpho.getEdges(image); cv::namedWindow( "Edges Image", CV_WINDOW_AUTOSIZE); cv::imshow( "Edges Image", edges); // Get the corners cv::Mat corners; corners = morpho.getCorners(image); // Display the corner on the image morpho.drawOnImage(corners, image); cv::namedWindow( "Corners on Image", CV_WINDOW_AUTOSIZE); cv::imshow( "Corners on Image", image); cv::waitKey(0); return 0;}
results:
Segmenting images using watersheds
watershedSegment.h
#include <opencv2/core/core.hpp>#include <opencv2/highgui/highgui.hpp>#include <opencv2/imgproc/imgproc.hpp>class WatershedSegmenter {private: cv::Mat markers;public: void setMarkers(const cv::Mat &markerImage) { // Convert to image of ints markerImage.convertTo(markers, CV_32S); } cv::Mat process(const cv::Mat &image) { // Apply watershed cv::watershed(image, markers); return markers; } // Return result in the form of an image cv::Mat getSegmentation() { cv::Mat tmp; // all segment with label higher than 255 // will be assigned value 255 markers.convertTo(tmp, CV_8U); return tmp; } // Return watershed in the form of an image cv::Mat getWatersheds() { cv::Mat tmp; // Each pixel p is transform into // 255p + 255 befor conversion markers.convertTo(tmp, CV_8U, 255, 255); return tmp; }};
// Read input image cv::Mat image = cv::imread( "../group.jpg"); if (!image.data) { return 0; } // Display the image cv::namedWindow( "Original Image"); cv::imshow( "Original Image", image); // Get the binary image cv::Mat binary; binary = cv::imread( "../binary.bmp", 0); // Display the binary image cv::namedWindow( "Binary Image"); cv::imshow( "Binary Image", binary); // Eliminate noise and smaller objects cv::Mat fg; cv::erode(binary, fg, cv::Mat(), cv::Point(-1, -1), 6); // Display the foreground image cv::namedWindow( "Foreground Image"); cv::imshow( "Foreground Image", fg);
results:
// Identify image pixels without objects cv::Mat bg; cv::dilate(binary, bg, cv::Mat(), cv::Point(-1, -1), 6); cv::threshold(bg, bg, 1, 128, cv::THRESH_BINARY_INV); // Display the backgroud image cv::namedWindow( "Background Image"); cv::imshow( "Background Image", bg);
results:
// Show markers image cv::Mat markers(binary.size(), CV_8U, cv::Scalar(0)); markers = fg + bg; cv::namedWindow( "Markers"); cv::imshow( "Markers", markers);
// Create watershed segmentation object WatershedSegmenter segmenter; // Set markers and process segmenter.setMarkers(markers); segmenter.process(image); // Display segmentation result cv::namedWindow( "Segmentation"); cv::imshow( "Segmentation", segmenter.getSegmentation()); // Display watersheds cv::namedWindow( "Watershed"); cv::imshow( "Watershed", segmenter.getWatersheds());
// Open another image------------------------------------ image = cv::imread( "../tower.jpg"); // Identify background pixels cv::Mat imageMask(image.size(), CV_8U, cv::Scalar(0)); cv::rectangle(imageMask, cv::Point(5, 5), cv::Point(image.cols - 5, image.rows - 5), cv::Scalar(255), 3); // Identify forground pixels (in the middle of the image) cv::rectangle(imageMask, cv::Point(image.cols / 2 - 10, image.rows / 2 - 10), cv::Point(image.cols / 2 + 10, image.rows / 2 + 10), cv::Scalar(1), 10); // Set markers and process segmenter.setMarkers(imageMask); segmenter.process(image); // Display the image with markers cv::rectangle(image, cv::Point(5, 5), cv::Point(image.cols - 5, image.rows - 5), cv::Scalar(255, 255, 255), 3); cv::rectangle(image, cv::Point(image.cols / 2 - 10, image.rows / 2 - 10), cv::Point(image.cols / 2 + 10, image.rows / 2 + 10), cv::Scalar(1, 1, 1), 10); cv::namedWindow( "Image with marker"); cv::imshow( "Image with marker", image); // Display watersheds cv::namedWindow( "Watersheds of foreground object"); cv::imshow( "Watersheds of foreground object", segmenter.getWatersheds());
results:
Extracting foreground objects with the GrabCut algorithm
// Open another image image = cv::imread( "../tower.jpg"); // define bounding rectange cv::Rect rectangle(50, 70, image.cols - 150, image.rows - 180); cv::Mat result; // segmentation result (4 possible values) cv::Mat bgModel, fgModel; // the models (internally used) // GrabCut segmentation cv::grabCut(image, // input image result, // segmentation result rectangle, // rectangle containing foreground bgModel, fgModel, // models 1, //number of iterations cv::GC_INIT_WITH_RECT// use rectangle ); // Get the pixles marked as likely foreground cv::compare(result, cv::GC_PR_FGD, result, cv::CMP_EQ); // Generate output image cv::Mat foreground(image.size(), CV_8UC3, cv::Scalar(255, 255, 255)); image.copyTo(foreground, result); // bg pixels not copied // draw rectangle on original image cv::rectangle(image, rectangle, cv::Scalar(255,255,255),1); cv::namedWindow( "Image"); cv::imshow( "Image",image); // display result cv::namedWindow( "Segmented Image"); cv::imshow( "Segmented Image",foreground);
// Open another image image= cv::imread("../group.jpg"); // define bounding rectangle cv::Rect rectangle2(10,100,380,180); cv::Mat bkgModel,fgrModel; // the models (internally used) // GrabCut segmentation cv::grabCut(image, // input image result, // segmentation result rectangle2,bkgModel,fgrModel,5,cv::GC_INIT_WITH_RECT); // Get the pixels marked as likely foreground // cv::compare(result,cv::GC_PR_FGD,result,cv::CMP_EQ); result= result&1; foreground.create(image.size(),CV_8UC3); foreground.setTo(cv::Scalar(255,255,255)); image.copyTo(foreground,result); // bg pixels not copied // draw rectangle on original image cv::rectangle(image, rectangle2, cv::Scalar(255,255,255),1); cv::namedWindow( "Image 2"); cv::imshow( "Image 2",image); // display result cv::namedWindow( "Foreground objects"); cv::imshow( "Foreground objects",foreground);
声明:以上内容来自用户投稿及互联网公开渠道收集整理发布,本网站不拥有所有权,未作人工编辑处理,也不承担相关法律责任,若内容有误或涉及侵权可进行投诉: 投诉/举报 工作人员会在5个工作日内联系你,一经查实,本站将立刻删除涉嫌侵权内容。